code stringlengths 281 23.7M |
|---|
def test_service_get(service: Service, default_entity_dict):
service.create(ServeRequest(**default_entity_dict))
entity: ServerResponse = service.get(ServeRequest(**default_entity_dict))
with db.session() as session:
db_entity: ServeEntity = session.get(ServeEntity, entity.id)
assert (db_ent... |
()
def enqueue_block_list_command():
broker = connect_broker()
inspect_many_blocks_actor = dramatiq.actor(inspect_many_blocks_task, broker=broker, queue_name=LOW_PRIORITY_QUEUE, priority=LOW_PRIORITY)
for block_string in fileinput.input():
block = int(block_string)
logger.info(f'Sending {blo... |
('pyscf')
def test_run_dict_freeze():
run_dict = {'geom': {'type': 'cart', 'fn': 'lib:h2o.xyz', 'freeze_atoms': [0, 1]}, 'calc': {'type': 'pyscf', 'basis': 'sto3g', 'pal': 1, 'verbose': 0}, 'opt': {}}
result = run_from_dict(run_dict)
geom = result.opt_geom
assert (geom.energy == pytest.approx((- 74.))) |
_repr
class NotificationSettings(db.Model, Timestamp):
id = db.Column(db.Integer, primary_key=True)
type = db.Column(db.String, nullable=False)
enabled = db.Column(db.Boolean, default=True, nullable=False, server_default='True')
def is_enabled(type: str) -> bool:
(settings, _) = get_or_create(No... |
class IlluminaRunInfo():
def __init__(self, runinfo_xml):
self.runinfo_xml = runinfo_xml
self.run_id = None
self.run_number = None
self.instrument = None
self.flowcell = None
self.date = None
self.lane_count = None
self.reads = []
doc = xml.dom... |
def validate_not_quarterly_to_monthly(source_fiscal_period, destination_fiscal_period):
if (is_quarter_final_period(source_fiscal_period) and (not is_quarter_final_period(destination_fiscal_period))):
logger.warning('WARNING: You are attempting to copy quarterly data into a monthly period. This is fine, h... |
class CustomForms(db.Model):
class TYPE():
ATTENDEE = 'attendee'
SESSION = 'session'
SPEAKER = 'speaker'
__tablename__ = 'custom_forms'
__table_args__ = (UniqueConstraint('event_id', 'field_identifier', 'form', 'form_id', name='custom_form_identifier'),)
id = db.Column(db.Integer... |
def _start():
global patch, name, path, monitor
global timeout, lsl_name, lsl_type, ft_host, ft_port, ft_output, start, selected, streams, stream, inlet, type, source_id, match, lsl_id, channel_count, channel_format, nominal_srate, samples, blocksize
timeout = patch.getfloat('lsl', 'timeout', default=30)
... |
def test_pod_config():
with pytest.raises(user.FlyteValidationException):
Pod(pod_spec=None)
with pytest.raises(user.FlyteValidationException):
Pod(pod_spec=V1PodSpec(containers=[]), primary_container_name=None)
selector = {'node_group': 'memory'}
(task_config=Pod(pod_spec=V1PodSpec(cont... |
class TestDataFrameSample(TestData):
SEED = 42
def build_from_index(self, sample_ed_flights):
sample_pd_flights = self.pd_flights_small().loc[(sample_ed_flights.index, sample_ed_flights.columns)]
return sample_pd_flights
def test_sample(self):
ed_flights_small = self.ed_flights_small... |
_DATA_MAPPER_REGISTRY.register()
class RotatedDatasetMapper(D2GoDatasetMapper):
def _original_call(self, dataset_dict):
dataset_dict = copy.deepcopy(dataset_dict)
image = self._read_image(dataset_dict, format=self.img_format)
if (not self.backfill_size):
utils.check_image_size(da... |
class OptionPlotoptionsSankeySonificationTracksActivewhen(Options):
def crossingDown(self):
return self._config_get(None)
def crossingDown(self, num: float):
self._config(num, js_type=False)
def crossingUp(self):
return self._config_get(None)
def crossingUp(self, num: float):
... |
class OptionPlotoptionsDumbbell(Options):
def accessibility(self) -> 'OptionPlotoptionsDumbbellAccessibility':
return self._config_sub_data('accessibility', OptionPlotoptionsDumbbellAccessibility)
def allowPointSelect(self):
return self._config_get(False)
def allowPointSelect(self, flag: boo... |
class HTTPCachingDatabase(CachingDatabase, _HTTPLogic):
def __init__(self):
from staticdhcpdlib import config
if hasattr(config, 'X_HTTPDB_CONCURRENCY_LIMIT'):
CachingDatabase.__init__(self, concurrency_limit=config.X_HTTPDB_CONCURRENCY_LIMIT)
else:
CachingDatabase.__... |
class ClipFieldGrp(object):
def __init__(self):
self.parent = None
self.layout = None
self.shot_name_field = None
self.fbx_file_field = None
self.fbx_choose_button = None
self.video_file_field = None
self.video_choose_button = None
self.cut_in_field = ... |
def test_error_when_regression_is_false_and_target_is_continuous(df_discretise):
np.random.seed(42)
(mu, sigma) = (0, 3)
y = np.random.normal(mu, sigma, len(df_discretise))
with pytest.raises(ValueError):
transformer = DecisionTreeDiscretiser(regression=False)
transformer.fit(df_discreti... |
('config_name,overrides,expected', [param('error_duplicate_group', [], raises(ConfigCompositionException, match=re.escape('group1 appears more than once in the final defaults list')), id='error_duplicate_group'), param('error_duplicate_group_nested', [], raises(ConfigCompositionException, match=re.escape('group1/group2... |
def local_master_get_detection_dataset_dicts(*args, **kwargs):
logger.info('Only load dataset dicts on local master process ...')
dataset_dicts = (d2_get_detection_dataset_dicts(*args, **kwargs) if (comm.get_local_rank() == 0) else [])
comm.synchronize()
dataset_size = comm.all_gather(len(dataset_dicts)... |
def test_stop(tmpdir):
cloud_provider = MagicMock()
cloud_provider.start_instance.return_value = {'instance_id': 'test-id'}
cloud_provider.get_ips.return_value = ('127.0.0.1', '127.0.0.1')
states = [True, True, True, True, True, False, False, False, False, False]
def is_running(instance_id):
... |
.parametrize('path,expected_params', [('/cvt/teams/007', {'id': 7}), ('/cvt/teams/1234/members', {'id': 1234}), ('/cvt/teams/default/members/700-5', {'id': 700, 'tenure': 5}), ('/cvt/repos/org/repo/compare/xkcd:353', {'org': 'org', 'repo': 'repo', 'usr0': 'xkcd', 'branch0': 353}), ('/cvt/repos/org/repo/compare/gunmacha... |
class OptionSeriesLollipopStatesSelectHalo(Options):
def attributes(self):
return self._config_get(None)
def attributes(self, value: Any):
self._config(value, js_type=False)
def opacity(self):
return self._config_get(0.25)
def opacity(self, num: float):
self._config(num, ... |
class TypeCustomer(ModelSimple):
allowed_values = {('value',): {'CUSTOMER': 'customer'}}
validations = {}
additional_properties_type = None
_nullable = False
_property
def openapi_types():
return {'value': (str,)}
_property
def discriminator():
return None
attribute_m... |
def downgrade():
op.drop_column('connectionconfig', 'disabled_at')
op.drop_column('connectionconfig', 'disabled')
op.execute('alter type executionlogstatus rename to executionlogstatus_old')
op.execute("create type executionlogstatus as enum('in_processing', 'pending', 'complete', 'error', 'retrying', '... |
def test_data_integrity_test_columns_type() -> None:
current_dataset = pd.DataFrame({'numerical_feature': [1, 2, 3], 'target': ['1', '1', '1'], 'datetime': [datetime.now(), datetime.now(), datetime.now()]})
reference_dataset = pd.DataFrame({'numerical_feature': [1.0, 2.4, 3.0], 'target': [True, False, True], 'd... |
class Dataset():
def __init__(self):
self.data = []
def add_sample(self, sample):
self.data.append(sample)
def size(self):
return len(self.data)
def inputs(self):
return [sample.input for sample in self.data]
def shuffle(self):
random.shuffle(self.data)
de... |
def _load_sbv2_pub_key(keydata):
vk = serialization.load_pem_public_key(keydata, backend=default_backend())
if isinstance(vk, rsa.RSAPublicKey):
if (vk.key_size != 3072):
raise esptool.FatalError(('Key file has length %d bits. Secure boot v2 only supports RSA-3072.' % vk.key_size))
r... |
def extractHellYeah524(item):
(vol, chp, frag, postfix) = extractVolChapterFragmentPostfix(item['title'])
if ((not (chp or vol or frag)) or ('preview' in item['title'].lower())):
return None
if ((item['tags'] == ['Uncategorized']) and item['title'].startswith('Chapter')):
return buildRelease... |
class Chef(models.Model):
slug = models.SlugField()
first_name = models.CharField(max_length=60)
last_name = models.CharField(max_length=60)
uuid_id = models.UUIDField()
email_address = models.EmailField()
twitter_profile = models.URLField()
def __str__(self):
return 'Chef {} {}'.for... |
def test_json_parses_to_policy_document():
pd = PolicyDocument(Version='2012-10-17', Statement=[Statement(Effect='Allow', Action=[Action('autoscaling', 'DescribeLaunchConfigurations')], Resource=['*']), Statement(Effect='Allow', Action=[Action('sts', 'AssumeRole')], Resource=['arn:aws:iam:::role/someRole'])])
a... |
class OptionSeriesStreamgraphLabel(Options):
def boxesToAvoid(self):
return self._config_get(None)
def boxesToAvoid(self, value: Any):
self._config(value, js_type=False)
def connectorAllowed(self):
return self._config_get(False)
def connectorAllowed(self, flag: bool):
sel... |
def llamacpp_loader(llm_adapter: LLMModelAdapter, model_params: LlamaCppModelParameters):
try:
from dbgpt.model.llm.llama_cpp.llama_cpp import LlamaCppModel
except ImportError as exc:
raise ValueError('Could not import python package: llama-cpp-python Please install db-gpt llama support with `cd... |
class RunningMomentsTest(TestCaseBase):
def setUp(self) -> None:
self.outer_size = 10
self.inner_size = (4, 5)
self.running_moments = RunningMoments(self.inner_size)
self.rtol = 1e-06
self.atol = 1e-06
def test_single_update(self) -> None:
input = torch.rand(self.... |
class EventModifyRequest(ryu_event.EventRequestBase):
def __init__(self, system_id, func):
super(EventModifyRequest, self).__init__()
self.dst = 'OVSDB'
self.system_id = system_id
self.func = func
def __str__(self):
return ('%s<system_id=%s>' % (self.__class__.__name__, s... |
class LTTextLineHorizontal(LTTextLine):
def __init__(self, word_margin):
LTTextLine.__init__(self, word_margin)
self._x1 = (+ INF)
return
def add(self, obj):
if (isinstance(obj, LTChar) and self.word_margin):
margin = (self.word_margin * max(obj.width, obj.height))
... |
.usefixtures('use_tmpdir')
def test_run_version_env_with_user_env(monkeypatch, source_root):
wrapper_file_name = f'{os.getcwd()}/bin/rms_wrapper'
with open('rms_config.yml', 'w', encoding='utf-8') as f:
f.write(f'''executable: {os.getcwd()}/bin/rms
wrapper: {wrapper_file_name}
env:
10.1.3:
... |
def build_ensemble(cls, **kwargs):
ens = cls(**kwargs)
use = ['ExtraTrees', 'RandomForest', 'LogisticRegression-SAG', 'MLP-adam']
meta = RandomForestClassifier(n_estimators=100, random_state=0, n_jobs=(- 1))
base_learners = list()
for (est_name, est) in ESTIMATORS.items():
e = clone(est)
... |
def edit_forum_topic(token, chat_id, message_thread_id, name=None, icon_custom_emoji_id=None):
method_url = 'editForumTopic'
payload = {'chat_id': chat_id, 'message_thread_id': message_thread_id}
if (name is not None):
payload['name'] = name
if (icon_custom_emoji_id is not None):
payload... |
def format_invoice_document_data(original_response: Dict) -> InvoiceParserDataClass:
fields = original_response[0].get('fields', [])
items: Sequence[ItemLinesInvoice] = extract_item_lignes(fields, ItemLinesInvoice)
default_dict = defaultdict((lambda : None))
merchant_name = fields.get('companyName', def... |
class OptionPlotoptionsPolygonSonificationTracksActivewhen(Options):
def crossingDown(self):
return self._config_get(None)
def crossingDown(self, num: float):
self._config(num, js_type=False)
def crossingUp(self):
return self._config_get(None)
def crossingUp(self, num: float):
... |
def test_dependency_getter(tmp_path: Path) -> None:
fake_pyproject_toml = '\n[tool.poetry.dependencies]\npython = ">=3.7,<4.0"\nbar = { version = ">=2.5.1,<4.0.0", python = ">3.7" }\nfoo-bar = { version = ">=2.5.1,<4.0.0", optional = true, python = ">3.7" }\nfox-python = "*" # top level module is called "fox"\n\... |
def cbFun(snmpEngine, sendRequestHandle, errorIndication, errorStatus, errorIndex, varBindTable, cbCtx):
(authData, transportTarget) = cbCtx
print(('%s via %s' % (authData, transportTarget)))
if errorIndication:
print(errorIndication)
return
elif errorStatus:
print(('%s at %s' % ... |
class TestTachoMotorSpeedPValue(ptc.ParameterizedTestCase):
def test_speed_i_negative(self):
with self.assertRaises(IOError):
self._param['motor'].speed_p = (- 1)
def test_speed_p_zero(self):
self._param['motor'].speed_p = 0
self.assertEqual(self._param['motor'].speed_p, 0)
... |
class SigningDialogues(Model, BaseSigningDialogues):
def __init__(self, **kwargs: Any) -> None:
Model.__init__(self, **kwargs)
def role_from_first_message(message: Message, receiver_address: Address) -> BaseDialogue.Role:
return BaseSigningDialogue.Role.SKILL
BaseSigningDialogues... |
def add_commands(subparsers, efuses):
add_common_commands(subparsers, efuses)
burn_key = subparsers.add_parser('burn_key', help='Burn the key block with the specified name')
protect_options(burn_key)
add_force_write_always(burn_key)
add_show_sensitive_info_option(burn_key)
burn_key.add_argument(... |
class Number(Field):
numeric_type: typing.Optional[type] = None
errors = {'type': 'Must be a number.', 'null': 'May not be null.', 'integer': 'Must be an integer.', 'finite': 'Must be finite.', 'minimum': 'Must be greater than or equal to {minimum}.', 'exclusive_minimum': 'Must be greater than {exclusive_minimu... |
class Packer(_Packer):
def __init__(self, *args, **kwargs):
default = kwargs.get('default')
for encoder in msgpack_encoders.get_all().values():
default = functools.partial(encoder, chain=default)
kwargs['default'] = default
super(Packer, self).__init__(*args, **kwargs) |
def extractXvvCpuMybluehostMe(item):
(vol, chp, frag, postfix) = extractVolChapterFragmentPostfix(item['title'])
if ((not (chp or vol)) or ('preview' in item['title'].lower())):
return None
tagmap = [('PRC', 'PRC', 'translated'), ('Loiterous', 'Loiterous', 'oel')]
for (tagname, name, tl_type) in... |
class Region():
def __init__(self, key, title, *args):
self.key = key
self.title = title
self.inherited = ((args and (args[0] == 'inherited')) or False)
self._content_types = []
def __str__(self):
return force_str(self.title)
def content_types(self):
return [(... |
def _create_unfiltered_taxable_event_set(configuration: Configuration, input_data: InputData) -> TransactionSet:
transaction_set: TransactionSet
entry: AbstractEntry
transaction: AbstractTransaction
taxable_event_set: TransactionSet = TransactionSet(configuration, 'MIXED', input_data.asset, MIN_DATE, MA... |
def _subcfg(env, eff):
if isinstance(eff, Effects.effset):
return Effects.effset(eff.buffer, [_subcfg(env, e) for e in eff.loc], eff.names, (_subcfg(env, eff.pred) if eff.pred else None), eff.srcinfo)
elif isinstance(eff, Effects.config_eff):
value = (_subcfg(env, eff.value) if eff.value else No... |
def _jit_dipole_magnetic_field_cartesian(coordinates, dipoles, magnetic_moments, b_e, b_n, b_u):
(easting, northing, upward) = coordinates
(easting_p, northing_p, upward_p) = dipoles
for l in prange(easting.size):
for m in range(easting_p.size):
(easting_comp, northing_comp, upward_comp)... |
def main():
module_spec = schema_to_module_spec(versioned_schema)
mkeyname = 'id'
fields = {'access_token': {'required': False, 'type': 'str', 'no_log': True}, 'enable_log': {'required': False, 'type': 'bool', 'default': False}, 'vdom': {'required': False, 'type': 'str', 'default': 'root'}, 'member_path': {... |
class OrdnerMainPage(Webpage):
def __init__(self, ordner, index_splits):
self.ordner = ordner
self.index_splits = index_splits
self.filepath = 'build/html/ordner/index.html'
self.title = 'Ordner: index of real numbers'
self.pagetitle = 'Ordner: index of real numbers - Fungrim... |
def _check_dipoles_and_magnetic_moments(dipoles, magnetic_moments):
if (magnetic_moments.shape[1] != 3):
raise ValueError(f"Invalid magnetic moments with '{magnetic_moments.shape[1]}' elements. Magnetic moments vectors should have 3 components.")
if (magnetic_moments.shape[0] != dipoles[0].size):
... |
class OptionSeriesDependencywheelSonificationDefaultinstrumentoptionsMappingHighpassFrequency(Options):
def mapFunction(self):
return self._config_get(None)
def mapFunction(self, value: Any):
self._config(value, js_type=False)
def mapTo(self):
return self._config_get(None)
def ma... |
class SourceEmbedding(ABC):
def __init__(self, file_path, vector_store_config: {}, source_reader: Optional=None, text_splitter: Optional[TextSplitter]=None, embedding_args: Optional[Dict]=None):
self.file_path = file_path
self.vector_store_config = (vector_store_config or {})
self.source_rea... |
def test_version_mocked(capsys, monkeypatch):
monkeypatch.setattr(ert.shared, '__version__', '1.0.3')
try:
ert_parser(None, ['--version'])
except SystemExit as e:
assert (e.code == 0)
(ert_version, _) = capsys.readouterr()
ert_version = ert_version.rstrip('\n')
assert (ert_versio... |
class Consumer(ThreadDelegateConsumer):
logger = logger
RebalanceListener: ClassVar[Type[ConsumerRebalanceListener]]
RebalanceListener = ConsumerRebalanceListener
consumer_stopped_errors: ClassVar[Tuple[(Type[BaseException], ...)]] = (ConsumerStoppedError,)
def _new_consumer_thread(self) -> Consumer... |
class DEITP(DeltaE):
NAME = 'itp'
def __init__(self, scalar: float=720) -> None:
self.scalar = scalar
def distance(self, color: 'Color', sample: 'Color', scalar: Optional[float]=None, **kwargs: Any) -> float:
if (scalar is None):
scalar = self.scalar
(i1, t1, p1) = color.... |
def sflaf(x, d, M=128, P=5, mu_L=0.2, mu_FL=0.5):
nIters = (min(len(x), len(d)) - M)
Q = (P * 2)
sk = np.arange(0, (Q * M), 2)
ck = np.arange(1, (Q * M), 2)
pk = np.tile(np.arange(P), M)
u = np.zeros(M)
w_L = np.zeros(M)
w_FL = np.zeros((Q * M))
e = np.zeros(nIters)
for n in rang... |
class Ml_model(models.Model):
name = models.TextField(blank=False, null=True)
model_name = models.TextField(blank=False, null=True)
username = models.TextField(blank=False, null=True)
description = models.TextField(blank=False, null=True)
version = models.FloatField(blank=False, null=True)
eval_... |
def generate_html_detail(dataframe_report):
all_detail = list(dataframe_report._features.values())
if (dataframe_report._target is not None):
all_detail.append(dataframe_report._target)
for feature in all_detail:
compare_dict = feature.get('compare')
if ((feature['type'] == FeatureTy... |
def test_fmpz_factor():
assert (flint.fmpz(6).gcd(flint.fmpz(9)) == 3)
assert (flint.fmpz(6).gcd(9) == 3)
assert raises((lambda : flint.fmpz(2).gcd('asd')), TypeError)
assert (flint.fmpz(6).lcm(flint.fmpz(9)) == 18)
assert (flint.fmpz(6).lcm(9) == 18)
assert raises((lambda : flint.fmpz(2).lcm('a... |
def charlen_tsets():
VocabularyMatcher.instance()._make_mask_from_char_length(1)
l1 = tset(charlen=1)
token_lengths = VocabularyMatcher.instance().token_lengths
assert (token_lengths is not None), 'VocabularyMatcher.instance().token_lengths is None even though it should be fully initialized.'
length... |
def get_best_model():
import re
pattern = 'model.(?P<epoch>\\d+)-(?P<val_loss>[0-9]*\\.?[0-9]*).hdf5'
p = re.compile(pattern)
files = [f for f in os.listdir('models/') if p.match(f)]
filename = None
if (len(files) > 0):
accs = [float(p.match(f).groups()[1]) for f in files]
best_i... |
class OptionSeriesTimelineDatalabelsStyle(Options):
def fontSize(self):
return self._config_get('0.8em')
def fontSize(self, num: float):
self._config(num, js_type=False)
def fontWeight(self):
return self._config_get('normal')
def fontWeight(self, text: str):
self._config(... |
class HistoricalMeta(ModelNormal):
allowed_values = {}
validations = {}
_property
def additional_properties_type():
return (bool, date, datetime, dict, float, int, list, str, none_type)
_nullable = False
_property
def openapi_types():
return {'to': (str,), '_from': (str,), 'b... |
class OptionSeriesBulletSonificationContexttracksMappingHighpass(Options):
def frequency(self) -> 'OptionSeriesBulletSonificationContexttracksMappingHighpassFrequency':
return self._config_sub_data('frequency', OptionSeriesBulletSonificationContexttracksMappingHighpassFrequency)
def resonance(self) -> '... |
()
def mock_ app, monkeypatch, request) -> ClientSession:
marker = request.node.get_closest_marker('
options = SessionMarker(**{**{'status_code': HTTPStatus.OK, 'text': b'', 'json': None, 'json_iterator': None, 'max_failures': None}, **((marker.kwargs or {}) if marker else {})})
def raise_for_status():
... |
def test_reference_creates_references_with_absolute_paths_containing_env_var(create_test_data, create_maya_env):
data = create_test_data
maya_env = create_maya_env
vers1 = Version(task=data['asset1'], created_by=data['user1'])
DBSession.add(vers1)
DBSession.commit()
vers2 = Version(task=data['as... |
def test_non_ip():
instance = HostBase()
for (eth_type, test_output) in ((1, {'tshark_ipx': 0, 'tshark_nonip': 1}), (33079, {'tshark_ipx': 1, 'tshark_nonip': 1}), (2048, {'tshark_ipx': 0, 'tshark_nonip': 0})):
test_data = {field: None for field in WS_FIELDS}
test_data.update({'eth.type': eth_typ... |
def replace_line(lines, line_no, block_comment, templates, copy_note, org, year_range, offset=None):
assert isinstance(lines, list)
assert (isinstance(line_no, int) and (1 <= line_no <= len(lines)))
assert isinstance(block_comment, bool)
assert isinstance(templates, dict)
assert isinstance(copy_note... |
class Token_Buffer(Token_Generator):
def __init__(self, lexer, cfg):
assert isinstance(lexer, MATLAB_Lexer)
assert isinstance(cfg, Config)
super().__init__(lexer.language, lexer.filename, lexer.blockname)
self.cfg = cfg
self.pos = 0
self.tokens = []
self.mh = ... |
def test_match_accept_with_inform_serialization():
msg = FipaMessage(message_id=1, dialogue_reference=(str(0), ''), target=0, performative=FipaMessage.Performative.MATCH_ACCEPT_W_INFORM, info={'address': 'dummy_address', 'signature': 'my_signature'})
msg.to = 'receiver'
envelope = Envelope(to=msg.to, sender... |
.skip_ci
('pyscf')
def test_pyscf_stocastic():
geom = geom_loader('lib:benzene_and_no.xyz')
def calc_getter(calc_number):
calc_kwargs = {'charge': (+ 1), 'mult': 1, 'pal': 2, 'basis': '321g', 'calc_number': calc_number}
calc = PySCF(**calc_kwargs)
return calc
stoc_kwargs = {'calc_get... |
def extractWordpressCom(item):
(vol, chp, frag, postfix) = extractVolChapterFragmentPostfix(item['title'])
if ((not (chp or vol)) or ('preview' in item['title'].lower())):
return None
tagmap = [('PRC', 'PRC', 'translated'), ('Loiterous', 'Loiterous', 'oel')]
for (tagname, name, tl_type) in tagma... |
def get_yara_error(rules_file: (str | bytes)) -> (Exception | None):
try:
if isinstance(rules_file, bytes):
rules_file = rules_file.decode()
yara.compile(source=rules_file)
return None
except (yara.Error, TypeError, UnicodeDecodeError) as error:
return error |
def initializePlugin(mobject):
mplugin = OpenMayaMPx.MFnPlugin(mobject, 'Erkan Ozgur Yilmaz', '1.0.2')
try:
mplugin.registerNode(kPluginNodeTypeName, cpocPluginId, nodeCreator, nodeInitializer)
except:
sys.stderr.write(('Failed to register node: %s' % kPluginNodeTypeName))
raise |
def snapshot_diff(cur_items, snapshot_file):
try:
prev_items = list(frontend_utils.get_pages(snapshot_file))
except pickle.UnpicklingError as e:
frontend_utils.echo_error(f'Error unpickling the data from {snapshot_file}: {e}')
return None
differences = []
for cur_item in cur_item... |
def test_simple_bkz_reduction(block_size=10):
for n in dimensions:
FPLLL.set_random_seed(n)
A = make_integer_matrix(n)
LLL.reduction(A)
B = copy(A)
BKZ.reduction(B, BKZ.Param(block_size=block_size))
C = copy(A)
SimpleBKZ(C)(block_size=block_size)
asser... |
class bad_instruction_error_msg(error_msg):
version = 3
type = 1
err_type = 3
def __init__(self, xid=None, code=None, data=None):
if (xid != None):
self.xid = xid
else:
self.xid = None
if (code != None):
self.code = code
else:
... |
def tail(records, tail):
logging.info((('Applying _tail generator: limiting results to top ' + tail) + ' records.'))
if (tail == '+0'):
for record in records:
(yield record)
elif ('+' in tail):
tail = (int(tail) - 1)
for record in itertools.islice(records, tail, None):
... |
class Environment():
def __init__(self, white: VectorLike, adapting_luminance: float, surround: float, discounting: float) -> None:
self.xyz_w = util.xy_to_xyz(white)
self.surround = surround
self.yn = adapting_luminance
self.d = discounting
self.ram = self.calc_ram()
... |
class vectorize2():
def __init__(self, pyfunc: Callable[(..., Any)], doc: Optional[str]=None):
self.func = pyfunc
self.__name__ = self.func.__name__
self.__doc__ = (self.func.__doc__ if (doc is None) else doc)
def _vector_apply(self, a: VectorLike, b: VectorLike, **kwargs: Any) -> Vector... |
def find_bounding_shape(virtual_size, available_shape):
assert (virtual_size <= product(available_shape))
free_size = virtual_size
free_dims = set(range(len(available_shape)))
bounding_shape = ([None] * len(available_shape))
while (len(free_dims) > 0):
guess = ceiling_root(free_size, len(fre... |
def _replace_reads(ir, fwd, c, sym, repl, only_replace_attrs=True):
cur_fwd = (lambda x: x)
c = fwd(c)
for rd in match_pattern(c, f'{repr(sym)}[_]', use_sym_id=True):
rd = cur_fwd(rd)
if (not (c_repl := repl(rd))):
continue
(ir, fwd_rd) = _replace_helper(rd, c_repl, only_... |
(eq=False, order=False, frozen=True, repr=True, hash=False)
class DefaultDetailsUpdateHandler(ChangeSetHandler):
db = attr.ib()
plugin_manager = attr.ib()
validators = attr.ib(factory=list)
def apply_changeset(self, model, changeset):
accumulate_errors((lambda v: v.validate(model, changeset)), s... |
class conn_tracking_nw_src_masked(oxm):
type_len = 127240
def __init__(self, value=None, value_mask=None):
if (value != None):
self.value = value
else:
self.value = 0
if (value_mask != None):
self.value_mask = value_mask
else:
self.... |
class OptionSeriesAreasplineSonificationDefaultinstrumentoptionsMappingPan(Options):
def mapFunction(self):
return self._config_get(None)
def mapFunction(self, value: Any):
self._config(value, js_type=False)
def mapTo(self):
return self._config_get(None)
def mapTo(self, text: str... |
class ConvProfilerCacheTestCase(unittest.TestCase):
def _test(self, first_dim, logger, test_name='conv2d'):
target = detect_target()
X = Tensor(shape=[first_dim, 28, 28, 128], dtype='float16', name='input_0', is_input=True)
W = Tensor(shape=[256, 3, 3, 128], dtype='float16', name='input_1', ... |
_pytree_node_class
class JaxStructure(AbstractJaxStructure, JaxObject):
geometry: JaxGeometryType = pd.Field(..., title='Geometry', description='Geometry of the structure, which is jax-compatible.', jax_field=True, discriminator=TYPE_TAG_STR)
medium: JaxMediumType = pd.Field(..., title='Medium', description='Me... |
class TestEnumV37(TestCase):
def test_repr_with_dataclass(self):
from dataclasses import dataclass, field
(repr=False)
class Foo():
__qualname__ = 'Foo'
a: int
def __repr__(self):
return 'ha hah!'
class Entries(Foo, Enum):
... |
class OptionPlotoptionsPieSonificationContexttracksMappingHighpassResonance(Options):
def mapFunction(self):
return self._config_get(None)
def mapFunction(self, value: Any):
self._config(value, js_type=False)
def mapTo(self):
return self._config_get(None)
def mapTo(self, text: st... |
class WorkflowDialog(QDialog):
closeButtonPressed = Signal()
def __init__(self, title, widget, parent=None):
QDialog.__init__(self, parent)
self.setWindowTitle(title)
self.setModal(True)
self.setWindowFlags((self.windowFlags() & (~ Qt.WindowContextHelpButtonHint)))
self.s... |
class Branch(GenericBranch[Condition]):
def __init__(self, condition: Condition, tags: Optional[Tuple[(Tag, ...)]]=None):
super(Branch, self).__init__(condition, tags=tags)
def __repr__(self) -> str:
return f'if {repr(self.condition)}'
def __str__(self) -> str:
return f'if({self.cond... |
def read_characters_from_files(files: List[str], frecent: List[str], use_additional: bool) -> Dict[(str, str)]:
all_characters: Dict[(str, List[str])] = {}
for character in frecent:
all_characters[character] = []
for file in __resolve_all_filenames(files, use_additional):
characters_from_fil... |
class MasterList(object):
def __init__(self, local_master_info, callback_update_remote, callback_update_local):
self._services_initialized = False
self.pubchanges = rospy.Publisher('~changes', MasterState, queue_size=10)
self.pubstats = rospy.Publisher('~linkstats', LinkStatesStamped, queue_... |
class DataQualityTable(Base):
__tablename__ = 'data_quality'
id = Column(Integer, primary_key=True)
timestamp = Column(Float)
drift_share = Column(Float)
ds_drift_metric_number_of_columns = Column(Integer)
number_of_drifted_columns = Column(Integer)
share_of_drifted_columns = Column(Float)
... |
def _get_whole_time_units(value, units='s', *, allowsmall=False):
assert isinstance(value, Decimal)
if (units not in TIME_UNITS):
raise ValueError(f'unsupported units {units!r}')
orig = f'{value} {units}'
while (int(value) != value):
(units, multiplier) = TIME_UNITS_DESCENDING[units]
... |
_invites_misc_routes.route('/role_invites/user', methods=['POST'])
def fetch_user():
token = request.json['data']['token']
try:
role_invite = RoleInvite.query.filter_by(hash=token).one()
except NoResultFound:
raise NotFoundError({'source': ''}, 'Role Invite Not Found')
else:
retu... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.