function
stringlengths
11
56k
repo_name
stringlengths
5
60
features
list
def test_update_domain_empty(self): self.assertRaises(exc.MissingDNSSettings, self.client.update_domain, self.domain)
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_delete(self): clt = self.client mgr = clt._manager dom = self.domain mgr._async_call = Mock(return_value=({}, {})) uri = "/domains/%s" % utils.get_id(dom) clt.delete(dom) mgr._async_call.assert_called_once_with(uri, method="DELETE", error_...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_list_subdomains(self): clt = self.client mgr = clt._manager dom = self.domain resp_body = {'Something': 'here'} clt.method_get = Mock(return_value=({}, resp_body)) uri = "/domains?name=%s&limit=5" % dom.name clt.list_subdomains(dom, limit=5) clt.m...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_search_records(self): clt = self.client mgr = clt._manager dom = self.domain typ = "A" uri = "/domains/%s/records?type=%s" % (utils.get_id(dom), typ) ret_body = {"records": [{"type": typ}]} mgr.count = 0 def mock_get(uri): if mgr.coun...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_find_record(self): clt = self.client mgr = clt._manager dom = self.domain typ = "A" nm = utils.random_unicode() data = "0.0.0.0" ret_body = {"records": [{ "accountId": "728829", "created": "2012-09-21T21:32:27.000+0000", ...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_find_record_not_unique(self): clt = self.client mgr = clt._manager dom = self.domain typ = "A" nm = utils.random_unicode() data = "0.0.0.0" ret_body = {"records": [{ "accountId": "728829", "created": "2012-09-21T21:32:27.00...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_get_record(self): clt = self.client mgr = clt._manager dom = self.domain nm = utils.random_unicode() rec_id = utils.random_unicode() rec_dict = {"id": rec_id, "name": nm} mgr.api.method_get = Mock(return_value=(None, rec_dict)) ret = clt.get_recor...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_delete_record(self): clt = self.client mgr = clt._manager dom = self.domain rec = CloudDNSRecord(mgr, {"id": utils.random_unicode()}) mgr._async_call = Mock(return_value=({}, {})) uri = "/domains/%s/records/%s" % (utils.get_id(dom), utils.get_id(rec)) clt...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_resolve_device_type_invalid(self): clt = self.client mgr = clt._manager device = object() self.assertRaises(exc.InvalidDeviceType, mgr._resolve_device_type, device)
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_list_ptr_records(self): clt = self.client mgr = clt._manager dvc = fakes.FakeDNSDevice() href = "%s/%s" % (example_uri, dvc.id) svc_name = "cloudServersOpenStack" uri = "/rdns/%s?href=%s" % (svc_name, href) mgr._get_ptr_details = Mock(return_value=(href, ...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_add_ptr_records(self): clt = self.client mgr = clt._manager dvc = fakes.FakeDNSDevice() href = "%s/%s" % (example_uri, dvc.id) svc_name = "cloudServersOpenStack" rec = {"foo": "bar"} body = {"recordsList": {"records": [rec]}, "link": {"con...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_delete_ptr_records(self): clt = self.client mgr = clt._manager dvc = fakes.FakeDNSDevice() href = "%s/%s" % (example_uri, dvc.id) svc_name = "cloudServersOpenStack" ip_address = "0.0.0.0" uri = "/rdns/%s?href=%s&ip=%s" % (svc_name, href, ip_address) ...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_get_rate_limits(self): clt = self.client limits = [{"uri": "fake1", "limit": 1}, {"uri": "fake2", "limit": 2}] resp = {"limits": {"rate": limits}} resp_limits = [{"uri": "fake1", "limits": 1}, {"uri": "fake2", "limits": 2}] clt.method_get ...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_iter(self): clt = self.client mgr = clt._manager res_iter = DomainResultsIterator(mgr) ret = res_iter.__iter__() self.assertTrue(ret is res_iter)
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_iter_items_first_fetch(self): clt = self.client mgr = clt._manager fake_name = utils.random_unicode() ret_body = {"domains": [{"name": fake_name}]} clt.method_get = Mock(return_value=({}, ret_body)) res_iter = DomainResultsIterator(mgr) ret = res_iter.nex...
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_iter_items_next_stop(self): clt = self.client mgr = clt._manager res_iter = DomainResultsIterator(mgr) res_iter.next_uri = None self.assertRaises(StopIteration, res_iter.next)
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_record_iter(self): clt = self.client mgr = clt._manager res_iter = RecordResultsIterator(mgr) self.assertEqual(res_iter.paging_service, "record")
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def test_client_empty_get_body_error(self): clt = self.client self.assertRaises(exc.ServiceResponseFailure, clt.get_absolute_limits)
rackspace/pyrax
[ 238, 218, 238, 75, 1348707957 ]
def __init__( self, *, bucket_key: str, bucket_name: Optional[str] = None, wildcard_match: bool = False, aws_conn_id: str = 'aws_default', verify: Optional[Union[str, bool]] = None, **kwargs,
apache/incubator-airflow
[ 29418, 12032, 29418, 869, 1428948298 ]
def poke(self, context): if self.bucket_name is None: parsed_url = urlparse(self.bucket_key) if parsed_url.netloc == '': raise AirflowException('If key is a relative path from root, please provide a bucket_name') self.bucket_name = parsed_url.netloc ...
apache/incubator-airflow
[ 29418, 12032, 29418, 869, 1428948298 ]
def check_fn(self, data: List) -> bool: return any(f.get('Size', 0) > 1048576 for f in data if isinstance(f, dict))
apache/incubator-airflow
[ 29418, 12032, 29418, 869, 1428948298 ]
def __init__( self, *, check_fn: Optional[Callable[..., bool]] = None, **kwargs,
apache/incubator-airflow
[ 29418, 12032, 29418, 869, 1428948298 ]
def poke(self, context): if super().poke(context=context) is False: return False s3_objects = self.get_files(s3_hook=self.get_hook()) if not s3_objects: return False check_fn = self.check_fn if self.check_fn_user is None else self.check_fn_user return che...
apache/incubator-airflow
[ 29418, 12032, 29418, 869, 1428948298 ]
def Enable(cls): """Enables this hook.""" cls._use = 1
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def Disable(cls): """Disables this hook.""" cls._use = 0
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def EnableMan(cls): """Enables copying the man pages.""" cls._use_man = 1
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def DisableMan(cls): """Disables copying the man pages.""" cls._use_man = 0
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def IsEnabled(cls): """Returns whether this hook is activated.""" return cls._use
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def IsManEnabled(cls): """Returns whether man pages will be copied.""" return cls._use_man
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def AddFile(cls, vFile): """Adds a required file to the hook to be copied into the initramfs.""" cls._files.append(vFile)
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def RemoveFile(cls, vFile): """Deletes a required file from the hook.""" try: cls._files.remove(vFile) except ValueError: Tools.Fail('The file "' + vFile + '" was not found on the list!')
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def PrintFiles(cls): """Prints the required files in this hook.""" for file in cls.GetFiles(): print("File: " + file)
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def GetFiles(cls): """Returns the list of required files.""" return cls._files
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def GetOptionalFiles(cls): """Returns the list of optional files.""" return cls._optional_files
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def GetDirectories(cls): """Returns the list of required directories.""" return cls._directories
fearedbliss/bliss-initramfs
[ 39, 23, 39, 2, 1407558349 ]
def main(verbose=False): D = 10 ** 6 return 3 * int((D - 5) / 7.0) + 2
dhermes/project-euler
[ 11, 3, 11, 1, 1299471955 ]
def latest_jar(): global __JARS__ return __JARS__[-1]
cosminbasca/rdftools
[ 5, 1, 5, 1, 1417629469 ]
def check_java(message=""): if call(['java', '-version'], stderr=DEVNULL) != 0: raise JavaNotFoundException( 'Java is not installed in the system path. {0}'.format(message))
cosminbasca/rdftools
[ 5, 1, 5, 1, 1417629469 ]
def run_lubm_generator(num_universities, index, generator_seed, ontology, output_path, xms=XMS, xmx=XMX): run_tool("com.rdftools.LubmGenerator", xms, xmx, "--num_universities", num_universities, "--start_index", index, "--seed", generator_seed, "--ont...
cosminbasca/rdftools
[ 5, 1, 5, 1, 1417629469 ]
def run_jvmvoid_generator(source, dataset_id, output_path, xms=XMS, xmx=XMX): run_tool("com.rdftools.VoIDGenerator", xms, xmx, "--source", source, "--dataset_id", dataset_id, "--output_path", output_path)
cosminbasca/rdftools
[ 5, 1, 5, 1, 1417629469 ]
def ud_exception(w: str, tag: str) -> str: if w == "การ" or w == "ความ": return "NOUN" return tag
PyThaiNLP/pythainlp
[ 786, 237, 786, 35, 1466693846 ]
def post_process( word_tags: List[Tuple[str, str]], to_ud: bool = False
PyThaiNLP/pythainlp
[ 786, 237, 786, 35, 1466693846 ]
def setUpClass(cls): super(NeutronResourcesTestJSON, cls).setUpClass() if not CONF.orchestration.image_ref: raise cls.skipException("No image available to test") os = clients.Manager() if not CONF.service_available.neutron: raise cls.skipException("Neutron support...
Mirantis/tempest
[ 2, 7, 2, 1, 1327963146 ]
def test_created_resources(self): """Verifies created neutron resources.""" resources = [('Network', 'OS::Neutron::Net'), ('Subnet', 'OS::Neutron::Subnet'), ('RouterInterface', 'OS::Neutron::RouterInterface'), ('Server', 'OS::Nova::Server')]...
Mirantis/tempest
[ 2, 7, 2, 1, 1327963146 ]
def test_created_network(self): """Verifies created network.""" network_id = self.test_resources.get('Network')['physical_resource_id'] resp, body = self.network_client.show_network(network_id) self.assertEqual('200', resp['status']) network = body['network'] self.assertI...
Mirantis/tempest
[ 2, 7, 2, 1, 1327963146 ]
def test_created_subnet(self): """Verifies created subnet.""" subnet_id = self.test_resources.get('Subnet')['physical_resource_id'] resp, body = self.network_client.show_subnet(subnet_id) self.assertEqual('200', resp['status']) subnet = body['subnet'] network_id = self.te...
Mirantis/tempest
[ 2, 7, 2, 1, 1327963146 ]
def test_created_router(self): """Verifies created router.""" router_id = self.test_resources.get('Router')['physical_resource_id'] resp, body = self.network_client.show_router(router_id) self.assertEqual('200', resp['status']) router = body['router'] self.assertEqual('Ne...
Mirantis/tempest
[ 2, 7, 2, 1, 1327963146 ]
def test_created_router_interface(self): """Verifies created router interface.""" router_id = self.test_resources.get('Router')['physical_resource_id'] network_id = self.test_resources.get('Network')['physical_resource_id'] subnet_id = self.test_resources.get('Subnet')['physical_resource...
Mirantis/tempest
[ 2, 7, 2, 1, 1327963146 ]
def main(config="../../config.yaml", namespace=""): # obtain config if isinstance(config, str): config = load_job_config(config) lr_param = { "name": "hetero_lr_0", "penalty": "L2", "optimizer": "rmsprop", "tol": 0.0001, "alpha": 0.01, "max_iter": 30,...
FederatedAI/FATE
[ 4887, 1449, 4887, 637, 1548325963 ]
def get_stats_for_region(region): try: session = boto3.Session(region_name=region) num_instances = len(list(session.resource("ec2").instances.all())) num_amis = len(list(session.resource("ec2").images.filter(Owners=["self"]))) num_vpcs = len(list(session.resource("ec2").vpcs.all())) ...
kislyuk/aegea
[ 67, 17, 67, 23, 1456962813 ]
def __init__(self, name, data): """Initialize the Action object. Actions do not have explicit partition attributes, the are implied by the partition of the rule to which they belong. """ super(Action, self).__init__(name, partition=None) # Actions are Only supported on ...
f5devcentral/f5-cccl
[ 10, 45, 10, 23, 1491942422 ]
def __str__(self): return str(self._data)
f5devcentral/f5-cccl
[ 10, 45, 10, 23, 1491942422 ]
def setUp(self): self.clock = task.Clock() self.clock.spawnProcess = Mock() treq_mock = create_autospec(treq) response_mock = Mock() response_mock.text.return_value = defer.succeed("") treq_mock.request.return_value = defer.succeed(response_mock) self.config = Con...
dpnova/pynerstat
[ 4, 1, 4, 14, 1498828453 ]
def test_start_stop(self): yield self.service.startService() self.service.rig.start.assert_called_with() yield self.service.stopService() self.service.rig.stop.assert_called_with()
dpnova/pynerstat
[ 4, 1, 4, 14, 1498828453 ]
def setUp(self): self.config = Config("a", "b", "w", "p") self.prot = MinerStatRemoteProtocol(self.config)
dpnova/pynerstat
[ 4, 1, 4, 14, 1498828453 ]
def test_dlconf(self): pass
dpnova/pynerstat
[ 4, 1, 4, 14, 1498828453 ]
def test_algo_check(self): pass
dpnova/pynerstat
[ 4, 1, 4, 14, 1498828453 ]
def test_poll_remote(self): pass
dpnova/pynerstat
[ 4, 1, 4, 14, 1498828453 ]
def __init__(self, adapter_agent, config): self.adapter_agent = adapter_agent self.config = config self.descriptor = Adapter( id=self.name, vendor='Voltha project', version='0.1', config=AdapterConfig(log_level=LogLevel.INFO) ) self...
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def stop(self): log.debug('stopping') log.info('stopped')
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def device_types(self): return DeviceTypes(items=self.supported_device_types)
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def change_master_state(self, master): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def reconcile_device(self, device): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def disable_device(self, device): device.oper_status = OperStatus.UNKNOWN self.adapter_agent.update_device(device)
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def reboot_device(self, device): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def get_image_download_status(self, device, request): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def activate_image_update(self, device, request): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def delete_device(self, device): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def update_pm_config(self, device, pm_configs): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def _simulate_device_activation(self, device): # first we verify that we got parent reference and proxy info assert device.parent_id assert device.proxy_address.device_id assert device.proxy_address.channel_id # we pretend that we were able to contact the device and obtain ...
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def update_flows_incrementally(self, device, flow_changes, group_changes): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def receive_proxied_message(self, proxy_address, msg): # just place incoming message to a list self.incoming_messages.put((proxy_address, msg))
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def update_interface(self, device, data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def receive_onu_detect_state(self, device_id, state): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def update_tcont(self, device, tcont_data, traffic_descriptor_data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def create_gemport(self, device, data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def remove_gemport(self, device, data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def update_multicast_gemport(self, device, data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def create_multicast_distribution_set(self, device, data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def remove_multicast_distribution_set(self, device, data): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def unsuppress_alarm(self, filter): raise NotImplementedError()
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def _simulate_message_exchange(self, device): # register for receiving async messages self.adapter_agent.register_for_proxied_messages(device.proxy_address) # reset incoming message queue while self.incoming_messages.pending: _ = yield self.incoming_messages.get() ...
opencord/voltha
[ 73, 117, 73, 17, 1484694318 ]
def create_tasks(self, evidence): """Create task for Partition Enumeration. Args: evidence: List of evidence objects to process Returns: A list of tasks to schedule. """ tasks = [PartitionEnumerationTask() for _ in evidence] return tasks
google/turbinia
[ 625, 155, 625, 117, 1442335840 ]
def check_response(func): """ Decorator checking first REST response. :return: """ def _decorator(self, *args, **kwargs): try: response = func(self, *args, **kwargs) except ServerNotFoundError as e: raise OperationRetry( 'Warning: {0}. ' ...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def __init__(self, config, logger, scope=constants.COMPUTE_SCOPE, discovery=constants.COMPUTE_DISCOVERY, api_version=constants.API_V1): """ GoogleCloudApi class constructor. Create API discovery object that will be making GCP REST API calls. ...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def discovery(self): """ Lazily load the discovery so we don't make API calls during __init__ """ if hasattr(self, '_discovery'): return self._discovery self._discovery = self.create_discovery(self.__discovery, self.scope, ...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def create_discovery(self, discovery, scope, api_version): """ Create Google Cloud API discovery object and perform authentication. :param discovery: name of the API discovery to be created :param scope: scope the API discovery will have :param api_version: version of the API ...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def __init__(self, config, logger, name, additional_settings=None, scope=constants.COMPUTE_SCOPE, discovery=constants.COMPUTE_DISCOVERY, api_version=constants.API_V1): """ GoogleCloudPlatform class constructor. Create API discov...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def get_common_instance_metadata(self): """ Get project's common instance metadata. :return: CommonInstanceMetadata list extracted from REST response get project metadata. """ self.logger.info( 'Get commonInstanceMetadata for project {0}'.format(self.project)...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def ZONES(self): if not hasattr(self, '_ZONES'): zones = {} request = self.discovery.zones().list(project=self.project) while request is not None: response = request.execute() for zone in response['items']: zones[zone['name...
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def __init__(self, message): super(GCPError, self).__init__(message)
cloudify-cosmo/cloudify-gcp-plugin
[ 6, 13, 6, 6, 1428599218 ]
def __init__(self, device, cpu=GUESTSHELL_CPU, memory=GUESTSHELL_MEMORY, disk=GUESTSHELL_DISK, log=None): self.device = device self.guestshell = partial(self.device.api.exec_opcmd, msg_type='cli_show_ascii') self.cli = self.device.api.exec_opcmd self.lo...
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def state(self): cmd = 'show virtual-service detail name guestshell+' try: got = self.cli(cmd) except CommandError: # means there is no guestshell self.exists = False self._state = 'None' return self._state try: sel...
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def size(self): self._get_sz_info() return self.sz_has
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def setup(self): self.log.info("/START(guestshell): setup") state = self.state self.log.info("/INFO(guestshell): current state: %s" % state) if 'Activated' == state: self._get_sz_info() if self.sz_need != self.sz_has: self.log.info("/INFO(guestsh...
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def enable(self): self.guestshell('guestshell enable') self._wait_state('Activated')
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def disable(self): self.guestshell('guestshell disable') self._wait_state('Deactivated')
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def resize_memory(self, memory): value = min(memory, self.sz_max['memory']) self.guestshell('guestshell resize memory {}'.format(value))
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def resize(self): self.resize_cpu(self.sz_need.cpu) self.resize_memory(self.sz_need.memory) self.resize_disk(self.sz_need.disk)
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]
def sudoers(self, enable): """ This method is used to enable/disable bash sudo commands running through the guestshell virtual service. By default sudo access is prevented due to the setting in the 'sudoers' file. Therefore the setting must be disabled in the file to enable sud...
Apstra/aeon-venos
[ 2, 4, 2, 1, 1463407794 ]