function stringlengths 11 56k | repo_name stringlengths 5 60 | features list |
|---|---|---|
def test_update_domain_empty(self):
self.assertRaises(exc.MissingDNSSettings, self.client.update_domain,
self.domain) | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_delete(self):
clt = self.client
mgr = clt._manager
dom = self.domain
mgr._async_call = Mock(return_value=({}, {}))
uri = "/domains/%s" % utils.get_id(dom)
clt.delete(dom)
mgr._async_call.assert_called_once_with(uri, method="DELETE",
error_... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_list_subdomains(self):
clt = self.client
mgr = clt._manager
dom = self.domain
resp_body = {'Something': 'here'}
clt.method_get = Mock(return_value=({}, resp_body))
uri = "/domains?name=%s&limit=5" % dom.name
clt.list_subdomains(dom, limit=5)
clt.m... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_search_records(self):
clt = self.client
mgr = clt._manager
dom = self.domain
typ = "A"
uri = "/domains/%s/records?type=%s" % (utils.get_id(dom), typ)
ret_body = {"records": [{"type": typ}]}
mgr.count = 0
def mock_get(uri):
if mgr.coun... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_find_record(self):
clt = self.client
mgr = clt._manager
dom = self.domain
typ = "A"
nm = utils.random_unicode()
data = "0.0.0.0"
ret_body = {"records": [{
"accountId": "728829",
"created": "2012-09-21T21:32:27.000+0000",
... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_find_record_not_unique(self):
clt = self.client
mgr = clt._manager
dom = self.domain
typ = "A"
nm = utils.random_unicode()
data = "0.0.0.0"
ret_body = {"records": [{
"accountId": "728829",
"created": "2012-09-21T21:32:27.00... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_get_record(self):
clt = self.client
mgr = clt._manager
dom = self.domain
nm = utils.random_unicode()
rec_id = utils.random_unicode()
rec_dict = {"id": rec_id, "name": nm}
mgr.api.method_get = Mock(return_value=(None, rec_dict))
ret = clt.get_recor... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_delete_record(self):
clt = self.client
mgr = clt._manager
dom = self.domain
rec = CloudDNSRecord(mgr, {"id": utils.random_unicode()})
mgr._async_call = Mock(return_value=({}, {}))
uri = "/domains/%s/records/%s" % (utils.get_id(dom), utils.get_id(rec))
clt... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_resolve_device_type_invalid(self):
clt = self.client
mgr = clt._manager
device = object()
self.assertRaises(exc.InvalidDeviceType, mgr._resolve_device_type,
device) | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_list_ptr_records(self):
clt = self.client
mgr = clt._manager
dvc = fakes.FakeDNSDevice()
href = "%s/%s" % (example_uri, dvc.id)
svc_name = "cloudServersOpenStack"
uri = "/rdns/%s?href=%s" % (svc_name, href)
mgr._get_ptr_details = Mock(return_value=(href, ... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_add_ptr_records(self):
clt = self.client
mgr = clt._manager
dvc = fakes.FakeDNSDevice()
href = "%s/%s" % (example_uri, dvc.id)
svc_name = "cloudServersOpenStack"
rec = {"foo": "bar"}
body = {"recordsList": {"records": [rec]},
"link": {"con... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_delete_ptr_records(self):
clt = self.client
mgr = clt._manager
dvc = fakes.FakeDNSDevice()
href = "%s/%s" % (example_uri, dvc.id)
svc_name = "cloudServersOpenStack"
ip_address = "0.0.0.0"
uri = "/rdns/%s?href=%s&ip=%s" % (svc_name, href, ip_address)
... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_get_rate_limits(self):
clt = self.client
limits = [{"uri": "fake1", "limit": 1},
{"uri": "fake2", "limit": 2}]
resp = {"limits": {"rate": limits}}
resp_limits = [{"uri": "fake1", "limits": 1},
{"uri": "fake2", "limits": 2}]
clt.method_get ... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_iter(self):
clt = self.client
mgr = clt._manager
res_iter = DomainResultsIterator(mgr)
ret = res_iter.__iter__()
self.assertTrue(ret is res_iter) | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_iter_items_first_fetch(self):
clt = self.client
mgr = clt._manager
fake_name = utils.random_unicode()
ret_body = {"domains": [{"name": fake_name}]}
clt.method_get = Mock(return_value=({}, ret_body))
res_iter = DomainResultsIterator(mgr)
ret = res_iter.nex... | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_iter_items_next_stop(self):
clt = self.client
mgr = clt._manager
res_iter = DomainResultsIterator(mgr)
res_iter.next_uri = None
self.assertRaises(StopIteration, res_iter.next) | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_record_iter(self):
clt = self.client
mgr = clt._manager
res_iter = RecordResultsIterator(mgr)
self.assertEqual(res_iter.paging_service, "record") | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def test_client_empty_get_body_error(self):
clt = self.client
self.assertRaises(exc.ServiceResponseFailure, clt.get_absolute_limits) | rackspace/pyrax | [
238,
218,
238,
75,
1348707957
] |
def __init__(
self,
*,
bucket_key: str,
bucket_name: Optional[str] = None,
wildcard_match: bool = False,
aws_conn_id: str = 'aws_default',
verify: Optional[Union[str, bool]] = None,
**kwargs, | apache/incubator-airflow | [
29418,
12032,
29418,
869,
1428948298
] |
def poke(self, context):
if self.bucket_name is None:
parsed_url = urlparse(self.bucket_key)
if parsed_url.netloc == '':
raise AirflowException('If key is a relative path from root, please provide a bucket_name')
self.bucket_name = parsed_url.netloc
... | apache/incubator-airflow | [
29418,
12032,
29418,
869,
1428948298
] |
def check_fn(self, data: List) -> bool:
return any(f.get('Size', 0) > 1048576 for f in data if isinstance(f, dict)) | apache/incubator-airflow | [
29418,
12032,
29418,
869,
1428948298
] |
def __init__(
self,
*,
check_fn: Optional[Callable[..., bool]] = None,
**kwargs, | apache/incubator-airflow | [
29418,
12032,
29418,
869,
1428948298
] |
def poke(self, context):
if super().poke(context=context) is False:
return False
s3_objects = self.get_files(s3_hook=self.get_hook())
if not s3_objects:
return False
check_fn = self.check_fn if self.check_fn_user is None else self.check_fn_user
return che... | apache/incubator-airflow | [
29418,
12032,
29418,
869,
1428948298
] |
def Enable(cls):
"""Enables this hook."""
cls._use = 1 | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def Disable(cls):
"""Disables this hook."""
cls._use = 0 | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def EnableMan(cls):
"""Enables copying the man pages."""
cls._use_man = 1 | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def DisableMan(cls):
"""Disables copying the man pages."""
cls._use_man = 0 | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def IsEnabled(cls):
"""Returns whether this hook is activated."""
return cls._use | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def IsManEnabled(cls):
"""Returns whether man pages will be copied."""
return cls._use_man | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def AddFile(cls, vFile):
"""Adds a required file to the hook to be copied into the initramfs."""
cls._files.append(vFile) | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def RemoveFile(cls, vFile):
"""Deletes a required file from the hook."""
try:
cls._files.remove(vFile)
except ValueError:
Tools.Fail('The file "' + vFile + '" was not found on the list!') | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def PrintFiles(cls):
"""Prints the required files in this hook."""
for file in cls.GetFiles():
print("File: " + file) | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def GetFiles(cls):
"""Returns the list of required files."""
return cls._files | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def GetOptionalFiles(cls):
"""Returns the list of optional files."""
return cls._optional_files | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def GetDirectories(cls):
"""Returns the list of required directories."""
return cls._directories | fearedbliss/bliss-initramfs | [
39,
23,
39,
2,
1407558349
] |
def main(verbose=False):
D = 10 ** 6
return 3 * int((D - 5) / 7.0) + 2 | dhermes/project-euler | [
11,
3,
11,
1,
1299471955
] |
def latest_jar():
global __JARS__
return __JARS__[-1] | cosminbasca/rdftools | [
5,
1,
5,
1,
1417629469
] |
def check_java(message=""):
if call(['java', '-version'], stderr=DEVNULL) != 0:
raise JavaNotFoundException(
'Java is not installed in the system path. {0}'.format(message)) | cosminbasca/rdftools | [
5,
1,
5,
1,
1417629469
] |
def run_lubm_generator(num_universities, index, generator_seed, ontology, output_path, xms=XMS, xmx=XMX):
run_tool("com.rdftools.LubmGenerator",
xms, xmx,
"--num_universities", num_universities,
"--start_index", index,
"--seed", generator_seed,
"--ont... | cosminbasca/rdftools | [
5,
1,
5,
1,
1417629469
] |
def run_jvmvoid_generator(source, dataset_id, output_path, xms=XMS, xmx=XMX):
run_tool("com.rdftools.VoIDGenerator",
xms, xmx,
"--source", source,
"--dataset_id", dataset_id,
"--output_path", output_path) | cosminbasca/rdftools | [
5,
1,
5,
1,
1417629469
] |
def ud_exception(w: str, tag: str) -> str:
if w == "การ" or w == "ความ":
return "NOUN"
return tag | PyThaiNLP/pythainlp | [
786,
237,
786,
35,
1466693846
] |
def post_process(
word_tags: List[Tuple[str, str]], to_ud: bool = False | PyThaiNLP/pythainlp | [
786,
237,
786,
35,
1466693846
] |
def setUpClass(cls):
super(NeutronResourcesTestJSON, cls).setUpClass()
if not CONF.orchestration.image_ref:
raise cls.skipException("No image available to test")
os = clients.Manager()
if not CONF.service_available.neutron:
raise cls.skipException("Neutron support... | Mirantis/tempest | [
2,
7,
2,
1,
1327963146
] |
def test_created_resources(self):
"""Verifies created neutron resources."""
resources = [('Network', 'OS::Neutron::Net'),
('Subnet', 'OS::Neutron::Subnet'),
('RouterInterface', 'OS::Neutron::RouterInterface'),
('Server', 'OS::Nova::Server')]... | Mirantis/tempest | [
2,
7,
2,
1,
1327963146
] |
def test_created_network(self):
"""Verifies created network."""
network_id = self.test_resources.get('Network')['physical_resource_id']
resp, body = self.network_client.show_network(network_id)
self.assertEqual('200', resp['status'])
network = body['network']
self.assertI... | Mirantis/tempest | [
2,
7,
2,
1,
1327963146
] |
def test_created_subnet(self):
"""Verifies created subnet."""
subnet_id = self.test_resources.get('Subnet')['physical_resource_id']
resp, body = self.network_client.show_subnet(subnet_id)
self.assertEqual('200', resp['status'])
subnet = body['subnet']
network_id = self.te... | Mirantis/tempest | [
2,
7,
2,
1,
1327963146
] |
def test_created_router(self):
"""Verifies created router."""
router_id = self.test_resources.get('Router')['physical_resource_id']
resp, body = self.network_client.show_router(router_id)
self.assertEqual('200', resp['status'])
router = body['router']
self.assertEqual('Ne... | Mirantis/tempest | [
2,
7,
2,
1,
1327963146
] |
def test_created_router_interface(self):
"""Verifies created router interface."""
router_id = self.test_resources.get('Router')['physical_resource_id']
network_id = self.test_resources.get('Network')['physical_resource_id']
subnet_id = self.test_resources.get('Subnet')['physical_resource... | Mirantis/tempest | [
2,
7,
2,
1,
1327963146
] |
def main(config="../../config.yaml", namespace=""):
# obtain config
if isinstance(config, str):
config = load_job_config(config)
lr_param = {
"name": "hetero_lr_0",
"penalty": "L2",
"optimizer": "rmsprop",
"tol": 0.0001,
"alpha": 0.01,
"max_iter": 30,... | FederatedAI/FATE | [
4887,
1449,
4887,
637,
1548325963
] |
def get_stats_for_region(region):
try:
session = boto3.Session(region_name=region)
num_instances = len(list(session.resource("ec2").instances.all()))
num_amis = len(list(session.resource("ec2").images.filter(Owners=["self"])))
num_vpcs = len(list(session.resource("ec2").vpcs.all()))
... | kislyuk/aegea | [
67,
17,
67,
23,
1456962813
] |
def __init__(self, name, data):
"""Initialize the Action object.
Actions do not have explicit partition attributes, the are
implied by the partition of the rule to which they belong.
"""
super(Action, self).__init__(name, partition=None)
# Actions are Only supported on ... | f5devcentral/f5-cccl | [
10,
45,
10,
23,
1491942422
] |
def __str__(self):
return str(self._data) | f5devcentral/f5-cccl | [
10,
45,
10,
23,
1491942422
] |
def setUp(self):
self.clock = task.Clock()
self.clock.spawnProcess = Mock()
treq_mock = create_autospec(treq)
response_mock = Mock()
response_mock.text.return_value = defer.succeed("")
treq_mock.request.return_value = defer.succeed(response_mock)
self.config = Con... | dpnova/pynerstat | [
4,
1,
4,
14,
1498828453
] |
def test_start_stop(self):
yield self.service.startService()
self.service.rig.start.assert_called_with()
yield self.service.stopService()
self.service.rig.stop.assert_called_with() | dpnova/pynerstat | [
4,
1,
4,
14,
1498828453
] |
def setUp(self):
self.config = Config("a", "b", "w", "p")
self.prot = MinerStatRemoteProtocol(self.config) | dpnova/pynerstat | [
4,
1,
4,
14,
1498828453
] |
def test_dlconf(self):
pass | dpnova/pynerstat | [
4,
1,
4,
14,
1498828453
] |
def test_algo_check(self):
pass | dpnova/pynerstat | [
4,
1,
4,
14,
1498828453
] |
def test_poll_remote(self):
pass | dpnova/pynerstat | [
4,
1,
4,
14,
1498828453
] |
def __init__(self, adapter_agent, config):
self.adapter_agent = adapter_agent
self.config = config
self.descriptor = Adapter(
id=self.name,
vendor='Voltha project',
version='0.1',
config=AdapterConfig(log_level=LogLevel.INFO)
)
self... | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def stop(self):
log.debug('stopping')
log.info('stopped') | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def device_types(self):
return DeviceTypes(items=self.supported_device_types) | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def change_master_state(self, master):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def reconcile_device(self, device):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def disable_device(self, device):
device.oper_status = OperStatus.UNKNOWN
self.adapter_agent.update_device(device) | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def reboot_device(self, device):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def get_image_download_status(self, device, request):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def activate_image_update(self, device, request):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def delete_device(self, device):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def update_pm_config(self, device, pm_configs):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def _simulate_device_activation(self, device):
# first we verify that we got parent reference and proxy info
assert device.parent_id
assert device.proxy_address.device_id
assert device.proxy_address.channel_id
# we pretend that we were able to contact the device and obtain
... | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def update_flows_incrementally(self, device, flow_changes, group_changes):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def receive_proxied_message(self, proxy_address, msg):
# just place incoming message to a list
self.incoming_messages.put((proxy_address, msg)) | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def update_interface(self, device, data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def receive_onu_detect_state(self, device_id, state):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def update_tcont(self, device, tcont_data, traffic_descriptor_data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def create_gemport(self, device, data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def remove_gemport(self, device, data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def update_multicast_gemport(self, device, data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def create_multicast_distribution_set(self, device, data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def remove_multicast_distribution_set(self, device, data):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def unsuppress_alarm(self, filter):
raise NotImplementedError() | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def _simulate_message_exchange(self, device):
# register for receiving async messages
self.adapter_agent.register_for_proxied_messages(device.proxy_address)
# reset incoming message queue
while self.incoming_messages.pending:
_ = yield self.incoming_messages.get()
... | opencord/voltha | [
73,
117,
73,
17,
1484694318
] |
def create_tasks(self, evidence):
"""Create task for Partition Enumeration.
Args:
evidence: List of evidence objects to process
Returns:
A list of tasks to schedule.
"""
tasks = [PartitionEnumerationTask() for _ in evidence]
return tasks | google/turbinia | [
625,
155,
625,
117,
1442335840
] |
def check_response(func):
"""
Decorator checking first REST response.
:return:
"""
def _decorator(self, *args, **kwargs):
try:
response = func(self, *args, **kwargs)
except ServerNotFoundError as e:
raise OperationRetry(
'Warning: {0}. '
... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def __init__(self, config, logger,
scope=constants.COMPUTE_SCOPE,
discovery=constants.COMPUTE_DISCOVERY,
api_version=constants.API_V1):
"""
GoogleCloudApi class constructor.
Create API discovery object that will be making GCP REST API calls.
... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def discovery(self):
"""
Lazily load the discovery so we don't make API calls during __init__
"""
if hasattr(self, '_discovery'):
return self._discovery
self._discovery = self.create_discovery(self.__discovery, self.scope,
... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def create_discovery(self, discovery, scope, api_version):
"""
Create Google Cloud API discovery object and perform authentication.
:param discovery: name of the API discovery to be created
:param scope: scope the API discovery will have
:param api_version: version of the API
... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def __init__(self, config, logger, name,
additional_settings=None,
scope=constants.COMPUTE_SCOPE,
discovery=constants.COMPUTE_DISCOVERY,
api_version=constants.API_V1):
"""
GoogleCloudPlatform class constructor.
Create API discov... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def get_common_instance_metadata(self):
"""
Get project's common instance metadata.
:return: CommonInstanceMetadata list extracted from REST response get
project metadata.
"""
self.logger.info(
'Get commonInstanceMetadata for project {0}'.format(self.project)... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def ZONES(self):
if not hasattr(self, '_ZONES'):
zones = {}
request = self.discovery.zones().list(project=self.project)
while request is not None:
response = request.execute()
for zone in response['items']:
zones[zone['name... | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def __init__(self, message):
super(GCPError, self).__init__(message) | cloudify-cosmo/cloudify-gcp-plugin | [
6,
13,
6,
6,
1428599218
] |
def __init__(self, device,
cpu=GUESTSHELL_CPU, memory=GUESTSHELL_MEMORY,
disk=GUESTSHELL_DISK, log=None):
self.device = device
self.guestshell = partial(self.device.api.exec_opcmd, msg_type='cli_show_ascii')
self.cli = self.device.api.exec_opcmd
self.lo... | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def state(self):
cmd = 'show virtual-service detail name guestshell+'
try:
got = self.cli(cmd)
except CommandError:
# means there is no guestshell
self.exists = False
self._state = 'None'
return self._state
try:
sel... | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def size(self):
self._get_sz_info()
return self.sz_has | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def setup(self):
self.log.info("/START(guestshell): setup")
state = self.state
self.log.info("/INFO(guestshell): current state: %s" % state)
if 'Activated' == state:
self._get_sz_info()
if self.sz_need != self.sz_has:
self.log.info("/INFO(guestsh... | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def enable(self):
self.guestshell('guestshell enable')
self._wait_state('Activated') | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def disable(self):
self.guestshell('guestshell disable')
self._wait_state('Deactivated') | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def resize_memory(self, memory):
value = min(memory, self.sz_max['memory'])
self.guestshell('guestshell resize memory {}'.format(value)) | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def resize(self):
self.resize_cpu(self.sz_need.cpu)
self.resize_memory(self.sz_need.memory)
self.resize_disk(self.sz_need.disk) | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
def sudoers(self, enable):
"""
This method is used to enable/disable bash sudo commands running
through the guestshell virtual service. By default sudo access
is prevented due to the setting in the 'sudoers' file. Therefore
the setting must be disabled in the file to enable sud... | Apstra/aeon-venos | [
2,
4,
2,
1,
1463407794
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.