source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x() ->List[ServiceState]:
keys = DB.get_keys('states*')
LOG.debug('Loading list of known services.')
services = []
for key in keys:
values = key.split(':')
if len(values) == 4:
services.append(ServiceState(*values[1:4]))
return services
| get_service_state_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x() ->List[ServiceState]:\n keys = DB.get_keys('states*')\n LOG.debug('Loading list of known services.')\n services = []\n for key in keys:\n ... |
def x() ->List[tuple]:
keys = DB.get_keys('states*')
services = []
for key in keys:
values = key.split(':')
if len(values) == 4:
services.append(':'.join(values[1:]))
return services
| get_service_id_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x() ->List[tuple]:\n keys = DB.get_keys('states*')\n services = []\n for key in keys:\n values = key.split(':')\n if len(values) == 4:\n ... |
def x():
LOG.debug('GET Processing Block list')
_url = get_root_url()
block_ids = sorted(DB.get_processing_block_ids())
LOG.debug('Processing Block IDs: %s', block_ids)
response = dict(num_processing_blocks=len(block_ids), processing_blocks
=list())
for block in DB.get_block_details(block_ids):
block_id = block['id']
LOG.debug('Creating PB summary for %s', block_id)
block['links'] = dict(detail='{}/processing-block/{}'.format(_url,
block_id), scheduling_block='{}/scheduling-block/{}'.format(
_url, block_id.split(':')[0]))
response['processing_blocks'].append(block)
response['links'] = {'self': '{}'.format(request.url), 'home': '{}'.
format(_url)}
return response, HTTPStatus.OK
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n LOG.debug('GET Processing Block list')\n _url = get_root_url()\n block_ids = sorted(DB.get_processing_block_ids())\n LOG.debug('Processing Blo... |
def x(self) ->dict:
self._check_object_exists()
config_dict = DB.get_hash_dict(self.key)
for _, value in config_dict.items():
for char in ['[', '{']:
if char in value:
value = ast.literal_eval(value)
return config_dict
| config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->dict:\n self._check_object_exists()\n config_dict = DB.get_hash_dict(self.key)\n for _, value in config_dict.items():\n for char in ['[... |
def x(self, property_key: str) ->str:
self._check_object_exists()
return DB.get_hash_value(self.key, property_key)
| get_property | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, property_key: str) ->str:\n self._check_object_exists()\n return DB.get_hash_value(self.key, property_key)\n",
"role": "user"
},
{
"con... |
def x(self, value):
self._check_object_exists()
DB.set_hash_value(self.key, 'status', value)
self.publish('status_changed', event_data=dict(status=value))
| set_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n self._check_object_exists()\n DB.set_hash_value(self.key, 'status', value)\n self.publish('status_changed', event_data=dict(status=val... |
def x(self, event_type: str, event_data: dict=None):
import inspect
import os.path
_stack = inspect.stack()
_origin = os.path.basename(_stack[3][1]) + '::' + _stack[3][3
] + '::L{}'.format(_stack[3][2])
publish(event_type=event_type, event_data=event_data, object_type=self.
_type, object_id=self._id, object_key=self._key, origin=_origin)
| publish | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_type: str, event_data: dict=None):\n import inspect\n import os.path\n _stack = inspect.stack()\n _origin = os.path.basename(_stack[3][... |
def x(self) ->List[Event]:
LOG.debug('Getting events for %s', self.key)
return get_events(self.key)
| get_events | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->List[Event]:\n LOG.debug('Getting events for %s', self.key)\n return get_events(self.key)\n",
"role": "user"
},
{
"content": "get_eve... |
def x(self):
if not DB.get_keys(self.key):
raise KeyError("Object with key '{}' not exist".format(self.key))
| check_object_exists | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not DB.get_keys(self.key):\n raise KeyError(\"Object with key '{}' not exist\".format(self.key))\n",
"role": "user"
},
{
"con... |
def x(subsystem: str, name: str, version: str) ->str:
return '{}:{}:{}'.format(subsystem, name, version)
| get_service_state_object_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subsystem: str, name: str, version: str) ->str:\n return '{}:{}:{}'.format(subsystem, name, version)\n",
"role": "user"
},
{
"content": "get_ser... |
def x(self, compose_str: str) ->list:
if not self._manager:
raise RuntimeError('Services can only be run on swarm manager nodes')
services_ids = []
try:
service_config = yaml.load(compose_str)
service_list = copy.deepcopy(service_config)
service_config.pop('version')
service_config.pop('services')
for service_name in service_list['services']:
service_exist = self._client.services.list(filters={'name':
service_name})
if not service_exist:
service_config['name'] = service_name
service_spec = self._parse_services(service_config,
service_name, service_list)
created_service = self._client.services.create(**service_spec)
service_id = created_service.short_id
LOG.debug('Service created: %s', service_id)
services_ids.append(service_id)
else:
LOG.debug('Services already exists')
except yaml.YAMLError as exc:
print(exc)
return services_ids
| create_services | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, compose_str: str) ->list:\n if not self._manager:\n raise RuntimeError('Services can only be run on swarm manager nodes')\n services_ids = [... |
def x(self, volume_name: str, driver_spec: str=None):
if driver_spec:
driver = driver_spec
else:
driver = 'local'
if not self._manager:
raise RuntimeError(
'Services can only be deleted on swarm manager nodes')
self._client.volumes.create(name=volume_name, driver=driver)
| create_volume | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, volume_name: str, driver_spec: str=None):\n if driver_spec:\n driver = driver_spec\n else:\n driver = 'local'\n if not self._manag... |
def x(self, service: str):
if not self._manager:
raise RuntimeError(
'Services can only be deleted on swarm manager nodes')
self._api_client.remove_service(service)
| delete_service | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service: str):\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n self._api... |
def x(self):
if not self._manager:
raise RuntimeError(
'Services can only be deleted on swarm manager nodes')
service_list = self.get_service_list()
for services in service_list:
self._api_client.remove_service(services)
| delete_all_services | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n service_list = self.get... |
def x(self, volume_name: str):
if not self._manager:
raise RuntimeError('Volumes can only be deleted on swarm manager nodes'
)
self._api_client.remove_volume(volume_name)
| delete_volume | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, volume_name: str):\n if not self._manager:\n raise RuntimeError('Volumes can only be deleted on swarm manager nodes'\n )\n self._... |
def x(self):
if not self._manager:
raise RuntimeError('Volumes can only be deleted on swarm manager nodes'
)
volume_list = self.get_volume_list()
for volumes in volume_list:
self._api_client.remove_volume(volumes, force=True)
| delete_all_volumes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._manager:\n raise RuntimeError('Volumes can only be deleted on swarm manager nodes'\n )\n volume_list = self.get_v... |
def x(self) ->list:
services = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services.')
service_list = self._client.services.list()
for s_list in service_list:
services.append(s_list.short_id)
return services
| get_service_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n services = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the service... |
def x(self, service_id: str) ->str:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services details.'
)
service = self._client.services.get(service_id)
return service.name
| get_service_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services d... |
def x(self, service_id: str) ->dict:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services details.'
)
service = self._client.services.get(service_id)
return service.attrs
| get_service_details | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->dict:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services ... |
def x(self, service_id: str) ->str:
service = self._client.services.get(service_id)
for service_task in service.tasks():
service_state = service_task['DesiredState']
return service_state
| get_service_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n service = self._client.services.get(service_id)\n for service_task in service.tasks():\n service_state = service_t... |
def x(self) ->list:
nodes = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the nodes.')
node_list = self._client.nodes.list()
for n_list in node_list:
nodes.append(n_list.id)
return nodes
| get_node_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n nodes = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the nodes.')\n... |
def x(self, node_id: list) ->dict:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve node details.')
node = self._client.nodes.get(node_id)
return node.attrs
| get_node_details | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node_id: list) ->dict:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve node details.')\n ... |
def x(self) ->list:
containers = []
containers_list = self._client.containers.list()
for c_list in containers_list:
containers.append(c_list.short_id)
return containers
| get_container_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n containers = []\n containers_list = self._client.containers.list()\n for c_list in containers_list:\n containers.append(c_list.... |
def x(self, container_id_or_name: str) ->dict:
container = self._client.containers.get(container_id_or_name)
return container.attrs
| get_container_details | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, container_id_or_name: str) ->dict:\n container = self._client.containers.get(container_id_or_name)\n return container.attrs\n",
"role": "user"
... |
def x(self) ->list:
volumes = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services.')
volume_list = self._client.volumes.list()
for v_list in volume_list:
volumes.append(v_list.name)
return volumes
| get_volume_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n volumes = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services... |
def x(self, volume_name: str) ->dict:
if volume_name not in self.volumes:
raise RuntimeError('No such volume found: ', volume_name)
volume = self._client.volumes.get(volume_name)
return volume.attrs
| get_volume_details | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, volume_name: str) ->dict:\n if volume_name not in self.volumes:\n raise RuntimeError('No such volume found: ', volume_name)\n volume = self.... |
def x(self, service_id: str) ->str:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve replication level of the service'
)
service_details = self.get_service_details(service_id)
actual_replica = service_details['Spec']['Mode']['Replicated']['Replicas']
return actual_replica
| get_actual_replica | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve replication level ... |
def x(self, service_id: str) ->str:
replicas = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve replication level of the service'
)
service_tasks = self._client.services.get(service_id).tasks()
for task in service_tasks:
if task['Status']['State'] == 'running':
replicas.append(task)
return len(replicas)
| get_replicas | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n replicas = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve... |
def x(self, node_name: str, labels: dict):
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can update node details.')
node_spec = {'Availability': 'active', 'Name': node_name, 'Role':
'manager', 'Labels': labels}
node = self._client.nodes.get(node_name)
node.update(node_spec)
| update_labels | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node_name: str, labels: dict):\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can update node details.... |
def x(self, service_config: dict, service_name: str, service_list: dict
) ->dict:
for key, value in service_list['services'][service_name].items():
service_config[key] = value
if 'command' in key:
key = 'args'
service_config['args'] = value
service_config.pop('command')
if 'ports' in key:
endpoint_spec = self._parse_ports(value)
service_config['endpoint_spec'] = endpoint_spec
service_config.pop('ports')
if 'volumes' in key:
volume_spec = self._parse_volumes(value)
service_config['mounts'] = volume_spec
service_config.pop('volumes')
if 'deploy' in key:
self._parse_deploy(value, service_config)
service_config.pop('deploy')
if 'networks' in key:
network_spec = self._parse_networks(service_list)
service_config['networks'] = network_spec
if 'logging' in key:
self._parse_logging(value, service_config)
service_config.pop('logging')
if 'environment' in key:
service_config['env'] = value
service_config.pop('environment')
return service_config
| parse_services | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_config: dict, service_name: str, service_list: dict\n ) ->dict:\n for key, value in service_list['services'][service_name].items():\n ... |
def x(self, deploy_values: dict, service_config: dict):
mode = {}
for d_value in deploy_values:
if 'restart_policy' in d_value:
restart_spec = docker.types.RestartPolicy(**deploy_values[d_value])
service_config['restart_policy'] = restart_spec
if 'placement' in d_value:
for constraints_key, constraints_value in deploy_values[d_value
].items():
service_config[constraints_key] = constraints_value
if 'mode' in d_value:
mode[d_value] = deploy_values[d_value]
if 'replicas' in d_value:
mode[d_value] = deploy_values[d_value]
if 'resources' in d_value:
resource_spec = self._parse_resources(deploy_values, d_value)
service_config['resources'] = resource_spec
mode_spec = docker.types.ServiceMode(**mode)
service_config['mode'] = mode_spec
| parse_deploy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, deploy_values: dict, service_config: dict):\n mode = {}\n for d_value in deploy_values:\n if 'restart_policy' in d_value:\n resta... |
def x(port_values: dict) ->dict:
endpoints = {}
for port_element in port_values:
target_port = port_element.split(':')
for port in target_port:
endpoints[int(port)] = int(port)
endpoint_spec = docker.types.EndpointSpec(ports=endpoints)
return endpoint_spec
| parse_ports | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(port_values: dict) ->dict:\n endpoints = {}\n for port_element in port_values:\n target_port = port_element.split(':')\n for port in target... |
def x(volume_values: dict) ->str:
for v_values in volume_values:
for v_key, v_value in v_values.items():
if v_key == 'source':
if v_value == '.':
source = os.path.dirname(os.path.abspath(__file__))
else:
source = v_value
if v_key == 'target':
target = v_value
volume_spec = [source + ':' + target]
return volume_spec
| parse_volumes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(volume_values: dict) ->str:\n for v_values in volume_values:\n for v_key, v_value in v_values.items():\n if v_key == 'source':\n ... |
def x(resource_values: dict, resource_name: str) ->dict:
resources = {}
for r_values in resource_values[resource_name]:
if 'limits' in r_values:
for r_key, r_value in resource_values[resource_name][r_values
].items():
if 'cpu' in r_key:
cpu_value = float(r_value) * 10 ** 9
cpu_key = r_key[:3] + '_limit'
resources[cpu_key] = int(cpu_value)
if 'mem' in r_key:
mem_value = re.sub('M', '', r_value)
mem_key = r_key[:3] + '_limit'
resources[mem_key] = int(mem_value) * 1048576
resources_spec = docker.types.Resources(**resources)
return resources_spec
| parse_resources | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(resource_values: dict, resource_name: str) ->dict:\n resources = {}\n for r_values in resource_values[resource_name]:\n if 'limits' in r_values:\n... |
def x(service_list: dict) ->list:
networks = []
for n_values in service_list['networks'].values():
for n_key, n_value in n_values.items():
if 'name' in n_key:
networks.append(n_value)
return networks
| parse_networks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service_list: dict) ->list:\n networks = []\n for n_values in service_list['networks'].values():\n for n_key, n_value in n_values.items():\n ... |
def x(log_values: dict, service_config: dict):
for log_key, log_value in log_values.items():
if 'driver' in log_key:
service_config['log_driver'] = log_value
if 'options' in log_key:
service_config['log_driver_options'] = log_value
| parse_logging | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(log_values: dict, service_config: dict):\n for log_key, log_value in log_values.items():\n if 'driver' in log_key:\n service_config['log_d... |
def x():
LOG.info('Initialising Processing Block queue.')
queue = ProcessingBlockQueue()
active_pb_ids = ProcessingBlockList().active
LOG.info('Initialising PC PB queue: %s', active_pb_ids)
for pb_id in active_pb_ids:
pb = ProcessingBlock(pb_id)
queue.put(pb.id, pb.priority, pb.type)
return queue
| init_queue | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n LOG.info('Initialising Processing Block queue.')\n queue = ProcessingBlockQueue()\n active_pb_ids = ProcessingBlockList().active\n LOG.info('I... |
def x(self):
LOG.info('Starting to monitor PB events')
check_counter = 0
while True:
if check_counter == 50:
check_counter = 0
LOG.debug('Checking for PB events...')
published_events = self._pb_events.get_published_events()
for event in published_events:
if event.type == 'status_changed':
LOG.info('PB status changed event: %s', event.data['status'])
if event.data['status'] == 'created':
LOG.info(
'Acknowledged PB created event (%s) for %s, [timestamp: %s]'
, event.id, event.object_id, event.timestamp)
pb = ProcessingBlock(event.object_id)
self._queue.put(event.object_id, pb.priority, pb.type)
if event.data['status'] == 'completed':
LOG.info(
'Acknowledged PB completed event (%s) for %s, [timestamp: %s]'
, event.id, event.object_id, event.timestamp)
self._num_pbcs -= 1
if self._num_pbcs < 0:
self._num_pbcs = 0
time.sleep(0.1)
check_counter += 1
| monitor_events | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting to monitor PB events')\n check_counter = 0\n while True:\n if check_counter == 50:\n check_counter = 0\n... |
def x(self):
LOG.info('Starting Processing Block queue reporter.')
while True:
LOG.info('PB queue length = %d', len(self._queue))
time.sleep(self._report_interval)
if active_count() != 5:
LOG.critical(
'Processing Controller not running correctly! (%d/%d threads active)'
, active_count(), 5)
| processing_controller_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting Processing Block queue reporter.')\n while True:\n LOG.info('PB queue length = %d', len(self._queue))\n time.sl... |
def x(self):
LOG.info('Starting to Schedule Processing Blocks.')
while True:
time.sleep(0.5)
if not self._queue:
continue
if self._num_pbcs >= self._max_pbcs:
LOG.warning('Resource limit reached!')
continue
_inspect = Inspect(app=APP)
if self._queue and _inspect.active() is not None:
next_pb = self._queue[-1]
LOG.info('Considering %s for execution...', next_pb[2])
utc_now = datetime.datetime.utcnow()
time_in_queue = utc_now - datetime_from_isoformat(next_pb[4])
if time_in_queue.total_seconds() >= 10:
item = self._queue.get()
LOG.info('------------------------------------')
LOG.info('>>> Executing %s! <<<', item)
LOG.info('------------------------------------')
execute_processing_block.delay(item)
self._num_pbcs += 1
else:
LOG.info('Waiting for resources for %s', next_pb[2])
| schedule_processing_blocks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting to Schedule Processing Blocks.')\n while True:\n time.sleep(0.5)\n if not self._queue:\n continue\n ... |
def x(self):
LOG.info('Starting to Monitor PBC status.')
inspect = celery.current_app.control.inspect()
workers = inspect.ping()
start_time = time.time()
while workers is None:
time.sleep(0.1)
elapsed = time.time() - start_time
if elapsed > 20.0:
LOG.warning('PBC not found!')
break
if workers is not None:
for worker in workers:
_tasks = inspect.registered_tasks()[worker]
LOG.info('Worker: %s tasks:', worker)
for task_index, task_name in enumerate(_tasks):
LOG.info(' %02d : %s', task_index, task_name)
while True:
LOG.info('Checking PBC status (%d/%d)', self._num_pbcs, self._max_pbcs)
celery_app = celery.current_app
inspect = celery_app.control.inspect()
workers = inspect.ping()
if workers is None:
LOG.warning('PBC service not found!')
else:
LOG.info('PBC state: %s', celery_app.events.State())
_active = inspect.active()
_scheduled = inspect.scheduled()
for worker in workers:
LOG.info(' Worker %s: scheduled: %s, active: %s', worker,
_active[worker], _scheduled[worker])
time.sleep(self._report_interval)
| monitor_pbc_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting to Monitor PBC status.')\n inspect = celery.current_app.control.inspect()\n workers = inspect.ping()\n start_time = tim... |
def x(self):
scheduler_threads = [Thread(target=self._monitor_events, daemon=True),
Thread(target=self._processing_controller_status, daemon=True),
Thread(target=self._schedule_processing_blocks, daemon=True),
Thread(target=self._monitor_pbc_status, daemon=True)]
for thread in scheduler_threads:
thread.start()
try:
for thread in scheduler_threads:
thread.join()
except KeyboardInterrupt:
LOG.info('Keyboard interrupt!')
sys.exit(0)
finally:
LOG.info('Finally!')
| start | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n scheduler_threads = [Thread(target=self._monitor_events, daemon=True),\n Thread(target=self._processing_controller_status, daemon=True),\n ... |
def x():
if request.method == 'POST':
response = {'message': 'POST Accepted'}
logging.info('alarm POSTED!')
data = request.data
logging.info(data)
string = json.dumps(data)
producer.send('SIP-alarms', string.encode())
return response
return ''
| alarm | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if request.method == 'POST':\n response = {'message': 'POST Accepted'}\n logging.info('alarm POSTED!')\n data = request.data\n ... |
def x(service: ServiceState):
LOG.debug('Setting current state from target state for %s', service.id)
service.update_current_state(service.target_state)
| update_service_current_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service: ServiceState):\n LOG.debug('Setting current state from target state for %s', service.id)\n service.update_current_state(service.target_state)\n"... |
def x(sdp_target_state: str):
service_states = get_service_state_list()
for service in service_states:
if service.current_state != sdp_target_state:
LOG.debug('Setting the current state of %s to be %s', service.
id, sdp_target_state)
service.update_current_state(sdp_target_state)
| update_services_instant_gratification | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_target_state: str):\n service_states = get_service_state_list()\n for service in service_states:\n if service.current_state != sdp_target_stat... |
def x(sdp_target_state: str):
service_states = get_service_state_list()
for service in service_states:
if service.current_state != sdp_target_state:
LOG.debug('Setting the target state of %s to be %s', service.id,
sdp_target_state)
service.update_target_state(sdp_target_state)
| update_services_target_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_target_state: str):\n service_states = get_service_state_list()\n for service in service_states:\n if service.current_state != sdp_target_stat... |
def x(sdp_state: SDPState):
LOG.info('Handling SDP target state updated event...')
LOG.info('SDP target state: %s', sdp_state.target_state)
if sdp_state.target_state == 'off':
_update_services_target_state('off')
sdp_state.update_current_state(sdp_state.target_state)
| handle_sdp_target_state_updated | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_state: SDPState):\n LOG.info('Handling SDP target state updated event...')\n LOG.info('SDP target state: %s', sdp_state.target_state)\n if sdp_sta... |
def x():
parser = argparse.ArgumentParser(description='{} service.'.format(
__service_id__))
parser.add_argument('--random_errors', action='store_true', help=
'Enable random errors')
parser.add_argument('-v', action='store_true', help=
'Verbose mode (enable debug printing)')
parser.add_argument('-vv', action='store_true', help='Extra verbose mode')
args = parser.parse_args()
if args.vv:
init_logger(log_level='DEBUG', show_log_origin=True)
elif args.v:
init_logger(logger_name='sip.ec.master_controller', log_level='DEBUG')
else:
init_logger(log_level='INFO')
return args
| parse_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description='{} service.'.format(\n __service_id__))\n parser.add_argument('--random_errors', action='store_... |
def x(sdp_state: SDPState):
LOG.info('Initialising: %s', __service_id__)
_services = ['ExecutionControl:AlarmReceiver:1.0.0',
'ExecutionControl:AlertManager:1.0.0',
'ExecutionControl:ConfigurationDatabase:5.0.1',
'ExecutionControl:MasterController:1.3.0',
'ExecutionControl:ProcessingController:1.2.6',
'ExecutionControl:ProcessingBlockController:1.3.0',
'TangoControl:Database:1.0.4', 'TangoControl:MySQL:1.0.3',
'TangoControl:SDPMaster:1.2.1', 'TangoControl:Subarrays:1.2.0',
'TangoControl:ProcessingBlocks:1.2.0', 'Platform:Kafka:2.1.1',
'Platform:Prometheus:1.0.0', 'Platform:PrometheusPushGateway:0.7.0',
'Platform:RedisCommander:210.0.0', 'Platform:Zookeeper:3.4.13']
for service_id in _services:
subsystem, name, version = service_id.split(':')
ServiceState(subsystem, name, version)
if sdp_state.current_state in ['unknown', 'off']:
try:
LOG.info("Setting the SDPState to 'init'")
sdp_state.update_current_state('init', force=True)
except ValueError as error:
LOG.critical('Unable to set the State of SDP to init! %s', str(
error))
LOG.info('Updating Service States')
service_state_list = get_service_state_list()
for service_state in service_state_list:
if service_state.current_state in ['unknown', 'off']:
service_state.update_current_state('init', force=True)
for service_state in service_state_list:
if service_state.current_state == 'init':
time.sleep(random.uniform(0, 0.2))
service_state.update_current_state('on')
services_on = [(service.current_state == 'on') for service in
service_state_list]
if all(services_on):
LOG.info('All Services are online!.')
sdp_state.update_current_state('standby')
else:
LOG.critical('Master Controller failed to initialise.')
return service_state_list
| init | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_state: SDPState):\n LOG.info('Initialising: %s', __service_id__)\n _services = ['ExecutionControl:AlarmReceiver:1.0.0',\n 'ExecutionControl:Al... |
def x(event: Event, sdp_state: SDPState, service_states: List[ServiceState]):
LOG.debug('Event detected! (id : "%s", type: "%s", data: "%s")', event.
object_id, event.type, event.data)
if event.object_id == 'SDP' and event.type == 'current_state_updated':
LOG.info('SDP current state updated, no action required!')
if event.object_id == 'SDP' and event.type == 'target_state_updated':
LOG.info("SDP target state changed to '%s'", sdp_state.target_state)
if sdp_state.target_state == sdp_state.current_state:
LOG.warning('SDP already in %s state', sdp_state.current_state)
return
if not sdp_state.is_target_state_allowed(sdp_state.target_state):
LOG.error('Transition to %s is not allowed when in state %s',
sdp_state.target_state, sdp_state.current_state)
sdp_state.target_state = sdp_state.current_state
return
_update_services_target_state(sdp_state.target_state)
if sdp_state.target_state == 'off':
LOG.info('Turning off services!')
for service_state in service_states:
service_state.update_target_state('off')
service_state.update_current_state('off')
LOG.info('Processing target state change request ...')
time.sleep(0.1)
LOG.info('Done processing target state change request!')
sdp_state.update_current_state(sdp_state.target_state)
if sdp_state.current_state == 'alarm':
LOG.debug('raising SDP state alarm')
SIP_STATE_ALARM.set(1)
else:
SIP_STATE_ALARM.set(0)
try:
push_to_gateway('platform_pushgateway:9091', job='SIP',
registry=COLLECTOR_REGISTRY)
except urllib.error.URLError:
LOG.warning('Unable to connect to the Alarms service!')
| process_event | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event: Event, sdp_state: SDPState, service_states: List[ServiceState]):\n LOG.debug('Event detected! (id : \"%s\", type: \"%s\", data: \"%s\")', event.\n ... |
def x():
sdp_state = SDPState()
service_states = get_service_state_list()
state_events = sdp_state.get_event_queue(subscriber=__service_name__)
state_is_off = sdp_state.current_state == 'off'
counter = 0
while True:
time.sleep(0.1)
if not state_is_off:
if counter % 1000 == 0:
LOG.debug('Checking published events ... %d', counter / 1000)
_published_events = state_events.get_published_events(process
=True)
for _state_event in _published_events:
_process_event(_state_event, sdp_state, service_states)
else:
_state_event = state_events.get()
if _state_event:
_process_event(_state_event, sdp_state, service_states)
state_is_off = sdp_state.current_state == 'off'
counter += 1
| process_state_change_events | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n sdp_state = SDPState()\n service_states = get_service_state_list()\n state_events = sdp_state.get_event_queue(subscriber=__service_name__)\n s... |
def x():
_parse_args()
LOG.info('Starting: %s', __service_id__)
LOG.info('Subscribing to state change events (subscriber = %s)',
__service_name__)
sdp_state = SDPState()
_ = sdp_state.subscribe(subscriber=__service_name__)
_ = _init(sdp_state)
LOG.info('Finished initialising!')
LOG.info('Responding to state change events ...')
try:
_process_state_change_events()
except ValueError as error:
LOG.critical('Value error: %s', str(error))
except KeyboardInterrupt as err:
LOG.debug('Keyboard Interrupt %s', err)
LOG.info('Exiting!')
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n _parse_args()\n LOG.info('Starting: %s', __service_id__)\n LOG.info('Subscribing to state change events (subscriber = %s)',\n __service_na... |
def x():
stream_config = spead2.send.StreamConfig(max_packet_size=16356, rate=
1000000000.0, burst_size=10, max_heaps=1)
item_group = spead2.send.ItemGroup(flavour=spead2.Flavour(4, 64, 48, 0))
num_baselines = 512 * 513 // 2
dtype = [('TCI', 'i1'), ('FD', 'u1'), ('VIS', '<c8', 4)]
item_group.add_item(id=24576, name='visibility_timestamp_count',
description='', shape=tuple(), format=None, dtype='<u4')
item_group.add_item(id=24577, name='visibility_timestamp_fraction',
description='', shape=tuple(), format=None, dtype='<u4')
item_group.add_item(id=24581, name='visibility_baseline_count',
description='', shape=tuple(), format=None, dtype='<u4')
item_group.add_item(id=24584, name='scan_id', description='', shape=
tuple(), format=None, dtype='<u8')
item_group.add_item(id=24586, name='correlator_output_data',
description='', shape=(num_baselines,), dtype=dtype)
streams = []
num_streams = 2
for i in range(num_streams):
stream = spead2.send.UdpStream(thread_pool=spead2.ThreadPool(
threads=1), hostname='127.0.0.1', port=41000 + i, config=
stream_config)
stream.send_heap(item_group.get_start())
streams.append(stream)
vis = numpy.zeros(shape=(num_baselines,), dtype=dtype)
num_heaps = 200
start_time = time.time()
for stream in streams:
item_group['visibility_timestamp_count'].value = 1
item_group['visibility_timestamp_fraction'].value = 0
item_group['visibility_baseline_count'].value = num_baselines
item_group['scan_id'].value = 100000000
item_group['correlator_output_data'].value = vis
for i in range(num_heaps):
stream.send_heap(item_group.get_heap(descriptors='all', data='all')
)
duration = time.time() - start_time
data_size = num_streams * num_heaps * (vis.nbytes / 1000000.0)
print('Sent %.3f MB in %.3f sec (%.3f MB/sec)' % (data_size, duration,
data_size / duration))
for stream in streams:
stream.send_heap(item_group.get_end())
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n stream_config = spead2.send.StreamConfig(max_packet_size=16356, rate=\n 1000000000.0, burst_size=10, max_heaps=1)\n item_group = spead2.send.... |
def x(cls, config: dict):
timestamp = config.get('timestamp', None)
return cls(config.get('id'), config.get('type'), config.get('data',
dict()), config.get('origin', None), timestamp, config.get(
'object_type', None), config.get('object_id', None), config.get(
'object_key', None))
| from_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, config: dict):\n timestamp = config.get('timestamp', None)\n return cls(config.get('id'), config.get('type'), config.get('data',\n dict()), c... |
def x(filename, imager, grid_data, grid_norm, grid_weights):
ms = oskar.MeasurementSet.open(filename)
block_start = 0
num_rows = ms.num_rows
num_baselines = ms.num_stations * (ms.num_stations - 1) // 2
while block_start < num_rows:
block_size = num_rows - block_start
if block_size > num_baselines:
block_size = num_baselines
uvw = ms.read_column('UVW', block_start, block_size)
vis_weights = ms.read_column('WEIGHT', block_start, block_size)
if ms.num_pols == 4:
vis_weights = 0.5 * (vis_weights[:, 0] + vis_weights[:, 3])
for j in range(ms.num_channels):
coords = uvw * (ms.freq_start_hz + j * ms.freq_inc_hz
) / 299792458.0
vis_data = None
if not imager.coords_only:
vis_data = ms.read_vis(block_start, j, 1, block_size)
if ms.num_pols == 4:
vis_data = 0.5 * (vis_data[0, :, 0] + vis_data[0, :, 3])
grid_norm = imager.update_plane(coords[:, 0], coords[:, 1],
coords[:, 2], vis_data, vis_weights, grid_data, grid_norm,
grid_weights)
block_start += block_size
return grid_norm
| process_input_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, imager, grid_data, grid_norm, grid_weights):\n ms = oskar.MeasurementSet.open(filename)\n block_start = 0\n num_rows = ms.num_rows\n num_... |
def x():
if len(sys.argv) < 2:
raise RuntimeError(
'Usage: mpiexec -n <np> python mpi_imager_test.py <settings_file> <dir>'
)
comm = MPI.COMM_WORLD
settings = None
inputs = None
grid_weights = None
log = logging.getLogger()
log.setLevel(logging.DEBUG)
if len(log.handlers) == 0:
log.addHandler(logging.StreamHandler(sys.stdout))
if comm.Get_rank() == 0:
with open(sys.argv[1]) as f:
settings = json.load(f)
data_dir = str(sys.argv[2])
inputs = glob(join(data_dir, '*.ms')) + glob(join(data_dir, '*.MS'))
inputs = filter(None, inputs)
log.info('Found input Measurement Sets: %s', ', '.join(inputs))
inputs = chunks(inputs, comm.Get_size())
comm.barrier()
settings = comm.bcast(settings)
inputs = comm.scatter(inputs)
log.debug('Rank %d, processing [%s]', comm.Get_rank(), ', '.join(inputs))
precision = settings['precision']
imager = oskar.Imager(precision)
for key, value in settings['imager'].items():
setattr(imager, key, value)
grid_norm = 0.0
grid_dim = [imager.plane_size, imager.plane_size]
grid_data = numpy.zeros(grid_dim, dtype='c8' if precision == 'single' else
'c16')
if settings['combine']:
if imager.weighting == 'Uniform' or imager.algorithm == 'W-projection':
local_weights = None
if imager.weighting == 'Uniform':
grid_weights = numpy.zeros(grid_dim, dtype=precision)
local_weights = numpy.zeros(grid_dim, dtype=precision)
imager.coords_only = True
for f in inputs:
log.info('Reading coordinates from %s', f)
process_input_data(f, imager, None, 0.0, local_weights)
imager.coords_only = False
num_w_planes = imager.num_w_planes
num_w_planes = comm.allreduce(num_w_planes, op=MPI.MAX)
imager.num_w_planes = num_w_planes
if local_weights is not None:
comm.Allreduce(local_weights, grid_weights, op=MPI.SUM)
for f in inputs:
log.info('Reading visibilities from %s', f)
grid_norm = process_input_data(f, imager, grid_data, grid_norm,
grid_weights)
grid = numpy.zeros_like(grid_data)
comm.Reduce(grid_data, grid, op=MPI.SUM)
grid_norm = comm.reduce(grid_norm, op=MPI.SUM)
if comm.Get_rank() == 0:
save_image(imager, grid, grid_norm, settings['output_file'])
log.info('Finished. Output file is %s', settings['output_file'])
else:
for f in inputs:
grid_norm = 0.0
grid_data.fill(0)
if imager.weighting == 'Uniform':
grid_weights = numpy.zeros(grid_dim, dtype=precision)
if (imager.weighting == 'Uniform' or imager.algorithm ==
'W-projection'):
imager.coords_only = True
log.info('Reading coordinates from %s', f)
process_input_data(f, imager, None, 0.0, grid_weights)
imager.coords_only = False
log.info('Reading visibilities from %s', f)
grid_norm = process_input_data(f, imager, grid_data, grid_norm,
grid_weights)
output_file = splitext(f)[0] + '.fits'
save_image(imager, grid_data, grid_norm, output_file)
log.info('Finished. Output file is %s', output_file)
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if len(sys.argv) < 2:\n raise RuntimeError(\n 'Usage: mpiexec -n <np> python mpi_imager_test.py <settings_file> <dir>'\n )... |
def x():
LOG.debug('GET list of SBIs.')
_url = get_root_url()
response = dict(scheduling_blocks=[], links=dict(home='{}'.format(_url)))
block_ids = DB.get_sched_block_instance_ids()
for block in DB.get_block_details(block_ids):
block_id = block['id']
LOG.debug('Adding SBI %s to list', block_id)
LOG.debug(block)
block['num_processing_blocks'] = len(block['processing_block_ids'])
temp = ['OK'] * 10 + ['WAITING'] * 4 + ['FAILED'] * 2
block['status'] = choice(temp)
try:
del block['processing_block_ids']
except KeyError:
pass
block['links'] = {'detail': '{}/scheduling-block/{}'.format(_url,
block_id)}
response['scheduling_blocks'].append(block)
return response, HTTPStatus.OK
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n LOG.debug('GET list of SBIs.')\n _url = get_root_url()\n response = dict(scheduling_blocks=[], links=dict(home='{}'.format(_url)))\n block_ids... |
def x():
response = dict(blocks=[])
block_ids = DB.get_sched_block_instance_ids()
for index, block_id in enumerate(block_ids):
block = DB.get_block_details([block_id]).__next__()
info = [index, block['id'], block['sub_array_id'], len(block[
'processing_blocks'])]
response['blocks'].append(info)
return response, HTTPStatus.OK
| get_table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n response = dict(blocks=[])\n block_ids = DB.get_sched_block_instance_ids()\n for index, block_id in enumerate(block_ids):\n block = DB.get... |
def x(self, key, field, value):
self._db.hset(key, field, value)
| set_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, field, value):\n self._db.hset(key, field, value)\n",
"role": "user"
},
{
"content": "set_value",
"role": "assistant"
}
] |
def x(self, event_name, event_type, block_id):
self._db.rpush(event_name, dict(type=event_type, id=block_id))
| push_event | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_name, event_type, block_id):\n self._db.rpush(event_name, dict(type=event_type, id=block_id))\n",
"role": "user"
},
{
"content": "pu... |
def x():
signal.signal(signal.SIGTERM, _sig_handler)
with open(sys.argv[1]) as fh:
config = json.load(fh)
os.chdir(os.path.expanduser('~'))
receiver = PulsarStart(config, logging.getLogger())
receiver.run()
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n signal.signal(signal.SIGTERM, _sig_handler)\n with open(sys.argv[1]) as fh:\n config = json.load(fh)\n os.chdir(os.path.expanduser('~'))\n... |
def x(func):
@wraps(func)
def x(*args, **kwargs):
try:
return func(*args, **kwargs)
except redis.exceptions.ConnectionError:
raise ConnectionError(
'Unable to connect to the Redis Configuration Database. host = {}, port = {}, id = {}.'
.format(REDIS_HOST, REDIS_PORT, REDIS_DB_ID))
return with_exception_handling
| check_connection | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func):\n\n @wraps(func)\n def x(*args, **kwargs):\n try:\n return func(*args, **kwargs)\n except redis.exceptions.ConnectionErro... |
def x(self, key: str, my_dict: dict, hierarchical: bool=False):
for _key, _value in my_dict.items():
if isinstance(_value, dict):
if not hierarchical:
self._db.hmset(key, {_key: json.dumps(_value)})
else:
self.save_dict(key + ':' + _key, _value, hierarchical)
elif isinstance(_value, list):
if not hierarchical:
self._db.hmset(key, {_key: str(_value)})
else:
print('saving list at ', key + ':' + _key)
self._db.lpush(key + ':' + _key, *_value[::-1])
elif isinstance(_value, bool):
self._db.hmset(key, {_key: str(_value)})
else:
self._db.hmset(key, {_key: _value})
| save_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key: str, my_dict: dict, hierarchical: bool=False):\n for _key, _value in my_dict.items():\n if isinstance(_value, dict):\n if not h... |
def x(my_dict, keys, values):
temp = my_dict
for depth, key in enumerate(keys):
if depth < len(keys) - 1:
if key not in temp:
temp[key] = dict()
temp = temp[key]
elif key not in temp:
temp[key] = values
else:
temp[key] = {**temp[key], **values}
return my_dict
| build_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(my_dict, keys, values):\n temp = my_dict\n for depth, key in enumerate(keys):\n if depth < len(keys) - 1:\n if key not in temp:\n ... |
def x(self, db_key: str) ->dict:
if self._db.type(db_key) == 'list':
db_values = self._db.lrange(db_key, 0, -1)
for i, value in enumerate(db_values):
try:
db_values[i] = ast.literal_eval(value)
except SyntaxError:
pass
except ValueError:
pass
else:
db_values = self._db.hgetall(db_key)
for _key, _value in db_values.items():
try:
db_values[_key] = ast.literal_eval(_value)
except SyntaxError:
pass
except ValueError:
pass
return db_values
| load_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str) ->dict:\n if self._db.type(db_key) == 'list':\n db_values = self._db.lrange(db_key, 0, -1)\n for i, value in enumerate(db_v... |
def x(self, db_key: str) ->dict:
db_keys = self._db.keys(pattern=db_key + '*')
my_dict = {}
for _db_key in db_keys:
if self._db.type(_db_key) == 'list':
db_values = self._db.lrange(_db_key, 0, -1)
for i, value in enumerate(db_values):
try:
db_values[i] = ast.literal_eval(value)
except SyntaxError:
pass
except ValueError:
pass
else:
db_values = self._db.hgetall(_db_key)
for _key, _value in db_values.items():
try:
db_values[_key] = ast.literal_eval(_value)
except SyntaxError:
pass
except ValueError:
pass
my_dict = self._build_dict(my_dict, _db_key.split(':'), db_values)
return my_dict[db_key]
| load_dict_hierarchical | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str) ->dict:\n db_keys = self._db.keys(pattern=db_key + '*')\n my_dict = {}\n for _db_key in db_keys:\n if self._db.type(_db_key)... |
def x(self, db_key: str, hierarchical: bool=False) ->dict:
if not hierarchical:
db_values = self._db.hgetall(db_key)
for _key, _value in db_values.items():
if isinstance(_value, str):
db_values[_key] = ast.literal_eval(_value)
my_dict = db_values
else:
my_dict = self._load_dict_hierarchical(db_key)
return my_dict
| load_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str, hierarchical: bool=False) ->dict:\n if not hierarchical:\n db_values = self._db.hgetall(db_key)\n for _key, _value in db_va... |
def x(self, db_key: str, dict_keys: List[str], hierarchical: bool=False
) ->List:
result = []
if not hierarchical:
_values = self._db.hmget(db_key, *dict_keys)
result = [ast.literal_eval(_value) for _value in _values]
else:
db_keys = self._db.keys(pattern=db_key + '*')
for _db_key in db_keys:
for name in _db_key.split(':')[1:]:
if name in dict_keys:
_values = self._load_values(_db_key)
result.append(_values)
_values = self._db.hmget(_db_key, *dict_keys)
for i, value in enumerate(_values):
try:
_values[i] = ast.literal_eval(value)
except SyntaxError:
pass
except ValueError:
pass
result += [value for value in _values if value is not None]
return result
| load_dict_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str, dict_keys: List[str], hierarchical: bool=False\n ) ->List:\n result = []\n if not hierarchical:\n _values = self._db.hmget(d... |
def x(self, key, field, value, pipeline=False):
if pipeline:
self._pipeline.hset(key, field, str(value))
else:
self._db.hset(key, field, str(value))
| set_hash_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, field, value, pipeline=False):\n if pipeline:\n self._pipeline.hset(key, field, str(value))\n else:\n self._db.hset(key, field, ... |
def x(self, key, *value, pipeline=False):
if pipeline:
self._pipeline.lpush(key, *value)
else:
self._db.lpush(key, *value)
| prepend_to_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, *value, pipeline=False):\n if pipeline:\n self._pipeline.lpush(key, *value)\n else:\n self._db.lpush(key, *value)\n",
"role"... |
def x(self, key, *value, pipeline=False):
if pipeline:
self._pipeline.rpush(key, *value)
else:
self._db.rpush(key, *value)
| append_to_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, *value, pipeline=False):\n if pipeline:\n self._pipeline.rpush(key, *value)\n else:\n self._db.rpush(key, *value)\n",
"role"... |
def x(self, key, pipeline=False):
if pipeline:
return self._pipeline.lrange(key, 0, -1)
return self._db.lrange(key, 0, -1)
| get_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, pipeline=False):\n if pipeline:\n return self._pipeline.lrange(key, 0, -1)\n return self._db.lrange(key, 0, -1)\n",
"role": "user"
... |
def x(self, *names: str, pipeline=False):
if pipeline:
self._pipeline.delete(*names)
else:
self._db.delete(*names)
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *names: str, pipeline=False):\n if pipeline:\n self._pipeline.delete(*names)\n else:\n self._db.delete(*names)\n",
"role": "user"... |
def x(self, event_name, event_history=None):
if event_history is None:
event_history = event_name + '_history'
return self._db.rpoplpush(event_name, event_history)
| get_event | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_name, event_history=None):\n if event_history is None:\n event_history = event_name + '_history'\n return self._db.rpoplpush(event_nam... |
def x(self, key: str, value, count: int=0, pipeline: bool=False):
if pipeline:
if redis.__version__ == '2.10.6':
self._pipeline.lrem(name=key, value=value, num=count)
else:
self._pipeline.lrem(key, count, value)
elif self._db.exists(key):
if redis.__version__ == '2.10.6':
self._db.lrem(name=key, value=value, num=count)
else:
self._db.lrem(key, count, value)
| remove_from_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key: str, value, count: int=0, pipeline: bool=False):\n if pipeline:\n if redis.__version__ == '2.10.6':\n self._pipeline.lrem(name=... |
def x(self, key, pipeline=False):
if pipeline:
self._pipeline.watch(key)
else:
self._db.watch(key)
| watch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, pipeline=False):\n if pipeline:\n self._pipeline.watch(key)\n else:\n self._db.watch(key)\n",
"role": "user"
},
{
"c... |
def x(self, channel, message, pipeline=False):
if pipeline:
self._pipeline.publish(channel, message)
else:
self._db.publish(channel, message)
| publish | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, channel, message, pipeline=False):\n if pipeline:\n self._pipeline.publish(channel, message)\n else:\n self._db.publish(channel, mess... |
def x(self) ->dict:
services_health = {}
services_ids = self._get_services()
for service_id in services_ids:
service_name = DC.get_service_name(service_id)
if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):
services_health[service_name] = 'Unhealthy'
else:
services_health[service_name] = 'Healthy'
return services_health
| get_services_health | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->dict:\n services_health = {}\n services_ids = self._get_services()\n for service_id in services_ids:\n service_name = DC.get_service_na... |
def x(self) ->str:
services_health_status = self.get_services_health()
health_status = all(status == 'Healthy' for status in
services_health_status.values())
if health_status:
overall_status = 'Healthy'
else:
overall_status = 'Unhealthy'
return overall_status
| get_overall_services_health | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->str:\n services_health_status = self.get_services_health()\n health_status = all(status == 'Healthy' for status in\n services_health_statu... |
def x(service_id: str) ->str:
if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):
health_status = 'Unhealthy'
else:
health_status = 'Healthy'
return health_status
| get_service_health | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service_id: str) ->str:\n if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):\n health_status = 'Unhealthy'\n else:\n heal... |
def x():
uptime = time.time() - START_TIME
response = dict(uptime=f'{uptime:.2f}s', links=dict(root='{}'.format(
get_root_url())))
return response, HTTPStatus.OK
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n uptime = time.time() - START_TIME\n response = dict(uptime=f'{uptime:.2f}s', links=dict(root='{}'.format(\n get_root_url())))\n return res... |
def x():
if len(sys.argv) < 2:
raise RuntimeError('Usage: python3 async_recv.py <json config>')
sip_logging.init_logger(show_thread=True)
spead_config = json.loads(sys.argv[1])
receiver = SpeadReceiver(spead_config)
receiver.run()
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if len(sys.argv) < 2:\n raise RuntimeError('Usage: python3 async_recv.py <json config>')\n sip_logging.init_logger(show_thread=True)\n spe... |
def x(self, i_block, receive_buffer):
self._log.info('Worker thread processing block %i', i_block)
time_overall0 = time.time()
time_unpack = 0.0
time_write = 0.0
for i_heap, heap in enumerate(receive_buffer.result()):
if isinstance(heap, spead2.recv.IncompleteHeap):
self._log.info('Dropped incomplete heap %i', heap.cnt + 1)
continue
items = self._item_group.update(heap)
i_chan = i_heap // self._num_buffer_times
i_time = i_heap % self._num_buffer_times
if 'correlator_output_data' in items:
vis_data = items['correlator_output_data'].value['VIS']
if self._block is None:
num_baselines = vis_data.shape[0]
num_pols = vis_data[0].shape[0]
self._block = numpy.zeros((self._num_buffer_times, self.
_num_streams, num_baselines), dtype=('c8', num_pols))
self._block[:, :, :] = 0
time_unpack0 = time.time()
self._block[i_time, i_chan, :] = vis_data
time_unpack += time.time() - time_unpack0
val = self._block[i_time, i_chan, -1][-1].real
self._log.debug('Data: %.3f', val)
if self._block is not None:
if self._config['process_data']:
pass
if self._config['write_data']:
time_write0 = time.time()
with open(self._config['filename'], 'ab') as f:
numpy.save(f, self._block, allow_pickle=False)
time_write += time.time() - time_write0
time_overall = time.time() - time_overall0
self._log.info('Total processing time: %.1f ms', 1000 * time_overall)
self._log.info('Unpack was %.1f %%', 100 * time_unpack / time_overall)
self._log.info('Write was %.1f %%', 100 * time_write / time_overall)
if time_unpack != 0.0:
self._log.info('Memory speed %.1f MB/s', self._block.nbytes * 1e-06 /
time_unpack)
if time_write != 0.0:
self._log.info('Write speed %.1f MB/s', self._block.nbytes * 1e-06 /
time_write)
| process_buffer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, i_block, receive_buffer):\n self._log.info('Worker thread processing block %i', i_block)\n time_overall0 = time.time()\n time_unpack = 0.0\n ... |
async def _run_loop(self, executor):
"""Main loop."""
loop = asyncio.get_event_loop()
self._log.info('Waiting for %d streams to start...', self._num_streams)
for stream in self._streams:
await stream.get(loop=loop)
i_block = 0
receive_buffer = [None, None]
while True:
processing_tasks = None
if i_block > 0:
i_buffer_proc = (i_block - 1) % 2
processing_tasks = loop.run_in_executor(executor, self.
process_buffer, i_block - 1, receive_buffer[i_buffer_proc])
i_buffer_recv = i_block % 2
receive_tasks = []
for stream in self._streams:
for _ in range(self._num_buffer_times):
receive_tasks.append(loop.create_task(stream.get(loop=loop)))
receive_buffer[i_buffer_recv] = asyncio.gather(*receive_tasks)
self._log.info('Receiving block %i', i_block)
try:
await receive_buffer[i_buffer_recv]
self._log.info('Received block %i', i_block)
except spead2.Stopped:
self._log.info('Stream Stopped')
break
if processing_tasks:
await processing_tasks
i_block += 1
| run_loop | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def _run_loop(self, executor):\n \"\"\"Main loop.\"\"\"\n loop = asyncio.get_event_loop()\n self._log.info('Waiting for %d streams to start...', self.... |
def x(self):
executor = concurrent.futures.ThreadPoolExecutor(max_workers=1)
loop = asyncio.get_event_loop()
loop.run_until_complete(self._run_loop(executor))
self._log.info('Shutting down...')
executor.shutdown()
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n executor = concurrent.futures.ThreadPoolExecutor(max_workers=1)\n loop = asyncio.get_event_loop()\n loop.run_until_complete(self._run_loop(ex... |
def x(self):
self.seek(0)
while True:
brace_count = 0
quoted = False
json_string = ''
while True:
b = self.read(1)
if b == b'':
return
c = b.decode()
if c == '\\':
json_string += c
json_string += self.read(1).decode()
continue
if c == '"':
if quoted:
quoted = False
else:
quoted = True
if c == '{':
brace_count += 1
elif c == '}':
brace_count -= 1
json_string += c
if brace_count == 0:
break
meta = json.loads(json_string)
n_channels = meta['data_cube']['n_channels']
n_sub_integrations = meta['data_cube']['n_sub_integrations']
f = open('{0}_{1}_{2}.json'.format(meta['metadata'][
'observation_id'], meta['metadata']['beam_id'], meta['metadata'
]['name']), 'w')
f.write(json.dumps(meta))
f.close()
data = self.read(n_channels * n_sub_integrations)
f = open('{0}_{1}_{2}.data'.format(meta['metadata'][
'observation_id'], meta['metadata']['beam_id'], meta['metadata'
]['name']), 'wb')
f.write(data)
f.close()
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.seek(0)\n while True:\n brace_count = 0\n quoted = False\n json_string = ''\n while True:\n b = self... |
def x(self):
self._log.info('Starting Pulsar Search Interface')
authorizer = DummyAuthorizer()
authorizer.add_user(self._config['login']['user'], self._config['login'
]['psswd'], '.', perm=self._config['login']['perm'])
authorizer.add_anonymous(os.getcwd())
handler = FTPHandler
handler.authorizer = authorizer
handler.abstracted_fs = PulsarFileSystem
handler.banner = 'SKA SDP pulsar search interface.'
address = self._config['address']['listen'], self._config['address']['port'
]
server = FTPServer(address, handler)
server.max_cons = 256
server.max_cons_per_ip = 5
server.serve_forever()
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._log.info('Starting Pulsar Search Interface')\n authorizer = DummyAuthorizer()\n authorizer.add_user(self._config['login']['user'], self... |
def x(self) ->str:
self._config = self._load_config()
return self._config.get('status')
| status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->str:\n self._config = self._load_config()\n return self._config.get('status')\n",
"role": "user"
},
{
"content": "status",
"role"... |
def x(self, value):
pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)
stages = DB.get_hash_value(pb_key, 'workflow_stages')
stages = ast.literal_eval(stages)
stages[self._index]['status'] = value
DB.set_hash_value(pb_key, 'workflow_stages', stages)
| status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)\n stages = DB.get_hash_value(pb_key, 'workflow_stages')\n stages = ast.literal_... |
def x(self):
pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)
stages = DB.get_hash_value(pb_key, 'workflow_stages')
stages = ast.literal_eval(stages)
return stages[self._index]
| load_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)\n stages = DB.get_hash_value(pb_key, 'workflow_stages')\n stages = ast.literal_eval(st... |
def x():
parser = argparse.ArgumentParser(description='Register PB devices.')
parser.add_argument('num_pb', type=int, help=
'Number of PBs devices to register.')
return parser.parse_args()
| parse_command_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description='Register PB devices.')\n parser.add_argument('num_pb', type=int, help=\n 'Number of PBs devices... |
def x(num_pbs: int=100):
tango_db = Database()
LOG.info('Registering PB devices:')
dev_info = DbDevInfo()
dev_info._class = 'ProcessingBlockDevice'
dev_info.server = 'processing_block_ds/1'
for index in range(num_pbs):
dev_info.name = 'sip_sdp/pb/{:05d}'.format(index)
LOG.info('\t%s', dev_info.name)
tango_db.add_device(dev_info)
| register_pb_devices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_pbs: int=100):\n tango_db = Database()\n LOG.info('Registering PB devices:')\n dev_info = DbDevInfo()\n dev_info._class = 'ProcessingBlockDevic... |
def x(cls, config_dict: dict, schema_path: str=None):
if schema_path is None:
schema_path = join(dirname(__file__), 'schema', 'configure_sbi.json')
with open(schema_path, 'r') as file:
schema = json.loads(file.read())
validate(config_dict, schema)
config_dict['status'] = 'created'
if 'subarray_id' not in config_dict:
config_dict['subarray_id'] = 'None'
timestamp = datetime.datetime.utcnow().isoformat()
config_dict['created'] = timestamp
config_dict['updated'] = timestamp
pb_list = copy.deepcopy(config_dict['processing_blocks'])
config_dict.pop('processing_blocks', None)
config_dict['processing_block_ids'] = []
for pb in pb_list:
config_dict['processing_block_ids'].append(pb['id'])
key = SchedulingObject.get_key(SBI_KEY, config_dict['id'])
DB.save_dict(key, config_dict, hierarchical=False)
key = '{}:active'.format(SBI_KEY)
DB.append_to_list(key, config_dict['id'])
sbi = SchedulingObject(SBI_KEY, config_dict['id'])
sbi.set_status('created')
for pb in pb_list:
pb['sbi_id'] = config_dict['id']
cls._add_pb(pb)
return cls(config_dict['id'])
| from_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, config_dict: dict, schema_path: str=None):\n if schema_path is None:\n schema_path = join(dirname(__file__), 'schema', 'configure_sbi.json')\n ... |
def x(self):
self.set_status('aborted')
DB.remove_from_list('{}:active'.format(self._type), self._id)
DB.append_to_list('{}:aborted'.format(self._type), self._id)
sbi_pb_ids = ast.literal_eval(DB.get_hash_value(self._key,
'processing_block_ids'))
for pb_id in sbi_pb_ids:
pb = ProcessingBlock(pb_id)
pb.abort()
| abort | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.set_status('aborted')\n DB.remove_from_list('{}:active'.format(self._type), self._id)\n DB.append_to_list('{}:aborted'.format(self._type... |
def x(self) ->List[str]:
values = DB.get_hash_value(self._key, 'processing_block_ids')
return ast.literal_eval(values)
| get_pb_ids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->List[str]:\n values = DB.get_hash_value(self._key, 'processing_block_ids')\n return ast.literal_eval(values)\n",
"role": "user"
},
{
... |
def x(date=None, project: str='sip', instance_id: int=None) ->str:
if date is None:
date = datetime.datetime.utcnow()
if isinstance(date, datetime.datetime):
date = date.strftime('%Y%m%d')
if instance_id is None:
instance_id = randint(0, 9999)
return 'SBI-{}-{}-{:04d}'.format(date, project, instance_id)
| get_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(date=None, project: str='sip', instance_id: int=None) ->str:\n if date is None:\n date = datetime.datetime.utcnow()\n if isinstance(date, datetime... |
def x(pb_config: dict):
pb_config['status'] = 'created'
timestamp = datetime.datetime.utcnow().isoformat()
pb_config['created'] = timestamp
pb_config['updated'] = timestamp
if 'priority' not in pb_config:
pb_config['priority'] = 0
SchedulingBlockInstance._update_workflow_definition(pb_config)
keys = ['resources_required', 'resources_assigned', 'dependencies']
for key in keys:
if key not in pb_config:
pb_config[key] = []
for stage in pb_config['workflow_stages']:
if key not in stage:
stage[key] = []
key = SchedulingObject.get_key(PB_KEY, pb_config['id'])
DB.save_dict(key, pb_config, hierarchical=False)
key = '{}:active'.format(PB_KEY)
DB.append_to_list(key, pb_config['id'])
key = '{}:active:{}'.format(PB_KEY, pb_config['type'])
DB.append_to_list(key, pb_config['id'])
pb = SchedulingObject(PB_KEY, pb_config['id'])
pb.set_status('created')
| add_pb | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pb_config: dict):\n pb_config['status'] = 'created'\n timestamp = datetime.datetime.utcnow().isoformat()\n pb_config['created'] = timestamp\n pb_co... |
def x(pb_config: dict):
known_workflows = get_workflows()
workflow_id = pb_config['workflow']['id']
workflow_version = pb_config['workflow']['version']
if (workflow_id not in known_workflows or workflow_version not in
known_workflows[workflow_id]):
raise RuntimeError('Unknown workflow definition: {}:{}'.format(
workflow_id, workflow_version))
workflow = get_workflow(workflow_id, workflow_version)
for stage in workflow['stages']:
stage['status'] = 'none'
pb_config['workflow_parameters'] = pb_config['workflow']['parameters']
pb_config['workflow_id'] = pb_config['workflow']['id']
pb_config['workflow_version'] = pb_config['workflow']['version']
pb_config['workflow_stages'] = workflow['stages']
pb_config.pop('workflow', None)
| update_workflow_definition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pb_config: dict):\n known_workflows = get_workflows()\n workflow_id = pb_config['workflow']['id']\n workflow_version = pb_config['workflow']['version'... |
def x():
response = {'links': {'message':
'Welcome to the SIP Processing Controller Interface', 'items': [{
'href': '{}health'.format(request.url)}, {'href': '{}subarrays'.
format(request.url)}, {'href': '{}scheduling_blocks'.format(request
.url)}, {'href': '{}processing_blocks'.format(request.url)}]}}
return response, HTTPStatus.OK
| root | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n response = {'links': {'message':\n 'Welcome to the SIP Processing Controller Interface', 'items': [{\n 'href': '{}health'.format(request.... |
def x(num_blocks, project='test'):
_date = strftime('%Y%m%d', gmtime())
_project = project
for i in range(num_blocks):
yield '{}-{}-sbi{:03d}'.format(_date, _project, i)
| generate_scheduling_block_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_blocks, project='test'):\n _date = strftime('%Y%m%d', gmtime())\n _project = project\n for i in range(num_blocks):\n yield '{}-{}-sbi{:03d}... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.