source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, duration=None, preauthenticated=False, filename=None, project=
None, **kwargs):
args = {'preauthenticated': preauthenticated}
if duration is not None:
args['duration'] = duration
if filename is not None:
args['filename'] = filename
if project is None and 'DX_JOB_ID' not in os.environ:
project_from_handler = self.get_proj_id()
if project_from_handler and object_exists_in_project(self.get_id(),
project_from_handler):
project = project_from_handler
if project is not None and project is not DXFile.NO_PROJECT_HINT:
args['project'] = project
if '_DX_DUMP_BILLED_PROJECT' in os.environ:
with open(os.environ['_DX_DUMP_BILLED_PROJECT'], 'w') as fd:
if project is not None and project != DXFile.NO_PROJECT_HINT:
fd.write(project)
with self._url_download_mutex:
if (self._download_url is None or self._download_url_expires < time
.time()):
if 'timeout' not in kwargs:
kwargs['timeout'] = FILE_REQUEST_TIMEOUT
resp = dxpy.api.file_download(self._dxid, args, **kwargs)
self._download_url = resp['url']
self._download_url_headers = _validate_headers(resp.get(
'headers', {}))
if preauthenticated:
self._download_url_expires = resp['expires'] / 1000 - 60
else:
self._download_url_expires = 32503680000
retval_download_url = self._download_url
retval_download_url_headers = copy.copy(self._download_url_headers)
return retval_download_url, retval_download_url_headers
| get_download_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, duration=None, preauthenticated=False, filename=None, project=\n None, **kwargs):\n args = {'preauthenticated': preauthenticated}\n if duration ... |
def x(self, length=None, use_compression=None, project=None, **kwargs):
if self._file_length == None:
desc = self.describe(**kwargs)
if desc['state'] != 'closed':
raise DXFileError(
'Cannot read from file until it is in the closed state')
self._file_length = int(desc['size'])
get_first_chunk_sequentially = (self._file_length > 128 * 1024 and self
._pos == 0 and dxpy.JOB_ID)
if self._pos == self._file_length:
return b''
if length == None or length > self._file_length - self._pos:
length = self._file_length - self._pos
buf = self._read_buf
buf_remaining_bytes = dxpy.utils.string_buffer_length(buf) - buf.tell()
if length <= buf_remaining_bytes:
self._pos += length
return buf.read(length)
else:
orig_buf_pos = buf.tell()
orig_file_pos = self._pos
buf.seek(0, os.SEEK_END)
self._pos += buf_remaining_bytes
while self._pos < orig_file_pos + length:
remaining_len = orig_file_pos + length - self._pos
if self._response_iterator is None:
self._request_iterator = self._generate_read_requests(start_pos
=self._pos, project=project, **kwargs)
content = self._next_response_content(get_first_chunk_sequentially
=get_first_chunk_sequentially)
if len(content) < remaining_len:
buf.write(content)
self._pos += len(content)
else:
buf.write(content[:remaining_len])
self._pos += remaining_len
self._read_buf = BytesIO()
self._read_buf.write(content[remaining_len:])
self._read_buf.seek(0)
buf.seek(orig_buf_pos)
return buf.read()
| read2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, length=None, use_compression=None, project=None, **kwargs):\n if self._file_length == None:\n desc = self.describe(**kwargs)\n if desc['... |
def x(job_id, local_port, remote_port):
cmd = ['dx', 'ssh', '--suppress-running-check', job_id, '-o',
'StrictHostKeyChecking no']
cmd += ['-f', '-L', '{0}:localhost:{1}'.format(local_port, remote_port),
'-N']
subprocess.check_call(cmd)
| setup_ssh_tunnel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_id, local_port, remote_port):\n cmd = ['dx', 'ssh', '--suppress-running-check', job_id, '-o',\n 'StrictHostKeyChecking no']\n cmd += ['-f', '-... |
def x(job_id):
sys.stdout.write('Waiting for server in {0} to initialize ...'.format(
job_id))
sys.stdout.flush()
desc = dxpy.describe(job_id)
while SERVER_READY_TAG not in desc['tags'] and desc['state'] != 'failed':
time.sleep(SLEEP_PERIOD)
sys.stdout.write('.')
sys.stdout.flush()
desc = dxpy.describe(job_id)
if desc['state'] == 'failed':
msg = RED('Error:') + ' Server failed to run.\n'
msg += 'You may want to check the job logs by running:'
msg += BOLD('dx watch {0}'.format(job_id))
err_exit(msg)
| poll_for_server_running | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_id):\n sys.stdout.write('Waiting for server in {0} to initialize ...'.format(\n job_id))\n sys.stdout.flush()\n desc = dxpy.describe(job_id... |
def x(cmd):
if platform == 'linux' or platform == 'linux2':
cmd = ['xdg-open', cmd]
elif platform == 'darwin':
cmd = ['open', cmd]
elif platform == 'win32':
cmd = ['start', cmd]
subprocess.check_call(cmd)
| multi_platform_open | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cmd):\n if platform == 'linux' or platform == 'linux2':\n cmd = ['xdg-open', cmd]\n elif platform == 'darwin':\n cmd = ['open', cmd]\n e... |
def x():
notebook_apps = dxpy.find_apps(name=NOTEBOOK_APP, all_versions=True)
versions = [str(dxpy.describe(app['id'])['version']) for app in
notebook_apps]
return versions
| get_notebook_app_versions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n notebook_apps = dxpy.find_apps(name=NOTEBOOK_APP, all_versions=True)\n versions = [str(dxpy.describe(app['id'])['version']) for app in\n note... |
def x(args, ssh_config_check):
ssh_config_check()
if args.only_check_config:
return
if args.version is not None:
executable = get_app_from_path('app-{0}/{1}'.format(NOTEBOOK_APP,
args.version))
if executable is not None and 'id' in executable:
executable = executable['id']
else:
msg = RED('Warning:'
) + ' Invalid notebook version: {0}\nValid versions are: '.format(
args.version)
msg += BOLD('{0}'.format(str(get_notebook_app_versions())))
err_exit(msg)
else:
executable = 'app-{0}'.format(NOTEBOOK_APP)
cmd = ['dx', 'run', executable, '-inotebook_type={0}'.format(args.
notebook_type)]
cmd += ['-iinput_files={0}'.format(f) for f in args.notebook_files]
cmd += ['-itimeout={0}'.format(args.timeout), '-y', '--brief',
'--allow-ssh', '--instance-type', args.instance_type]
if args.spark:
cmd += ['-iinstall_spark=true']
if args.snapshot:
cmd += ['-isnapshot={0}'.format(args.snapshot)]
job_id = subprocess.check_output(cmd).strip()
poll_for_server_running(job_id)
if args.notebook_type in {'jupyter', 'jupyter_lab', 'jupyter_notebook'}:
remote_port = 8888
setup_ssh_tunnel(job_id, args.port, remote_port)
if args.open_server:
multi_platform_open('http://localhost:{0}'.format(args.port))
print(
'A web browser should have opened to connect you to your notebook.'
)
print(
'If no browser appears, or if you need to reopen a browser at any point, you should be able to point your browser to http://localhost:{0}'
.format(args.port))
| run_notebook | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, ssh_config_check):\n ssh_config_check()\n if args.only_check_config:\n return\n if args.version is not None:\n executable = get_ap... |
def x(word, known_words):
NWORDS = _train(known_words)
candidates = _known([word], NWORDS) or _known(_edits1(word), NWORDS
) or _known_edits2(word, NWORDS) or [word]
return max(candidates, key=NWORDS.get)
| correct | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(word, known_words):\n NWORDS = _train(known_words)\n candidates = _known([word], NWORDS) or _known(_edits1(word), NWORDS\n ) or _known_edits2(word... |
def x(cls, job_config):
params = job_config.input_reader_params
shard_count = job_config.shard_count
query_spec = cls._get_query_spec(params)
if not property_range.should_shard_by_property_range(query_spec.filters):
return super(ModelDatastoreInputReader, cls).split_input(job_config)
p_range = property_range.PropertyRange(query_spec.filters, query_spec.
model_class_path)
p_ranges = p_range.split(shard_count)
if query_spec.ns:
ns_range = namespace_range.NamespaceRange(namespace_start=
query_spec.ns, namespace_end=query_spec.ns, _app=query_spec.app)
ns_ranges = [copy.copy(ns_range) for _ in p_ranges]
else:
ns_keys = namespace_range.get_namespace_keys(query_spec.app, cls.
MAX_NAMESPACES_FOR_KEY_SHARD + 1)
if not ns_keys:
return
if len(ns_keys) <= cls.MAX_NAMESPACES_FOR_KEY_SHARD:
ns_ranges = [namespace_range.NamespaceRange(_app=query_spec.app
) for _ in p_ranges]
else:
ns_ranges = namespace_range.NamespaceRange.split(n=shard_count,
contiguous=False, can_query=lambda : True, _app=query_spec.app)
p_ranges = [copy.copy(p_range) for _ in ns_ranges]
assert len(p_ranges) == len(ns_ranges)
iters = [db_iters.RangeIteratorFactory.create_property_range_iterator(p,
ns, query_spec) for p, ns in zip(p_ranges, ns_ranges)]
return [cls(i) for i in iters]
| split_input | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_config):\n params = job_config.input_reader_params\n shard_count = job_config.shard_count\n query_spec = cls._get_query_spec(params)\n if ... |
def x(value):
if ndb is not None and isinstance(value, ndb.Model):
return None
if getattr(value, '_populate_internal_entity', None):
return value._populate_internal_entity()
return value
| normalize_entity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n if ndb is not None and isinstance(value, ndb.Model):\n return None\n if getattr(value, '_populate_internal_entity', None):\n retu... |
def x(value):
if ndb is not None and isinstance(value, (ndb.Model, ndb.Key)):
return None
if getattr(value, 'key', None):
return value.key()
elif isinstance(value, basestring):
return datastore.Key(value)
else:
return value
| normalize_key | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n if ndb is not None and isinstance(value, (ndb.Model, ndb.Key)):\n return None\n if getattr(value, 'key', None):\n return value.ke... |
def x(self, item):
if self.should_flush():
self.flush()
self.items.append(item)
| append | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, item):\n if self.should_flush():\n self.flush()\n self.items.append(item)\n",
"role": "user"
},
{
"content": "append",
"role... |
def x(self, entity):
actual_entity = _normalize_entity(entity)
if actual_entity is None:
return self.ndb_put(entity)
self.puts.append(actual_entity)
| put | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity):\n actual_entity = _normalize_entity(entity)\n if actual_entity is None:\n return self.ndb_put(entity)\n self.puts.append(actual_... |
def x(self, entity):
assert ndb is not None and isinstance(entity, ndb.Model)
self.ndb_puts.append(entity)
| ndb_put | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity):\n assert ndb is not None and isinstance(entity, ndb.Model)\n self.ndb_puts.append(entity)\n",
"role": "user"
},
{
"content": "... |
def x(self, entity):
key = _normalize_key(entity)
if key is None:
return self.ndb_delete(entity)
self.deletes.append(key)
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity):\n key = _normalize_key(entity)\n if key is None:\n return self.ndb_delete(entity)\n self.deletes.append(key)\n",
"role": "us... |
def x(self, entity_or_key):
if ndb is not None and isinstance(entity_or_key, ndb.Model):
key = entity_or_key.key
else:
key = entity_or_key
self.ndb_deletes.append(key)
| ndb_delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity_or_key):\n if ndb is not None and isinstance(entity_or_key, ndb.Model):\n key = entity_or_key.key\n else:\n key = entity_or_ke... |
def x(self):
self.puts.flush()
self.deletes.flush()
self.ndb_puts.flush()
self.ndb_deletes.flush()
| flush | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.puts.flush()\n self.deletes.flush()\n self.ndb_puts.flush()\n self.ndb_deletes.flush()\n",
"role": "user"
},
{
"content":... |
def x(self, items, options):
datastore.Put(items, config=self._create_config(options))
| flush_puts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, items, options):\n datastore.Put(items, config=self._create_config(options))\n",
"role": "user"
},
{
"content": "flush_puts",
"role": ... |
def x(self, items, options):
datastore.Delete(items, config=self._create_config(options))
| flush_deletes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, items, options):\n datastore.Delete(items, config=self._create_config(options))\n",
"role": "user"
},
{
"content": "flush_deletes",
"r... |
def x(self, items, options):
assert ndb is not None
ndb.put_multi(items, config=self._create_config(options))
| flush_ndb_puts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, items, options):\n assert ndb is not None\n ndb.put_multi(items, config=self._create_config(options))\n",
"role": "user"
},
{
"content"... |
def x(self, items, options):
assert ndb is not None
ndb.delete_multi(items, config=self._create_config(options))
| flush_ndb_deletes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, items, options):\n assert ndb is not None\n ndb.delete_multi(items, config=self._create_config(options))\n",
"role": "user"
},
{
"conte... |
def x(self, options):
return datastore.CreateConfig(deadline=options['deadline'],
force_writes=self.force_writes)
| create_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, options):\n return datastore.CreateConfig(deadline=options['deadline'],\n force_writes=self.force_writes)\n",
"role": "user"
},
{
"... |
def x(self, counter_name, delta=1):
self._shard_state.counters_map.increment(counter_name, delta)
| increment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, counter_name, delta=1):\n self._shard_state.counters_map.increment(counter_name, delta)\n",
"role": "user"
},
{
"content": "increment",
... |
def x(self):
cursor = self._cursor
if self._query is not None:
if isinstance(self._query, db.Query):
cursor = self._query.cursor()
else:
cursor = self._query.cursor_after()
if cursor is None or isinstance(cursor, basestring):
cursor_object = False
else:
cursor_object = True
cursor = cursor.to_websafe_string()
return {'property_range': self._property_range.to_json(), 'query_spec':
self._query_spec.to_json(), 'cursor': cursor, 'ns_range': self.
_ns_range.to_json_object(), 'name': self.__class__.__name__,
'cursor_object': cursor_object}
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cursor = self._cursor\n if self._query is not None:\n if isinstance(self._query, db.Query):\n cursor = self._query.cursor()\n ... |
def x(cls, json):
obj = cls(property_range.PropertyRange.from_json(json['property_range']
), namespace_range.NamespaceRange.from_json_object(json['ns_range']
), model.QuerySpec.from_json(json['query_spec']))
cursor = json['cursor']
if cursor and json['cursor_object']:
obj._cursor = datastore_query.Cursor.from_websafe_string(cursor)
else:
obj._cursor = cursor
return obj
| from_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json):\n obj = cls(property_range.PropertyRange.from_json(json['property_range']\n ), namespace_range.NamespaceRange.from_json_object(json['ns_r... |
def x(self):
json = {'name': self.__class__.__name__, 'num_ranges': len(self._iters)}
for i in xrange(len(self._iters)):
json_item = self._iters[i].to_json()
query_spec = json_item['query_spec']
item_name = json_item['name']
del json_item['query_spec']
del json_item['name']
json[str(i)] = json_item
json['query_spec'] = query_spec
json['item_name'] = item_name
return json
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n json = {'name': self.__class__.__name__, 'num_ranges': len(self._iters)}\n for i in xrange(len(self._iters)):\n json_item = self._iters[i... |
def x(cls, json):
num_ranges = int(json['num_ranges'])
query_spec = json['query_spec']
item_name = json['item_name']
p_range_iters = []
for i in xrange(num_ranges):
json_item = json[str(i)]
json_item['query_spec'] = query_spec
json_item['name'] = item_name
p_range_iters.append(_PropertyRangeModelIterator.from_json(json_item))
obj = cls(p_range_iters)
return obj
| from_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json):\n num_ranges = int(json['num_ranges'])\n query_spec = json['query_spec']\n item_name = json['item_name']\n p_range_iters = []\n for ... |
def x(self):
current_iter = None
if self._current_iter:
current_iter = self._current_iter.to_json()
return {'key_ranges': self._key_ranges.to_json(), 'query_spec': self.
_query_spec.to_json(), 'current_iter': current_iter,
'key_range_iter_cls': self._key_range_iter_cls.__name__, 'name':
self.__class__.__name__}
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n current_iter = None\n if self._current_iter:\n current_iter = self._current_iter.to_json()\n return {'key_ranges': self._key_ranges.to... |
def x(cls, json):
key_range_iter_cls = _KEY_RANGE_ITERATORS[json['key_range_iter_cls']]
obj = cls(key_ranges.KeyRangesFactory.from_json(json['key_ranges']),
model.QuerySpec.from_json(json['query_spec']), key_range_iter_cls)
current_iter = None
if json['current_iter']:
current_iter = key_range_iter_cls.from_json(json['current_iter'])
obj._current_iter = current_iter
return obj
| from_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json):\n key_range_iter_cls = _KEY_RANGE_ITERATORS[json['key_range_iter_cls']]\n obj = cls(key_ranges.KeyRangesFactory.from_json(json['key_ranges'])... |
def x(self):
cursor = self._get_cursor()
cursor_object = False
if cursor and isinstance(cursor, datastore_query.Cursor):
cursor = cursor.to_websafe_string()
cursor_object = True
return {'key_range': self._key_range.to_json(), 'query_spec': self.
_query_spec.to_json(), 'cursor': cursor, 'cursor_object': cursor_object
}
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cursor = self._get_cursor()\n cursor_object = False\n if cursor and isinstance(cursor, datastore_query.Cursor):\n cursor = cursor.to_w... |
def x(cls, json):
obj = cls(key_range.KeyRange.from_json(json['key_range']), model.
QuerySpec.from_json(json['query_spec']))
cursor = json['cursor']
if cursor and json['cursor_object']:
obj._cursor = datastore_query.Cursor.from_websafe_string(cursor)
else:
obj._cursor = cursor
return obj
| from_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json):\n obj = cls(key_range.KeyRange.from_json(json['key_range']), model.\n QuerySpec.from_json(json['query_spec']))\n cursor = json['cursor... |
def x(status_file=__file__):
checked = set()
yaml = _find_mapreduce_yaml(os.path.dirname(status_file), checked)
if not yaml:
yaml = _find_mapreduce_yaml(os.getcwd(), checked)
return yaml
| find_mapreduce_yaml | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(status_file=__file__):\n checked = set()\n yaml = _find_mapreduce_yaml(os.path.dirname(status_file), checked)\n if not yaml:\n yaml = _find_map... |
def x(start, checked):
dir = start
while dir not in checked:
checked.add(dir)
for mr_yaml_name in MR_YAML_NAMES:
yaml_path = os.path.join(dir, mr_yaml_name)
if os.path.exists(yaml_path):
return yaml_path
dir = os.path.dirname(dir)
return None
| find_mapreduce_yaml | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(start, checked):\n dir = start\n while dir not in checked:\n checked.add(dir)\n for mr_yaml_name in MR_YAML_NAMES:\n yaml_path =... |
def x(parse=parse_mapreduce_yaml):
mr_yaml_path = find_mapreduce_yaml()
if not mr_yaml_path:
raise errors.MissingYamlError()
mr_yaml_file = open(mr_yaml_path)
try:
return parse(mr_yaml_file.read())
finally:
mr_yaml_file.close()
| get_mapreduce_yaml | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(parse=parse_mapreduce_yaml):\n mr_yaml_path = find_mapreduce_yaml()\n if not mr_yaml_path:\n raise errors.MissingYamlError()\n mr_yaml_file = o... |
def x(mapreduce_yaml):
all_configs = []
for config in mapreduce_yaml.mapreduce:
out = {'name': config.name, 'mapper_input_reader': config.mapper.
input_reader, 'mapper_handler': config.mapper.handler}
if config.mapper.params_validator:
out['mapper_params_validator'] = config.mapper.params_validator
if config.mapper.params:
param_defaults = {}
for param in config.mapper.params:
param_defaults[param.name] = param.default or param.value
out['mapper_params'] = param_defaults
if config.params:
param_defaults = {}
for param in config.params:
param_defaults[param.name] = param.default or param.value
out['params'] = param_defaults
if config.mapper.output_writer:
out['mapper_output_writer'] = config.mapper.output_writer
all_configs.append(out)
return all_configs
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mapreduce_yaml):\n all_configs = []\n for config in mapreduce_yaml.mapreduce:\n out = {'name': config.name, 'mapper_input_reader': config.mapper.\... |
def x(cls, job_config):
if job_config.input_reader_cls != cls:
raise errors.BadReaderParamsError(
'Expect input reader class %r, got %r.' % (cls, job_config.
input_reader_cls))
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_config):\n if job_config.input_reader_cls != cls:\n raise errors.BadReaderParamsError(\n 'Expect input reader class %r, got %r.' ... |
def x(records):
ctx = context.get()
l = len(records)
key_records = [None] * l
logging.debug('Parsing')
for i in range(l):
proto = kv_pb.KeyValue()
proto.ParseFromString(records[i])
key_records[i] = proto.key(), records[i]
logging.debug('Sorting')
key_records.sort(cmp=_compare_keys)
logging.debug('Writing')
mapper_spec = ctx.mapreduce_spec.mapper
params = input_readers._get_params(mapper_spec)
bucket_name = params.get('bucket_name')
filename = (ctx.mapreduce_spec.name + '/' + ctx.mapreduce_id +
'/output-' + ctx.shard_id + '-' + str(int(time.time())))
full_filename = '/%s/%s' % (bucket_name, filename)
filehandle = cloudstorage.open(full_filename, mode='w')
with output_writers.GCSRecordsPool(filehandle, ctx=ctx) as pool:
for key_record in key_records:
pool.append(key_record[1])
logging.debug('Finalizing')
filehandle.close()
entity = _OutputFile(key_name=full_filename, parent=_OutputFile.
get_root_key(ctx.mapreduce_id))
entity.put()
| sort_records_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(records):\n ctx = context.get()\n l = len(records)\n key_records = [None] * l\n logging.debug('Parsing')\n for i in range(l):\n proto = k... |
def x(key, values, partial):
proto = kv_pb.KeyValues()
proto.set_key(key)
proto.value_list().extend(values)
yield proto.Encode()
| merge_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key, values, partial):\n proto = kv_pb.KeyValues()\n proto.set_key(key)\n proto.value_list().extend(values)\n yield proto.Encode()\n",
"role": ... |
def x(binary_record):
proto = kv_pb.KeyValue()
proto.ParseFromString(binary_record)
yield proto.key(), proto.value()
| hashing_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(binary_record):\n proto = kv_pb.KeyValue()\n proto.ParseFromString(binary_record)\n yield proto.key(), proto.value()\n",
"role": "user"
},
{
... |
def x(cls, mapper_spec):
filelists = mapper_spec.params[cls.FILES_PARAM]
max_values_count = mapper_spec.params.get(cls.MAX_VALUES_COUNT_PARAM, -1)
max_values_size = mapper_spec.params.get(cls.MAX_VALUES_SIZE_PARAM, -1)
return [cls([0] * len(files), max_values_count, max_values_size) for
files in filelists]
| split_input | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapper_spec):\n filelists = mapper_spec.params[cls.FILES_PARAM]\n max_values_count = mapper_spec.params.get(cls.MAX_VALUES_COUNT_PARAM, -1)\n max... |
def x(cls, mapper_spec):
if mapper_spec.input_reader_class() != cls:
raise errors.BadReaderParamsError('Input reader class mismatch')
params = mapper_spec.params
if cls.FILES_PARAM not in params:
raise errors.BadReaderParamsError('Missing files parameter.')
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapper_spec):\n if mapper_spec.input_reader_class() != cls:\n raise errors.BadReaderParamsError('Input reader class mismatch')\n params = map... |
def x(cls, mapper_spec):
if mapper_spec.output_writer_class() != cls:
raise errors.BadWriterParamsError('Output writer class mismatch')
params = output_writers._get_params(mapper_spec)
if cls.BUCKET_NAME_PARAM not in params:
raise errors.BadWriterParamsError(
'%s is required for the _HashingGCSOutputWriter' % cls.
BUCKET_NAME_PARAM)
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapper_spec):\n if mapper_spec.output_writer_class() != cls:\n raise errors.BadWriterParamsError('Output writer class mismatch')\n params = o... |
def x(self):
for pool in self._pools:
if pool is not None:
pool.flush(True)
return {'filehandles': pickle.dumps(self._filehandles)}
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for pool in self._pools:\n if pool is not None:\n pool.flush(True)\n return {'filehandles': pickle.dumps(self._filehandles)}\n... |
def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None):
mapper_spec = mr_spec.mapper
params = output_writers._get_params(mapper_spec)
bucket_name = params.get(cls.BUCKET_NAME_PARAM)
shards = mapper_spec.shard_count
filehandles = []
filename = mr_spec.name + '/' + mr_spec.mapreduce_id + '/shard-' + str(
shard_number) + '-bucket-'
for i in range(shards):
full_filename = '/%s/%s%d' % (bucket_name, filename, i)
filehandles.append(cloudstorage.open(full_filename, mode='w'))
return cls(filehandles)
| create | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None):\n mapper_spec = mr_spec.mapper\n params = output_writers._get_params(mapper_spec)\n bu... |
def x(cls, mapreduce_state):
shards = mapreduce_state.mapreduce_spec.mapper.shard_count
filenames = []
for _ in range(shards):
filenames.append([None] * shards)
shard_states = model.ShardState.find_all_by_mapreduce_state(mapreduce_state
)
for x, shard_state in enumerate(shard_states):
shard_filenames = shard_state.writer_state['shard_filenames']
for y in range(shards):
filenames[y][x] = shard_filenames[y]
return filenames
| get_filenames | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_state):\n shards = mapreduce_state.mapreduce_spec.mapper.shard_count\n filenames = []\n for _ in range(shards):\n filenames.appe... |
def x(self, ctx, shard_state):
filenames = []
for filehandle in self._filehandles:
filenames.append(filehandle.name)
filehandle.close()
shard_state.writer_state = {'shard_filenames': filenames}
| finalize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ctx, shard_state):\n filenames = []\n for filehandle in self._filehandles:\n filenames.append(filehandle.name)\n filehandle.close()\n... |
def x(self, data):
ctx = context.get()
if len(data) != 2:
logging.error('Got bad tuple of length %d (2-tuple expected): %s',
len(data), data)
try:
key = str(data[0])
value = str(data[1])
except TypeError:
logging.error('Expecting a tuple, but got %s: %s', data.__class__.
__name__, data)
file_index = key.__hash__() % len(self._filehandles)
pool = self._pools[file_index]
if pool is None:
filehandle = self._filehandles[file_index]
pool = output_writers.GCSRecordsPool(filehandle=filehandle, ctx=ctx)
self._pools[file_index] = pool
proto = kv_pb.KeyValue()
proto.set_key(key)
proto.set_value(value)
pool.append(proto.Encode())
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n ctx = context.get()\n if len(data) != 2:\n logging.error('Got bad tuple of length %d (2-tuple expected): %s',\n len(data... |
def x(self):
reader_params = self.input_reader_cls.params_to_json(self.
input_reader_params)
return {'input_reader': reader_params, 'output_writer': self.
output_writer_params}
| get_mapper_params | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n reader_params = self.input_reader_cls.params_to_json(self.\n input_reader_params)\n return {'input_reader': reader_params, 'output_writer... |
def x(self):
from mapreduce import model
return model.MapperSpec(handler_spec=util._obj_to_path(self.mapper),
input_reader_spec=util._obj_to_path(self.input_reader_cls), params=
self._get_mapper_params(), shard_count=self.shard_count,
output_writer_spec=util._obj_to_path(self.output_writer_cls))
| get_mapper_spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from mapreduce import model\n return model.MapperSpec(handler_spec=util._obj_to_path(self.mapper),\n input_reader_spec=util._obj_to_path(... |
def x(self):
return {'force_writes': self._force_writes, 'done_callback': self.
done_callback_url, 'user_params': self.user_params,
'shard_max_attempts': self.shard_max_attempts, 'task_max_attempts':
self._task_max_attempts, 'task_max_data_processing_attempts': self.
_task_max_data_processing_attempts, 'queue_name': self.queue_name,
'base_path': self._base_path, 'app_id': self._app, 'api_version':
self._api_version}
| get_mr_params | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return {'force_writes': self._force_writes, 'done_callback': self.\n done_callback_url, 'user_params': self.user_params,\n 'shard_max... |
def x(cls):
cfg = cls(_lenient=True)
mr_params = cfg._get_mr_params()
mr_params['api_version'] = 0
return mr_params
| get_default_mr_params | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls):\n cfg = cls(_lenient=True)\n mr_params = cfg._get_mr_params()\n mr_params['api_version'] = 0\n return mr_params\n",
"role": "user"
},
... |
def x(cls, mr_spec, queue_name):
mapper_spec = mr_spec.mapper
api_version = mr_spec.params.get('api_version', 0)
old_api = api_version == 0
input_reader_cls = mapper_spec.input_reader_class()
input_reader_params = input_readers._get_params(mapper_spec)
if issubclass(input_reader_cls, input_reader.InputReader):
input_reader_params = input_reader_cls.params_from_json(
input_reader_params)
output_writer_cls = mapper_spec.output_writer_class()
output_writer_params = output_writers._get_params(mapper_spec)
return cls(_lenient=old_api, job_name=mr_spec.name, job_id=mr_spec.
mapreduce_id, mapper=util.for_name(mapper_spec.handler_spec),
input_reader_cls=input_reader_cls, input_reader_params=
input_reader_params, output_writer_cls=output_writer_cls,
output_writer_params=output_writer_params, shard_count=mapper_spec.
shard_count, queue_name=queue_name, user_params=mr_spec.params.get(
'user_params'), shard_max_attempts=mr_spec.params.get(
'shard_max_attempts'), done_callback_url=mr_spec.params.get(
'done_callback'), _force_writes=mr_spec.params.get('force_writes'),
_base_path=mr_spec.params['base_path'], _task_max_attempts=mr_spec.
params.get('task_max_attempts'), _task_max_data_processing_attempts
=mr_spec.params.get('task_max_data_processing_attempts'),
_hooks_cls=util.for_name(mr_spec.hooks_class_name), _app=mr_spec.
params.get('app_id'), _api_version=api_version)
| to_map_job_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mr_spec, queue_name):\n mapper_spec = mr_spec.mapper\n api_version = mr_spec.params.get('api_version', 0)\n old_api = api_version == 0\n input... |
def x(self, record_type, data):
length = len(data)
crc = crc32c.crc_update(crc32c.CRC_INIT, [record_type])
crc = crc32c.crc_update(crc, data)
crc = crc32c.crc_finalize(crc)
self.__writer.write(struct.pack(_HEADER_FORMAT, _mask_crc(crc), length,
record_type))
self.__writer.write(data)
self.__position += _HEADER_LENGTH + length
| write_record | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, record_type, data):\n length = len(data)\n crc = crc32c.crc_update(crc32c.CRC_INIT, [record_type])\n crc = crc32c.crc_update(crc, data)\n crc... |
def x(self, data):
block_remaining = _BLOCK_SIZE - self.__position % _BLOCK_SIZE
if block_remaining < _HEADER_LENGTH:
self.__writer.write('\x00' * block_remaining)
self.__position += block_remaining
block_remaining = _BLOCK_SIZE
if block_remaining < len(data) + _HEADER_LENGTH:
first_chunk = data[:block_remaining - _HEADER_LENGTH]
self.__write_record(_RECORD_TYPE_FIRST, first_chunk)
data = data[len(first_chunk):]
while True:
block_remaining = _BLOCK_SIZE - self.__position % _BLOCK_SIZE
if block_remaining >= len(data) + _HEADER_LENGTH:
self.__write_record(_RECORD_TYPE_LAST, data)
break
else:
chunk = data[:block_remaining - _HEADER_LENGTH]
self.__write_record(_RECORD_TYPE_MIDDLE, chunk)
data = data[len(chunk):]
else:
self.__write_record(_RECORD_TYPE_FULL, data)
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n block_remaining = _BLOCK_SIZE - self.__position % _BLOCK_SIZE\n if block_remaining < _HEADER_LENGTH:\n self.__writer.write('\\x00' ... |
def x(self):
pad_length = _BLOCK_SIZE - self.__position % _BLOCK_SIZE
if pad_length and pad_length != _BLOCK_SIZE:
self.__writer.write('\x00' * pad_length)
self.__position += pad_length
| pad_block | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pad_length = _BLOCK_SIZE - self.__position % _BLOCK_SIZE\n if pad_length and pad_length != _BLOCK_SIZE:\n self.__writer.write('\\x00' * p... |
def x(self):
block_remaining = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE
if block_remaining < _HEADER_LENGTH:
return '', _RECORD_TYPE_NONE
header = self.__reader.read(_HEADER_LENGTH)
if len(header) != _HEADER_LENGTH:
raise EOFError('Read %s bytes instead of %s' % (len(header),
_HEADER_LENGTH))
masked_crc, length, record_type = struct.unpack(_HEADER_FORMAT, header)
crc = _unmask_crc(masked_crc)
if length + _HEADER_LENGTH > block_remaining:
raise errors.InvalidRecordError('Length is too big')
data = self.__reader.read(length)
if len(data) != length:
raise EOFError('Not enough data read. Expected: %s but got %s' % (
length, len(data)))
if record_type == _RECORD_TYPE_NONE:
return '', record_type
actual_crc = crc32c.crc_update(crc32c.CRC_INIT, [record_type])
actual_crc = crc32c.crc_update(actual_crc, data)
actual_crc = crc32c.crc_finalize(actual_crc)
if actual_crc != crc:
raise errors.InvalidRecordError('Data crc does not match')
return data, record_type
| try_read_record | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n block_remaining = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE\n if block_remaining < _HEADER_LENGTH:\n return '', _RECORD_TYPE_NONE\... |
def x(self):
pad_length = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE
if pad_length and pad_length != _BLOCK_SIZE:
data = self.__reader.read(pad_length)
if len(data) != pad_length:
raise EOFError('Read %d bytes instead of %d' % (len(data),
pad_length))
| sync | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pad_length = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE\n if pad_length and pad_length != _BLOCK_SIZE:\n data = self.__reader.read(... |
def x(self, job_name, handler_spec, input_reader_spec, output_writer_spec=
None, params=None, shards=None, base_path=None):
if shards is None:
shards = parameters.config.SHARD_COUNT
if base_path is None:
base_path = parameters.config.BASE_PATH
mapreduce_id = control.start_map(job_name, handler_spec,
input_reader_spec, params or {}, mapreduce_parameters={
'done_callback': self.get_callback_url(), 'done_callback_method':
'GET', 'pipeline_id': self.pipeline_id, 'base_path': base_path},
shard_count=shards, output_writer_spec=output_writer_spec,
queue_name=self.queue_name)
self.fill(self.outputs.job_id, mapreduce_id)
self.set_status(console_url='%s/detail?mapreduce_id=%s' % (base_path,
mapreduce_id))
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, job_name, handler_spec, input_reader_spec, output_writer_spec=\n None, params=None, shards=None, base_path=None):\n if shards is None:\n sha... |
def x(self):
if self.was_aborted:
return
mapreduce_id = self.outputs.job_id.value
mapreduce_state = model.MapreduceState.get_by_job_id(mapreduce_id)
if mapreduce_state.result_status != model.MapreduceState.RESULT_SUCCESS:
self.retry('Job %s had status %s' % (mapreduce_id, mapreduce_state.
result_status))
return
mapper_spec = mapreduce_state.mapreduce_spec.mapper
outputs = []
output_writer_class = mapper_spec.output_writer_class()
if (output_writer_class and mapreduce_state.result_status == model.
MapreduceState.RESULT_SUCCESS):
outputs = output_writer_class.get_filenames(mapreduce_state)
self.fill(self.outputs.result_status, mapreduce_state.result_status)
self.fill(self.outputs.counters, mapreduce_state.counters_map.to_dict())
self.complete(outputs)
| callback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.was_aborted:\n return\n mapreduce_id = self.outputs.job_id.value\n mapreduce_state = model.MapreduceState.get_by_job_id(mapred... |
def x(name, handler_spec, reader_spec, mapper_parameters, shard_count=None,
output_writer_spec=None, mapreduce_parameters=None, base_path=None,
queue_name=None, eta=None, countdown=None, hooks_class_name=None, _app=
None, in_xg_transaction=False):
if shard_count is None:
shard_count = parameters.config.SHARD_COUNT
if mapper_parameters:
mapper_parameters = dict(mapper_parameters)
mr_params = map_job.JobConfig._get_default_mr_params()
if mapreduce_parameters:
mr_params.update(mapreduce_parameters)
if base_path:
mr_params['base_path'] = base_path
mr_params['queue_name'] = util.get_queue_name(queue_name)
mapper_spec = model.MapperSpec(handler_spec, reader_spec,
mapper_parameters, shard_count, output_writer_spec=output_writer_spec)
if in_xg_transaction and not db.is_in_transaction():
logging.warning(
'Expects an opened xg transaction to start mapreduce when transactional is True.'
)
return handlers.StartJobHandler._start_map(name, mapper_spec, mr_params,
queue_name=mr_params['queue_name'], eta=eta, countdown=countdown,
hooks_class_name=hooks_class_name, _app=_app, in_xg_transaction=
in_xg_transaction)
| start_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name, handler_spec, reader_spec, mapper_parameters, shard_count=None,\n output_writer_spec=None, mapreduce_parameters=None, base_path=None,\n queue_name=... |
def x(self):
self.__update_state()
if self._state.active:
return self.RUNNING
else:
return self._state.result_status
| get_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.__update_state()\n if self._state.active:\n return self.RUNNING\n else:\n return self._state.result_status\n",
"role":... |
def x(self, counter_name, default=0):
self.__update_state()
return self._state.counters_map.get(counter_name, default)
| get_counter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, counter_name, default=0):\n self.__update_state()\n return self._state.counters_map.get(counter_name, default)\n",
"role": "user"
},
{
... |
def x(self):
assert self.SUCCESS == self.get_status()
ss = model.ShardState.find_all_by_mapreduce_state(self._state)
for s in ss:
yield iter(s.writer_state.get('outs', []))
| get_outputs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n assert self.SUCCESS == self.get_status()\n ss = model.ShardState.find_all_by_mapreduce_state(self._state)\n for s in ss:\n yield iter(... |
def x(cls, job_config, in_xg_transaction=False):
cls.__validate_job_config(job_config)
mapper_spec = job_config._get_mapper_spec()
mapreduce_params = job_config._get_mr_params()
mapreduce_spec = model.MapreduceSpec(job_config.job_name, job_config.
job_id, mapper_spec.to_json(), mapreduce_params, util._obj_to_path(
job_config._hooks_cls))
if in_xg_transaction:
propagation = db.MANDATORY
else:
propagation = db.INDEPENDENT
state = None
@db.transactional(propagation=propagation)
def x():
state = cls.__create_and_save_state(job_config, mapreduce_spec)
cls.__add_kickoff_task(job_config, mapreduce_spec)
return state
state = _txn()
return cls(state)
| submit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_config, in_xg_transaction=False):\n cls.__validate_job_config(job_config)\n mapper_spec = job_config._get_mapper_spec()\n mapreduce_params = ... |
def x(self):
if self._state.active:
self._state = self.__get_state_by_id(self.job_config.job_id)
| update_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._state.active:\n self._state = self.__get_state_by_id(self.job_config.job_id)\n",
"role": "user"
},
{
"content": "update... |
def x(cls, job_id):
state = model.MapreduceState.get_by_job_id(job_id)
if state is None:
raise ValueError('Job state for job %s is missing.' % job_id)
return state
| get_state_by_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_id):\n state = model.MapreduceState.get_by_job_id(job_id)\n if state is None:\n raise ValueError('Job state for job %s is missing.' % job... |
def x(cls, job_config, mapreduce_spec):
state = model.MapreduceState.create_new(job_config.job_id)
state.mapreduce_spec = mapreduce_spec
state.active = True
state.active_shards = 0
state.app_id = job_config._app
config = datastore_rpc.Configuration(force_writes=job_config._force_writes)
state.put(config=config)
return state
| create_and_save_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_config, mapreduce_spec):\n state = model.MapreduceState.create_new(job_config.job_id)\n state.mapreduce_spec = mapreduce_spec\n state.active ... |
def x(cls, job_config, mapreduce_spec):
params = {'mapreduce_id': job_config.job_id}
kickoff_task = taskqueue.Task(url=job_config._base_path +
'/kickoffjob_callback/' + job_config.job_id, headers=util.
_get_task_headers(job_config.job_id), params=params)
if job_config._hooks_cls:
hooks = job_config._hooks_cls(mapreduce_spec)
try:
hooks.enqueue_kickoff_task(kickoff_task, job_config.queue_name)
return
except NotImplementedError:
pass
kickoff_task.add(job_config.queue_name, transactional=True)
| add_kickoff_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_config, mapreduce_spec):\n params = {'mapreduce_id': job_config.job_id}\n kickoff_task = taskqueue.Task(url=job_config._base_path +\n '/k... |
def x(self):
_json = self.to_json()
try:
return json.dumps(_json, sort_keys=True, cls=JsonEncoder)
except:
logging.exception('Could not serialize JSON: %r', _json)
raise
| to_json_str | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n _json = self.to_json()\n try:\n return json.dumps(_json, sort_keys=True, cls=JsonEncoder)\n except:\n logging.exception('Could ... |
def x(cls, json_str):
return cls.from_json(json.loads(json_str, cls=JsonDecoder))
| from_json_str | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json_str):\n return cls.from_json(json.loads(json_str, cls=JsonDecoder))\n",
"role": "user"
},
{
"content": "from_json_str",
"role": "a... |
def x(self, model_instance):
value = super(JsonProperty, self).get_value_for_datastore(model_instance)
if not value:
return None
json_value = value
if not isinstance(value, dict):
json_value = value.to_json()
if not json_value:
return None
return datastore_types.Text(json.dumps(json_value, sort_keys=True, cls=
JsonEncoder))
| get_value_for_datastore | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_instance):\n value = super(JsonProperty, self).get_value_for_datastore(model_instance)\n if not value:\n return None\n json_value =... |
def x(self, value):
if value is None:
return None
_json = json.loads(value, cls=JsonDecoder)
if self.data_type == dict:
return _json
return self.data_type.from_json(_json)
| make_value_from_datastore | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n if value is None:\n return None\n _json = json.loads(value, cls=JsonDecoder)\n if self.data_type == dict:\n return _json... |
def x(self, value):
if value is not None and not isinstance(value, self.data_type):
raise datastore_errors.BadValueError(
'Property %s must be convertible to a %s instance (%s)' % (self
.name, self.data_type, value))
return super(JsonProperty, self).validate(value)
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n if value is not None and not isinstance(value, self.data_type):\n raise datastore_errors.BadValueError(\n 'Property %s mus... |
def x(self, queue_name, transactional=False):
task = self.to_task()
task.add(queue_name, transactional)
| add | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, queue_name, transactional=False):\n task = self.to_task()\n task.add(queue_name, transactional)\n",
"role": "user"
},
{
"content": "add... |
def x(self):
return taskqueue.Task(url=self.url, payload=self._payload, name=self.
name, eta=self.eta, countdown=self.countdown, headers=self._headers)
| to_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return taskqueue.Task(url=self.url, payload=self._payload, name=self.\n name, eta=self.eta, countdown=self.countdown, headers=self._headers)... |
def x(cls, request):
if request.headers.get(cls.PAYLOAD_VERSION_HEADER) != cls.PAYLOAD_VERSION:
raise DeprecationWarning(
'Task is generated by an older incompatible version of mapreduce. Please kill this job manually'
)
return cls._decode_payload(request.body)
| decode_payload | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request):\n if request.headers.get(cls.PAYLOAD_VERSION_HEADER) != cls.PAYLOAD_VERSION:\n raise DeprecationWarning(\n 'Task is generat... |
def x(self, counter_name, delta):
current_value = self.counters.get(counter_name, 0)
new_value = current_value + delta
self.counters[counter_name] = new_value
return new_value
| increment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, counter_name, delta):\n current_value = self.counters.get(counter_name, 0)\n new_value = current_value + delta\n self.counters[counter_name] = n... |
def x(self, counters_map):
for counter_name in counters_map.counters:
self.increment(counter_name, counters_map.counters[counter_name])
| add_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, counters_map):\n for counter_name in counters_map.counters:\n self.increment(counter_name, counters_map.counters[counter_name])\n",
"role":... |
def x(self, counters_map):
for counter_name in counters_map.counters:
self.increment(counter_name, -counters_map.counters[counter_name])
| sub_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, counters_map):\n for counter_name in counters_map.counters:\n self.increment(counter_name, -counters_map.counters[counter_name])\n",
"role"... |
def x(self):
result = {'mapper_handler_spec': self.handler_spec,
'mapper_input_reader': self.input_reader_spec, 'mapper_params':
self.params, 'mapper_shard_count': self.shard_count}
if self.output_writer_spec:
result['mapper_output_writer'] = self.output_writer_spec
return result
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n result = {'mapper_handler_spec': self.handler_spec,\n 'mapper_input_reader': self.input_reader_spec, 'mapper_params':\n self.params, ... |
def x(self):
if self.__hooks is None and self.hooks_class_name is not None:
hooks_class = util.for_name(self.hooks_class_name)
if not isinstance(hooks_class, type):
raise ValueError(
'hooks_class_name must refer to a class, got %s' % type(
hooks_class).__name__)
if not issubclass(hooks_class, hooks.Hooks):
raise ValueError(
'hooks_class_name must refer to a hooks.Hooks subclass')
self.__hooks = hooks_class(self)
return self.__hooks
| get_hooks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.__hooks is None and self.hooks_class_name is not None:\n hooks_class = util.for_name(self.hooks_class_name)\n if not isinstan... |
def x(self):
mapper_spec = self.mapper.to_json()
return {'name': self.name, 'mapreduce_id': self.mapreduce_id,
'mapper_spec': mapper_spec, 'params': self.params,
'hooks_class_name': self.hooks_class_name}
| to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n mapper_spec = self.mapper.to_json()\n return {'name': self.name, 'mapreduce_id': self.mapreduce_id,\n 'mapper_spec': mapper_spec, 'params... |
def x(cls, json):
mapreduce_spec = cls(json['name'], json['mapreduce_id'], json[
'mapper_spec'], json.get('params'), json.get('hooks_class_name'))
return mapreduce_spec
| from_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json):\n mapreduce_spec = cls(json['name'], json['mapreduce_id'], json[\n 'mapper_spec'], json.get('params'), json.get('hooks_class_name'))\n ... |
def x(cls, mr_id):
key = 'GAE-MR-spec: %s' % mr_id
spec_json = memcache.get(key)
if spec_json:
return cls.from_json(spec_json)
state = MapreduceState.get_by_job_id(mr_id)
spec = state.mapreduce_spec
spec_json = spec.to_json()
memcache.set(key, spec_json)
return spec
| get_mapreduce_spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mr_id):\n key = 'GAE-MR-spec: %s' % mr_id\n spec_json = memcache.get(key)\n if spec_json:\n return cls.from_json(spec_json)\n state = M... |
def x(cls, mapreduce_id):
return db.Key.from_path(cls.kind(), str(mapreduce_id))
| get_key_by_job_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_id):\n return db.Key.from_path(cls.kind(), str(mapreduce_id))\n",
"role": "user"
},
{
"content": "get_key_by_job_id",
"role":... |
def x(self, shards_processed, shards_status):
chart = google_chart_api.BarChart()
def x(status_to_filter):
return [(count if status == status_to_filter else 0) for count,
status in zip(shards_processed, shards_status)]
if shards_status:
chart.stacked = True
chart.AddBars(filter_status('unknown'), color='404040')
chart.AddBars(filter_status('success'), color='00ac42')
chart.AddBars(filter_status('running'), color='3636a9')
chart.AddBars(filter_status('aborted'), color='e29e24')
chart.AddBars(filter_status('failed'), color='f6350f')
else:
chart.AddBars(shards_processed)
shard_count = len(shards_processed)
if shard_count > 95:
pixels_per_shard = 700.0 / shard_count
bar_thickness = int(pixels_per_shard * 0.9)
chart.style = bar_chart.BarChartStyle(bar_thickness=bar_thickness,
bar_gap=0.1, use_fractional_gap_spacing=True)
if shards_processed and shard_count <= 95:
stride_length = max(1, shard_count / 16)
chart.bottom.labels = []
for x in xrange(shard_count):
if x % stride_length == 0 or x == shard_count - 1:
chart.bottom.labels.append(x)
else:
chart.bottom.labels.append('')
chart.left.labels = ['0', str(max(shards_processed))]
chart.left.min = 0
self.chart_width = min(700, max(300, shard_count * 20))
self.chart_url = chart.display.Url(self.chart_width, 200)
| set_processed_counts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, shards_processed, shards_status):\n chart = google_chart_api.BarChart()\n\n def x(status_to_filter):\n return [(count if status == status_to... |
def x(mapreduce_id=None, gettime=datetime.datetime.now):
if not mapreduce_id:
mapreduce_id = MapreduceState.new_mapreduce_id()
state = MapreduceState(key_name=mapreduce_id, last_poll_time=gettime())
state.set_processed_counts([], [])
return state
| create_new | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mapreduce_id=None, gettime=datetime.datetime.now):\n if not mapreduce_id:\n mapreduce_id = MapreduceState.new_mapreduce_id()\n state = MapreduceSt... |
def x(self, output_writer):
self.input_reader = self.initial_input_reader
self.slice_id = 0
self.retries += 1
self.output_writer = output_writer
self.handler = self.mapreduce_spec.mapper.handler
| reset_for_retry | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, output_writer):\n self.input_reader = self.initial_input_reader\n self.slice_id = 0\n self.retries += 1\n self.output_writer = output_writer\... |
def x(self, recovery_slice=False):
if recovery_slice:
self.slice_id += 2
self.input_reader = self.input_reader.from_json(self._input_reader_json
)
else:
self.slice_id += 1
| advance_for_next_slice | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, recovery_slice=False):\n if recovery_slice:\n self.slice_id += 2\n self.input_reader = self.input_reader.from_json(self._input_reader_js... |
def x(self):
result = {'mapreduce_spec': self.mapreduce_spec.to_json_str(),
'shard_id': self.shard_id, 'slice_id': str(self.slice_id),
'input_reader_state': self.input_reader.to_json_str(),
'initial_input_reader_state': self.initial_input_reader.to_json_str
(), 'retries': str(self.retries)}
if self.output_writer:
result['output_writer_state'] = self.output_writer.to_json_str()
serialized_handler = util.try_serialize_handler(self.handler)
if serialized_handler:
result['serialized_handler'] = serialized_handler
return result
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n result = {'mapreduce_spec': self.mapreduce_spec.to_json_str(),\n 'shard_id': self.shard_id, 'slice_id': str(self.slice_id),\n 'input_... |
def x(cls, request):
mapreduce_spec = MapreduceSpec.from_json_str(request.get('mapreduce_spec'))
mapper_spec = mapreduce_spec.mapper
input_reader_spec_dict = json.loads(request.get('input_reader_state'),
cls=json_util.JsonDecoder)
input_reader = mapper_spec.input_reader_class().from_json(
input_reader_spec_dict)
initial_input_reader_spec_dict = json.loads(request.get(
'initial_input_reader_state'), cls=json_util.JsonDecoder)
initial_input_reader = mapper_spec.input_reader_class().from_json(
initial_input_reader_spec_dict)
output_writer = None
if mapper_spec.output_writer_class():
output_writer = mapper_spec.output_writer_class().from_json(json.
loads(request.get('output_writer_state', '{}'), cls=json_util.
JsonDecoder))
assert isinstance(output_writer, mapper_spec.output_writer_class()
), '%s.from_json returned an instance of wrong class: %s' % (
mapper_spec.output_writer_class(), output_writer.__class__)
handler = util.try_deserialize_handler(request.get('serialized_handler'))
if not handler:
handler = mapreduce_spec.mapper.handler
return cls(mapreduce_spec.params['base_path'], mapreduce_spec, str(
request.get('shard_id')), int(request.get('slice_id')),
input_reader, initial_input_reader, output_writer=output_writer,
retries=int(request.get('retries')), handler=handler)
| from_request | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request):\n mapreduce_spec = MapreduceSpec.from_json_str(request.get('mapreduce_spec'))\n mapper_spec = mapreduce_spec.mapper\n input_reader_spec... |
def x(self):
self.retries += 1
self.last_work_item = ''
self.active = True
self.result_status = None
self.input_finished = False
self.counters_map = CountersMap()
self.slice_id = 0
self.slice_start_time = None
self.slice_request_id = None
self.slice_retries = 0
self.acquired_once = False
| reset_for_retry | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.retries += 1\n self.last_work_item = ''\n self.active = True\n self.result_status = None\n self.input_finished = False\n self.c... |
def x(self, recovery_slice=False):
self.slice_start_time = None
self.slice_request_id = None
self.slice_retries = 0
self.acquired_once = False
if recovery_slice:
self.slice_id += 2
else:
self.slice_id += 1
| advance_for_next_slice | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, recovery_slice=False):\n self.slice_start_time = None\n self.slice_request_id = None\n self.slice_retries = 0\n self.acquired_once = False\n ... |
def x(self, other_state):
for prop in self.properties().values():
setattr(self, prop.name, getattr(other_state, prop.name))
| copy_from | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, other_state):\n for prop in self.properties().values():\n setattr(self, prop.name, getattr(other_state, prop.name))\n",
"role": "user"
},... |
def x(cls, mapreduce_state):
keys = cls.calculate_keys_by_mapreduce_state(mapreduce_state)
i = 0
while i < len(keys):
@db.non_transactional
def x(i):
return db.get(keys[i:i + cls._MAX_STATES_IN_MEMORY])
states = no_tx_get(i)
for s in states:
i += 1
if s is not None:
yield s
| find_all_by_mapreduce_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_state):\n keys = cls.calculate_keys_by_mapreduce_state(mapreduce_state)\n i = 0\n while i < len(keys):\n\n @db.non_transactional... |
def x(cls, mapreduce_state):
if mapreduce_state is None:
return []
keys = []
for i in range(mapreduce_state.mapreduce_spec.mapper.shard_count):
shard_id = cls.shard_id_from_number(mapreduce_state.key().name(), i)
keys.append(cls.get_key_by_shard_id(shard_id))
return keys
| calculate_keys_by_mapreduce_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_state):\n if mapreduce_state is None:\n return []\n keys = []\n for i in range(mapreduce_state.mapreduce_spec.mapper.shard_count... |
def x(cls, mapreduce_id, shard_number):
shard_id = cls.shard_id_from_number(mapreduce_id, shard_number)
state = cls(key_name=shard_id, mapreduce_id=mapreduce_id)
return state
| create_new | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_id, shard_number):\n shard_id = cls.shard_id_from_number(mapreduce_id, shard_number)\n state = cls(key_name=shard_id, mapreduce_id=mapredu... |
def x(cls, mapreduce_id):
return db.Key.from_path(cls.kind(), '%s:%s' % (mapreduce_id, cls._KEY_NAME)
)
| get_key_by_job_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_id):\n return db.Key.from_path(cls.kind(), '%s:%s' % (mapreduce_id, cls._KEY_NAME)\n )\n",
"role": "user"
},
{
"content": ... |
def x(cls, mapreduce_id, **kwargs):
cls(key_name='%s:%s' % (mapreduce_id, cls._KEY_NAME), command=cls.ABORT
).put(**kwargs)
| abort | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, mapreduce_id, **kwargs):\n cls(key_name='%s:%s' % (mapreduce_id, cls._KEY_NAME), command=cls.ABORT\n ).put(**kwargs)\n",
"role": "user"
},... |
def x(cls, job_config):
super(DatastoreInputReader, cls).validate(job_config)
params = job_config.input_reader_params
entity_kind = params[cls.ENTITY_KIND_PARAM]
if '.' in entity_kind:
logging.warning(
'. detected in entity kind %s specified for reader %s.Assuming entity kind contains the dot.'
, entity_kind, cls.__name__)
if cls.FILTERS_PARAM in params:
filters = params[cls.FILTERS_PARAM]
for f in filters:
if f[1] != '=':
raise errors.BadReaderParamsError(
'Only equality filters are supported: %s', f)
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, job_config):\n super(DatastoreInputReader, cls).validate(job_config)\n params = job_config.input_reader_params\n entity_kind = params[cls.ENTITY_... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.