source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, duration=None, preauthenticated=False, filename=None, project= None, **kwargs): args = {'preauthenticated': preauthenticated} if duration is not None: args['duration'] = duration if filename is not None: args['filename'] = filename if project is None and 'DX_JOB_ID' not in os.environ: project_from_handler = self.get_proj_id() if project_from_handler and object_exists_in_project(self.get_id(), project_from_handler): project = project_from_handler if project is not None and project is not DXFile.NO_PROJECT_HINT: args['project'] = project if '_DX_DUMP_BILLED_PROJECT' in os.environ: with open(os.environ['_DX_DUMP_BILLED_PROJECT'], 'w') as fd: if project is not None and project != DXFile.NO_PROJECT_HINT: fd.write(project) with self._url_download_mutex: if (self._download_url is None or self._download_url_expires < time .time()): if 'timeout' not in kwargs: kwargs['timeout'] = FILE_REQUEST_TIMEOUT resp = dxpy.api.file_download(self._dxid, args, **kwargs) self._download_url = resp['url'] self._download_url_headers = _validate_headers(resp.get( 'headers', {})) if preauthenticated: self._download_url_expires = resp['expires'] / 1000 - 60 else: self._download_url_expires = 32503680000 retval_download_url = self._download_url retval_download_url_headers = copy.copy(self._download_url_headers) return retval_download_url, retval_download_url_headers
get_download_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, duration=None, preauthenticated=False, filename=None, project=\n None, **kwargs):\n args = {'preauthenticated': preauthenticated}\n if duration ...
def x(self, length=None, use_compression=None, project=None, **kwargs): if self._file_length == None: desc = self.describe(**kwargs) if desc['state'] != 'closed': raise DXFileError( 'Cannot read from file until it is in the closed state') self._file_length = int(desc['size']) get_first_chunk_sequentially = (self._file_length > 128 * 1024 and self ._pos == 0 and dxpy.JOB_ID) if self._pos == self._file_length: return b'' if length == None or length > self._file_length - self._pos: length = self._file_length - self._pos buf = self._read_buf buf_remaining_bytes = dxpy.utils.string_buffer_length(buf) - buf.tell() if length <= buf_remaining_bytes: self._pos += length return buf.read(length) else: orig_buf_pos = buf.tell() orig_file_pos = self._pos buf.seek(0, os.SEEK_END) self._pos += buf_remaining_bytes while self._pos < orig_file_pos + length: remaining_len = orig_file_pos + length - self._pos if self._response_iterator is None: self._request_iterator = self._generate_read_requests(start_pos =self._pos, project=project, **kwargs) content = self._next_response_content(get_first_chunk_sequentially =get_first_chunk_sequentially) if len(content) < remaining_len: buf.write(content) self._pos += len(content) else: buf.write(content[:remaining_len]) self._pos += remaining_len self._read_buf = BytesIO() self._read_buf.write(content[remaining_len:]) self._read_buf.seek(0) buf.seek(orig_buf_pos) return buf.read()
read2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, length=None, use_compression=None, project=None, **kwargs):\n if self._file_length == None:\n desc = self.describe(**kwargs)\n if desc['...
def x(job_id, local_port, remote_port): cmd = ['dx', 'ssh', '--suppress-running-check', job_id, '-o', 'StrictHostKeyChecking no'] cmd += ['-f', '-L', '{0}:localhost:{1}'.format(local_port, remote_port), '-N'] subprocess.check_call(cmd)
setup_ssh_tunnel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_id, local_port, remote_port):\n cmd = ['dx', 'ssh', '--suppress-running-check', job_id, '-o',\n 'StrictHostKeyChecking no']\n cmd += ['-f', '-...
def x(job_id): sys.stdout.write('Waiting for server in {0} to initialize ...'.format( job_id)) sys.stdout.flush() desc = dxpy.describe(job_id) while SERVER_READY_TAG not in desc['tags'] and desc['state'] != 'failed': time.sleep(SLEEP_PERIOD) sys.stdout.write('.') sys.stdout.flush() desc = dxpy.describe(job_id) if desc['state'] == 'failed': msg = RED('Error:') + ' Server failed to run.\n' msg += 'You may want to check the job logs by running:' msg += BOLD('dx watch {0}'.format(job_id)) err_exit(msg)
poll_for_server_running
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_id):\n sys.stdout.write('Waiting for server in {0} to initialize ...'.format(\n job_id))\n sys.stdout.flush()\n desc = dxpy.describe(job_id...
def x(cmd): if platform == 'linux' or platform == 'linux2': cmd = ['xdg-open', cmd] elif platform == 'darwin': cmd = ['open', cmd] elif platform == 'win32': cmd = ['start', cmd] subprocess.check_call(cmd)
multi_platform_open
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd):\n if platform == 'linux' or platform == 'linux2':\n cmd = ['xdg-open', cmd]\n elif platform == 'darwin':\n cmd = ['open', cmd]\n e...
def x(): notebook_apps = dxpy.find_apps(name=NOTEBOOK_APP, all_versions=True) versions = [str(dxpy.describe(app['id'])['version']) for app in notebook_apps] return versions
get_notebook_app_versions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n notebook_apps = dxpy.find_apps(name=NOTEBOOK_APP, all_versions=True)\n versions = [str(dxpy.describe(app['id'])['version']) for app in\n note...
def x(args, ssh_config_check): ssh_config_check() if args.only_check_config: return if args.version is not None: executable = get_app_from_path('app-{0}/{1}'.format(NOTEBOOK_APP, args.version)) if executable is not None and 'id' in executable: executable = executable['id'] else: msg = RED('Warning:' ) + ' Invalid notebook version: {0}\nValid versions are: '.format( args.version) msg += BOLD('{0}'.format(str(get_notebook_app_versions()))) err_exit(msg) else: executable = 'app-{0}'.format(NOTEBOOK_APP) cmd = ['dx', 'run', executable, '-inotebook_type={0}'.format(args. notebook_type)] cmd += ['-iinput_files={0}'.format(f) for f in args.notebook_files] cmd += ['-itimeout={0}'.format(args.timeout), '-y', '--brief', '--allow-ssh', '--instance-type', args.instance_type] if args.spark: cmd += ['-iinstall_spark=true'] if args.snapshot: cmd += ['-isnapshot={0}'.format(args.snapshot)] job_id = subprocess.check_output(cmd).strip() poll_for_server_running(job_id) if args.notebook_type in {'jupyter', 'jupyter_lab', 'jupyter_notebook'}: remote_port = 8888 setup_ssh_tunnel(job_id, args.port, remote_port) if args.open_server: multi_platform_open('http://localhost:{0}'.format(args.port)) print( 'A web browser should have opened to connect you to your notebook.' ) print( 'If no browser appears, or if you need to reopen a browser at any point, you should be able to point your browser to http://localhost:{0}' .format(args.port))
run_notebook
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args, ssh_config_check):\n ssh_config_check()\n if args.only_check_config:\n return\n if args.version is not None:\n executable = get_ap...
def x(word, known_words): NWORDS = _train(known_words) candidates = _known([word], NWORDS) or _known(_edits1(word), NWORDS ) or _known_edits2(word, NWORDS) or [word] return max(candidates, key=NWORDS.get)
correct
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(word, known_words):\n NWORDS = _train(known_words)\n candidates = _known([word], NWORDS) or _known(_edits1(word), NWORDS\n ) or _known_edits2(word...
def x(cls, job_config): params = job_config.input_reader_params shard_count = job_config.shard_count query_spec = cls._get_query_spec(params) if not property_range.should_shard_by_property_range(query_spec.filters): return super(ModelDatastoreInputReader, cls).split_input(job_config) p_range = property_range.PropertyRange(query_spec.filters, query_spec. model_class_path) p_ranges = p_range.split(shard_count) if query_spec.ns: ns_range = namespace_range.NamespaceRange(namespace_start= query_spec.ns, namespace_end=query_spec.ns, _app=query_spec.app) ns_ranges = [copy.copy(ns_range) for _ in p_ranges] else: ns_keys = namespace_range.get_namespace_keys(query_spec.app, cls. MAX_NAMESPACES_FOR_KEY_SHARD + 1) if not ns_keys: return if len(ns_keys) <= cls.MAX_NAMESPACES_FOR_KEY_SHARD: ns_ranges = [namespace_range.NamespaceRange(_app=query_spec.app ) for _ in p_ranges] else: ns_ranges = namespace_range.NamespaceRange.split(n=shard_count, contiguous=False, can_query=lambda : True, _app=query_spec.app) p_ranges = [copy.copy(p_range) for _ in ns_ranges] assert len(p_ranges) == len(ns_ranges) iters = [db_iters.RangeIteratorFactory.create_property_range_iterator(p, ns, query_spec) for p, ns in zip(p_ranges, ns_ranges)] return [cls(i) for i in iters]
split_input
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config):\n params = job_config.input_reader_params\n shard_count = job_config.shard_count\n query_spec = cls._get_query_spec(params)\n if ...
def x(value): if ndb is not None and isinstance(value, ndb.Model): return None if getattr(value, '_populate_internal_entity', None): return value._populate_internal_entity() return value
normalize_entity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n if ndb is not None and isinstance(value, ndb.Model):\n return None\n if getattr(value, '_populate_internal_entity', None):\n retu...
def x(value): if ndb is not None and isinstance(value, (ndb.Model, ndb.Key)): return None if getattr(value, 'key', None): return value.key() elif isinstance(value, basestring): return datastore.Key(value) else: return value
normalize_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n if ndb is not None and isinstance(value, (ndb.Model, ndb.Key)):\n return None\n if getattr(value, 'key', None):\n return value.ke...
def x(self, item): if self.should_flush(): self.flush() self.items.append(item)
append
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, item):\n if self.should_flush():\n self.flush()\n self.items.append(item)\n", "role": "user" }, { "content": "append", "role...
def x(self, entity): actual_entity = _normalize_entity(entity) if actual_entity is None: return self.ndb_put(entity) self.puts.append(actual_entity)
put
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity):\n actual_entity = _normalize_entity(entity)\n if actual_entity is None:\n return self.ndb_put(entity)\n self.puts.append(actual_...
def x(self, entity): assert ndb is not None and isinstance(entity, ndb.Model) self.ndb_puts.append(entity)
ndb_put
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity):\n assert ndb is not None and isinstance(entity, ndb.Model)\n self.ndb_puts.append(entity)\n", "role": "user" }, { "content": "...
def x(self, entity): key = _normalize_key(entity) if key is None: return self.ndb_delete(entity) self.deletes.append(key)
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity):\n key = _normalize_key(entity)\n if key is None:\n return self.ndb_delete(entity)\n self.deletes.append(key)\n", "role": "us...
def x(self, entity_or_key): if ndb is not None and isinstance(entity_or_key, ndb.Model): key = entity_or_key.key else: key = entity_or_key self.ndb_deletes.append(key)
ndb_delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity_or_key):\n if ndb is not None and isinstance(entity_or_key, ndb.Model):\n key = entity_or_key.key\n else:\n key = entity_or_ke...
def x(self): self.puts.flush() self.deletes.flush() self.ndb_puts.flush() self.ndb_deletes.flush()
flush
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.puts.flush()\n self.deletes.flush()\n self.ndb_puts.flush()\n self.ndb_deletes.flush()\n", "role": "user" }, { "content":...
def x(self, items, options): datastore.Put(items, config=self._create_config(options))
flush_puts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, items, options):\n datastore.Put(items, config=self._create_config(options))\n", "role": "user" }, { "content": "flush_puts", "role": ...
def x(self, items, options): datastore.Delete(items, config=self._create_config(options))
flush_deletes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, items, options):\n datastore.Delete(items, config=self._create_config(options))\n", "role": "user" }, { "content": "flush_deletes", "r...
def x(self, items, options): assert ndb is not None ndb.put_multi(items, config=self._create_config(options))
flush_ndb_puts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, items, options):\n assert ndb is not None\n ndb.put_multi(items, config=self._create_config(options))\n", "role": "user" }, { "content"...
def x(self, items, options): assert ndb is not None ndb.delete_multi(items, config=self._create_config(options))
flush_ndb_deletes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, items, options):\n assert ndb is not None\n ndb.delete_multi(items, config=self._create_config(options))\n", "role": "user" }, { "conte...
def x(self, options): return datastore.CreateConfig(deadline=options['deadline'], force_writes=self.force_writes)
create_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, options):\n return datastore.CreateConfig(deadline=options['deadline'],\n force_writes=self.force_writes)\n", "role": "user" }, { "...
def x(self, counter_name, delta=1): self._shard_state.counters_map.increment(counter_name, delta)
increment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, counter_name, delta=1):\n self._shard_state.counters_map.increment(counter_name, delta)\n", "role": "user" }, { "content": "increment", ...
def x(self): cursor = self._cursor if self._query is not None: if isinstance(self._query, db.Query): cursor = self._query.cursor() else: cursor = self._query.cursor_after() if cursor is None or isinstance(cursor, basestring): cursor_object = False else: cursor_object = True cursor = cursor.to_websafe_string() return {'property_range': self._property_range.to_json(), 'query_spec': self._query_spec.to_json(), 'cursor': cursor, 'ns_range': self. _ns_range.to_json_object(), 'name': self.__class__.__name__, 'cursor_object': cursor_object}
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cursor = self._cursor\n if self._query is not None:\n if isinstance(self._query, db.Query):\n cursor = self._query.cursor()\n ...
def x(cls, json): obj = cls(property_range.PropertyRange.from_json(json['property_range'] ), namespace_range.NamespaceRange.from_json_object(json['ns_range'] ), model.QuerySpec.from_json(json['query_spec'])) cursor = json['cursor'] if cursor and json['cursor_object']: obj._cursor = datastore_query.Cursor.from_websafe_string(cursor) else: obj._cursor = cursor return obj
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json):\n obj = cls(property_range.PropertyRange.from_json(json['property_range']\n ), namespace_range.NamespaceRange.from_json_object(json['ns_r...
def x(self): json = {'name': self.__class__.__name__, 'num_ranges': len(self._iters)} for i in xrange(len(self._iters)): json_item = self._iters[i].to_json() query_spec = json_item['query_spec'] item_name = json_item['name'] del json_item['query_spec'] del json_item['name'] json[str(i)] = json_item json['query_spec'] = query_spec json['item_name'] = item_name return json
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n json = {'name': self.__class__.__name__, 'num_ranges': len(self._iters)}\n for i in xrange(len(self._iters)):\n json_item = self._iters[i...
def x(cls, json): num_ranges = int(json['num_ranges']) query_spec = json['query_spec'] item_name = json['item_name'] p_range_iters = [] for i in xrange(num_ranges): json_item = json[str(i)] json_item['query_spec'] = query_spec json_item['name'] = item_name p_range_iters.append(_PropertyRangeModelIterator.from_json(json_item)) obj = cls(p_range_iters) return obj
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json):\n num_ranges = int(json['num_ranges'])\n query_spec = json['query_spec']\n item_name = json['item_name']\n p_range_iters = []\n for ...
def x(self): current_iter = None if self._current_iter: current_iter = self._current_iter.to_json() return {'key_ranges': self._key_ranges.to_json(), 'query_spec': self. _query_spec.to_json(), 'current_iter': current_iter, 'key_range_iter_cls': self._key_range_iter_cls.__name__, 'name': self.__class__.__name__}
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n current_iter = None\n if self._current_iter:\n current_iter = self._current_iter.to_json()\n return {'key_ranges': self._key_ranges.to...
def x(cls, json): key_range_iter_cls = _KEY_RANGE_ITERATORS[json['key_range_iter_cls']] obj = cls(key_ranges.KeyRangesFactory.from_json(json['key_ranges']), model.QuerySpec.from_json(json['query_spec']), key_range_iter_cls) current_iter = None if json['current_iter']: current_iter = key_range_iter_cls.from_json(json['current_iter']) obj._current_iter = current_iter return obj
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json):\n key_range_iter_cls = _KEY_RANGE_ITERATORS[json['key_range_iter_cls']]\n obj = cls(key_ranges.KeyRangesFactory.from_json(json['key_ranges'])...
def x(self): cursor = self._get_cursor() cursor_object = False if cursor and isinstance(cursor, datastore_query.Cursor): cursor = cursor.to_websafe_string() cursor_object = True return {'key_range': self._key_range.to_json(), 'query_spec': self. _query_spec.to_json(), 'cursor': cursor, 'cursor_object': cursor_object }
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cursor = self._get_cursor()\n cursor_object = False\n if cursor and isinstance(cursor, datastore_query.Cursor):\n cursor = cursor.to_w...
def x(cls, json): obj = cls(key_range.KeyRange.from_json(json['key_range']), model. QuerySpec.from_json(json['query_spec'])) cursor = json['cursor'] if cursor and json['cursor_object']: obj._cursor = datastore_query.Cursor.from_websafe_string(cursor) else: obj._cursor = cursor return obj
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json):\n obj = cls(key_range.KeyRange.from_json(json['key_range']), model.\n QuerySpec.from_json(json['query_spec']))\n cursor = json['cursor...
def x(status_file=__file__): checked = set() yaml = _find_mapreduce_yaml(os.path.dirname(status_file), checked) if not yaml: yaml = _find_mapreduce_yaml(os.getcwd(), checked) return yaml
find_mapreduce_yaml
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(status_file=__file__):\n checked = set()\n yaml = _find_mapreduce_yaml(os.path.dirname(status_file), checked)\n if not yaml:\n yaml = _find_map...
def x(start, checked): dir = start while dir not in checked: checked.add(dir) for mr_yaml_name in MR_YAML_NAMES: yaml_path = os.path.join(dir, mr_yaml_name) if os.path.exists(yaml_path): return yaml_path dir = os.path.dirname(dir) return None
find_mapreduce_yaml
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(start, checked):\n dir = start\n while dir not in checked:\n checked.add(dir)\n for mr_yaml_name in MR_YAML_NAMES:\n yaml_path =...
def x(parse=parse_mapreduce_yaml): mr_yaml_path = find_mapreduce_yaml() if not mr_yaml_path: raise errors.MissingYamlError() mr_yaml_file = open(mr_yaml_path) try: return parse(mr_yaml_file.read()) finally: mr_yaml_file.close()
get_mapreduce_yaml
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(parse=parse_mapreduce_yaml):\n mr_yaml_path = find_mapreduce_yaml()\n if not mr_yaml_path:\n raise errors.MissingYamlError()\n mr_yaml_file = o...
def x(mapreduce_yaml): all_configs = [] for config in mapreduce_yaml.mapreduce: out = {'name': config.name, 'mapper_input_reader': config.mapper. input_reader, 'mapper_handler': config.mapper.handler} if config.mapper.params_validator: out['mapper_params_validator'] = config.mapper.params_validator if config.mapper.params: param_defaults = {} for param in config.mapper.params: param_defaults[param.name] = param.default or param.value out['mapper_params'] = param_defaults if config.params: param_defaults = {} for param in config.params: param_defaults[param.name] = param.default or param.value out['params'] = param_defaults if config.mapper.output_writer: out['mapper_output_writer'] = config.mapper.output_writer all_configs.append(out) return all_configs
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mapreduce_yaml):\n all_configs = []\n for config in mapreduce_yaml.mapreduce:\n out = {'name': config.name, 'mapper_input_reader': config.mapper.\...
def x(cls, job_config): if job_config.input_reader_cls != cls: raise errors.BadReaderParamsError( 'Expect input reader class %r, got %r.' % (cls, job_config. input_reader_cls))
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config):\n if job_config.input_reader_cls != cls:\n raise errors.BadReaderParamsError(\n 'Expect input reader class %r, got %r.' ...
def x(records): ctx = context.get() l = len(records) key_records = [None] * l logging.debug('Parsing') for i in range(l): proto = kv_pb.KeyValue() proto.ParseFromString(records[i]) key_records[i] = proto.key(), records[i] logging.debug('Sorting') key_records.sort(cmp=_compare_keys) logging.debug('Writing') mapper_spec = ctx.mapreduce_spec.mapper params = input_readers._get_params(mapper_spec) bucket_name = params.get('bucket_name') filename = (ctx.mapreduce_spec.name + '/' + ctx.mapreduce_id + '/output-' + ctx.shard_id + '-' + str(int(time.time()))) full_filename = '/%s/%s' % (bucket_name, filename) filehandle = cloudstorage.open(full_filename, mode='w') with output_writers.GCSRecordsPool(filehandle, ctx=ctx) as pool: for key_record in key_records: pool.append(key_record[1]) logging.debug('Finalizing') filehandle.close() entity = _OutputFile(key_name=full_filename, parent=_OutputFile. get_root_key(ctx.mapreduce_id)) entity.put()
sort_records_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(records):\n ctx = context.get()\n l = len(records)\n key_records = [None] * l\n logging.debug('Parsing')\n for i in range(l):\n proto = k...
def x(key, values, partial): proto = kv_pb.KeyValues() proto.set_key(key) proto.value_list().extend(values) yield proto.Encode()
merge_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key, values, partial):\n proto = kv_pb.KeyValues()\n proto.set_key(key)\n proto.value_list().extend(values)\n yield proto.Encode()\n", "role": ...
def x(binary_record): proto = kv_pb.KeyValue() proto.ParseFromString(binary_record) yield proto.key(), proto.value()
hashing_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(binary_record):\n proto = kv_pb.KeyValue()\n proto.ParseFromString(binary_record)\n yield proto.key(), proto.value()\n", "role": "user" }, { ...
def x(cls, mapper_spec): filelists = mapper_spec.params[cls.FILES_PARAM] max_values_count = mapper_spec.params.get(cls.MAX_VALUES_COUNT_PARAM, -1) max_values_size = mapper_spec.params.get(cls.MAX_VALUES_SIZE_PARAM, -1) return [cls([0] * len(files), max_values_count, max_values_size) for files in filelists]
split_input
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapper_spec):\n filelists = mapper_spec.params[cls.FILES_PARAM]\n max_values_count = mapper_spec.params.get(cls.MAX_VALUES_COUNT_PARAM, -1)\n max...
def x(cls, mapper_spec): if mapper_spec.input_reader_class() != cls: raise errors.BadReaderParamsError('Input reader class mismatch') params = mapper_spec.params if cls.FILES_PARAM not in params: raise errors.BadReaderParamsError('Missing files parameter.')
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapper_spec):\n if mapper_spec.input_reader_class() != cls:\n raise errors.BadReaderParamsError('Input reader class mismatch')\n params = map...
def x(cls, mapper_spec): if mapper_spec.output_writer_class() != cls: raise errors.BadWriterParamsError('Output writer class mismatch') params = output_writers._get_params(mapper_spec) if cls.BUCKET_NAME_PARAM not in params: raise errors.BadWriterParamsError( '%s is required for the _HashingGCSOutputWriter' % cls. BUCKET_NAME_PARAM)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapper_spec):\n if mapper_spec.output_writer_class() != cls:\n raise errors.BadWriterParamsError('Output writer class mismatch')\n params = o...
def x(self): for pool in self._pools: if pool is not None: pool.flush(True) return {'filehandles': pickle.dumps(self._filehandles)}
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for pool in self._pools:\n if pool is not None:\n pool.flush(True)\n return {'filehandles': pickle.dumps(self._filehandles)}\n...
def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None): mapper_spec = mr_spec.mapper params = output_writers._get_params(mapper_spec) bucket_name = params.get(cls.BUCKET_NAME_PARAM) shards = mapper_spec.shard_count filehandles = [] filename = mr_spec.name + '/' + mr_spec.mapreduce_id + '/shard-' + str( shard_number) + '-bucket-' for i in range(shards): full_filename = '/%s/%s%d' % (bucket_name, filename, i) filehandles.append(cloudstorage.open(full_filename, mode='w')) return cls(filehandles)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None):\n mapper_spec = mr_spec.mapper\n params = output_writers._get_params(mapper_spec)\n bu...
def x(cls, mapreduce_state): shards = mapreduce_state.mapreduce_spec.mapper.shard_count filenames = [] for _ in range(shards): filenames.append([None] * shards) shard_states = model.ShardState.find_all_by_mapreduce_state(mapreduce_state ) for x, shard_state in enumerate(shard_states): shard_filenames = shard_state.writer_state['shard_filenames'] for y in range(shards): filenames[y][x] = shard_filenames[y] return filenames
get_filenames
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_state):\n shards = mapreduce_state.mapreduce_spec.mapper.shard_count\n filenames = []\n for _ in range(shards):\n filenames.appe...
def x(self, ctx, shard_state): filenames = [] for filehandle in self._filehandles: filenames.append(filehandle.name) filehandle.close() shard_state.writer_state = {'shard_filenames': filenames}
finalize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ctx, shard_state):\n filenames = []\n for filehandle in self._filehandles:\n filenames.append(filehandle.name)\n filehandle.close()\n...
def x(self, data): ctx = context.get() if len(data) != 2: logging.error('Got bad tuple of length %d (2-tuple expected): %s', len(data), data) try: key = str(data[0]) value = str(data[1]) except TypeError: logging.error('Expecting a tuple, but got %s: %s', data.__class__. __name__, data) file_index = key.__hash__() % len(self._filehandles) pool = self._pools[file_index] if pool is None: filehandle = self._filehandles[file_index] pool = output_writers.GCSRecordsPool(filehandle=filehandle, ctx=ctx) self._pools[file_index] = pool proto = kv_pb.KeyValue() proto.set_key(key) proto.set_value(value) pool.append(proto.Encode())
write
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n ctx = context.get()\n if len(data) != 2:\n logging.error('Got bad tuple of length %d (2-tuple expected): %s',\n len(data...
def x(self): reader_params = self.input_reader_cls.params_to_json(self. input_reader_params) return {'input_reader': reader_params, 'output_writer': self. output_writer_params}
get_mapper_params
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n reader_params = self.input_reader_cls.params_to_json(self.\n input_reader_params)\n return {'input_reader': reader_params, 'output_writer...
def x(self): from mapreduce import model return model.MapperSpec(handler_spec=util._obj_to_path(self.mapper), input_reader_spec=util._obj_to_path(self.input_reader_cls), params= self._get_mapper_params(), shard_count=self.shard_count, output_writer_spec=util._obj_to_path(self.output_writer_cls))
get_mapper_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from mapreduce import model\n return model.MapperSpec(handler_spec=util._obj_to_path(self.mapper),\n input_reader_spec=util._obj_to_path(...
def x(self): return {'force_writes': self._force_writes, 'done_callback': self. done_callback_url, 'user_params': self.user_params, 'shard_max_attempts': self.shard_max_attempts, 'task_max_attempts': self._task_max_attempts, 'task_max_data_processing_attempts': self. _task_max_data_processing_attempts, 'queue_name': self.queue_name, 'base_path': self._base_path, 'app_id': self._app, 'api_version': self._api_version}
get_mr_params
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {'force_writes': self._force_writes, 'done_callback': self.\n done_callback_url, 'user_params': self.user_params,\n 'shard_max...
def x(cls): cfg = cls(_lenient=True) mr_params = cfg._get_mr_params() mr_params['api_version'] = 0 return mr_params
get_default_mr_params
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls):\n cfg = cls(_lenient=True)\n mr_params = cfg._get_mr_params()\n mr_params['api_version'] = 0\n return mr_params\n", "role": "user" }, ...
def x(cls, mr_spec, queue_name): mapper_spec = mr_spec.mapper api_version = mr_spec.params.get('api_version', 0) old_api = api_version == 0 input_reader_cls = mapper_spec.input_reader_class() input_reader_params = input_readers._get_params(mapper_spec) if issubclass(input_reader_cls, input_reader.InputReader): input_reader_params = input_reader_cls.params_from_json( input_reader_params) output_writer_cls = mapper_spec.output_writer_class() output_writer_params = output_writers._get_params(mapper_spec) return cls(_lenient=old_api, job_name=mr_spec.name, job_id=mr_spec. mapreduce_id, mapper=util.for_name(mapper_spec.handler_spec), input_reader_cls=input_reader_cls, input_reader_params= input_reader_params, output_writer_cls=output_writer_cls, output_writer_params=output_writer_params, shard_count=mapper_spec. shard_count, queue_name=queue_name, user_params=mr_spec.params.get( 'user_params'), shard_max_attempts=mr_spec.params.get( 'shard_max_attempts'), done_callback_url=mr_spec.params.get( 'done_callback'), _force_writes=mr_spec.params.get('force_writes'), _base_path=mr_spec.params['base_path'], _task_max_attempts=mr_spec. params.get('task_max_attempts'), _task_max_data_processing_attempts =mr_spec.params.get('task_max_data_processing_attempts'), _hooks_cls=util.for_name(mr_spec.hooks_class_name), _app=mr_spec. params.get('app_id'), _api_version=api_version)
to_map_job_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mr_spec, queue_name):\n mapper_spec = mr_spec.mapper\n api_version = mr_spec.params.get('api_version', 0)\n old_api = api_version == 0\n input...
def x(self, record_type, data): length = len(data) crc = crc32c.crc_update(crc32c.CRC_INIT, [record_type]) crc = crc32c.crc_update(crc, data) crc = crc32c.crc_finalize(crc) self.__writer.write(struct.pack(_HEADER_FORMAT, _mask_crc(crc), length, record_type)) self.__writer.write(data) self.__position += _HEADER_LENGTH + length
write_record
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, record_type, data):\n length = len(data)\n crc = crc32c.crc_update(crc32c.CRC_INIT, [record_type])\n crc = crc32c.crc_update(crc, data)\n crc...
def x(self, data): block_remaining = _BLOCK_SIZE - self.__position % _BLOCK_SIZE if block_remaining < _HEADER_LENGTH: self.__writer.write('\x00' * block_remaining) self.__position += block_remaining block_remaining = _BLOCK_SIZE if block_remaining < len(data) + _HEADER_LENGTH: first_chunk = data[:block_remaining - _HEADER_LENGTH] self.__write_record(_RECORD_TYPE_FIRST, first_chunk) data = data[len(first_chunk):] while True: block_remaining = _BLOCK_SIZE - self.__position % _BLOCK_SIZE if block_remaining >= len(data) + _HEADER_LENGTH: self.__write_record(_RECORD_TYPE_LAST, data) break else: chunk = data[:block_remaining - _HEADER_LENGTH] self.__write_record(_RECORD_TYPE_MIDDLE, chunk) data = data[len(chunk):] else: self.__write_record(_RECORD_TYPE_FULL, data)
write
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n block_remaining = _BLOCK_SIZE - self.__position % _BLOCK_SIZE\n if block_remaining < _HEADER_LENGTH:\n self.__writer.write('\\x00' ...
def x(self): pad_length = _BLOCK_SIZE - self.__position % _BLOCK_SIZE if pad_length and pad_length != _BLOCK_SIZE: self.__writer.write('\x00' * pad_length) self.__position += pad_length
pad_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n pad_length = _BLOCK_SIZE - self.__position % _BLOCK_SIZE\n if pad_length and pad_length != _BLOCK_SIZE:\n self.__writer.write('\\x00' * p...
def x(self): block_remaining = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE if block_remaining < _HEADER_LENGTH: return '', _RECORD_TYPE_NONE header = self.__reader.read(_HEADER_LENGTH) if len(header) != _HEADER_LENGTH: raise EOFError('Read %s bytes instead of %s' % (len(header), _HEADER_LENGTH)) masked_crc, length, record_type = struct.unpack(_HEADER_FORMAT, header) crc = _unmask_crc(masked_crc) if length + _HEADER_LENGTH > block_remaining: raise errors.InvalidRecordError('Length is too big') data = self.__reader.read(length) if len(data) != length: raise EOFError('Not enough data read. Expected: %s but got %s' % ( length, len(data))) if record_type == _RECORD_TYPE_NONE: return '', record_type actual_crc = crc32c.crc_update(crc32c.CRC_INIT, [record_type]) actual_crc = crc32c.crc_update(actual_crc, data) actual_crc = crc32c.crc_finalize(actual_crc) if actual_crc != crc: raise errors.InvalidRecordError('Data crc does not match') return data, record_type
try_read_record
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n block_remaining = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE\n if block_remaining < _HEADER_LENGTH:\n return '', _RECORD_TYPE_NONE\...
def x(self): pad_length = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE if pad_length and pad_length != _BLOCK_SIZE: data = self.__reader.read(pad_length) if len(data) != pad_length: raise EOFError('Read %d bytes instead of %d' % (len(data), pad_length))
sync
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n pad_length = _BLOCK_SIZE - self.__reader.tell() % _BLOCK_SIZE\n if pad_length and pad_length != _BLOCK_SIZE:\n data = self.__reader.read(...
def x(self, job_name, handler_spec, input_reader_spec, output_writer_spec= None, params=None, shards=None, base_path=None): if shards is None: shards = parameters.config.SHARD_COUNT if base_path is None: base_path = parameters.config.BASE_PATH mapreduce_id = control.start_map(job_name, handler_spec, input_reader_spec, params or {}, mapreduce_parameters={ 'done_callback': self.get_callback_url(), 'done_callback_method': 'GET', 'pipeline_id': self.pipeline_id, 'base_path': base_path}, shard_count=shards, output_writer_spec=output_writer_spec, queue_name=self.queue_name) self.fill(self.outputs.job_id, mapreduce_id) self.set_status(console_url='%s/detail?mapreduce_id=%s' % (base_path, mapreduce_id))
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, job_name, handler_spec, input_reader_spec, output_writer_spec=\n None, params=None, shards=None, base_path=None):\n if shards is None:\n sha...
def x(self): if self.was_aborted: return mapreduce_id = self.outputs.job_id.value mapreduce_state = model.MapreduceState.get_by_job_id(mapreduce_id) if mapreduce_state.result_status != model.MapreduceState.RESULT_SUCCESS: self.retry('Job %s had status %s' % (mapreduce_id, mapreduce_state. result_status)) return mapper_spec = mapreduce_state.mapreduce_spec.mapper outputs = [] output_writer_class = mapper_spec.output_writer_class() if (output_writer_class and mapreduce_state.result_status == model. MapreduceState.RESULT_SUCCESS): outputs = output_writer_class.get_filenames(mapreduce_state) self.fill(self.outputs.result_status, mapreduce_state.result_status) self.fill(self.outputs.counters, mapreduce_state.counters_map.to_dict()) self.complete(outputs)
callback
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.was_aborted:\n return\n mapreduce_id = self.outputs.job_id.value\n mapreduce_state = model.MapreduceState.get_by_job_id(mapred...
def x(name, handler_spec, reader_spec, mapper_parameters, shard_count=None, output_writer_spec=None, mapreduce_parameters=None, base_path=None, queue_name=None, eta=None, countdown=None, hooks_class_name=None, _app= None, in_xg_transaction=False): if shard_count is None: shard_count = parameters.config.SHARD_COUNT if mapper_parameters: mapper_parameters = dict(mapper_parameters) mr_params = map_job.JobConfig._get_default_mr_params() if mapreduce_parameters: mr_params.update(mapreduce_parameters) if base_path: mr_params['base_path'] = base_path mr_params['queue_name'] = util.get_queue_name(queue_name) mapper_spec = model.MapperSpec(handler_spec, reader_spec, mapper_parameters, shard_count, output_writer_spec=output_writer_spec) if in_xg_transaction and not db.is_in_transaction(): logging.warning( 'Expects an opened xg transaction to start mapreduce when transactional is True.' ) return handlers.StartJobHandler._start_map(name, mapper_spec, mr_params, queue_name=mr_params['queue_name'], eta=eta, countdown=countdown, hooks_class_name=hooks_class_name, _app=_app, in_xg_transaction= in_xg_transaction)
start_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name, handler_spec, reader_spec, mapper_parameters, shard_count=None,\n output_writer_spec=None, mapreduce_parameters=None, base_path=None,\n queue_name=...
def x(self): self.__update_state() if self._state.active: return self.RUNNING else: return self._state.result_status
get_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.__update_state()\n if self._state.active:\n return self.RUNNING\n else:\n return self._state.result_status\n", "role":...
def x(self, counter_name, default=0): self.__update_state() return self._state.counters_map.get(counter_name, default)
get_counter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, counter_name, default=0):\n self.__update_state()\n return self._state.counters_map.get(counter_name, default)\n", "role": "user" }, { ...
def x(self): assert self.SUCCESS == self.get_status() ss = model.ShardState.find_all_by_mapreduce_state(self._state) for s in ss: yield iter(s.writer_state.get('outs', []))
get_outputs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n assert self.SUCCESS == self.get_status()\n ss = model.ShardState.find_all_by_mapreduce_state(self._state)\n for s in ss:\n yield iter(...
def x(cls, job_config, in_xg_transaction=False): cls.__validate_job_config(job_config) mapper_spec = job_config._get_mapper_spec() mapreduce_params = job_config._get_mr_params() mapreduce_spec = model.MapreduceSpec(job_config.job_name, job_config. job_id, mapper_spec.to_json(), mapreduce_params, util._obj_to_path( job_config._hooks_cls)) if in_xg_transaction: propagation = db.MANDATORY else: propagation = db.INDEPENDENT state = None @db.transactional(propagation=propagation) def x(): state = cls.__create_and_save_state(job_config, mapreduce_spec) cls.__add_kickoff_task(job_config, mapreduce_spec) return state state = _txn() return cls(state)
submit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config, in_xg_transaction=False):\n cls.__validate_job_config(job_config)\n mapper_spec = job_config._get_mapper_spec()\n mapreduce_params = ...
def x(self): if self._state.active: self._state = self.__get_state_by_id(self.job_config.job_id)
update_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._state.active:\n self._state = self.__get_state_by_id(self.job_config.job_id)\n", "role": "user" }, { "content": "update...
def x(cls, job_id): state = model.MapreduceState.get_by_job_id(job_id) if state is None: raise ValueError('Job state for job %s is missing.' % job_id) return state
get_state_by_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_id):\n state = model.MapreduceState.get_by_job_id(job_id)\n if state is None:\n raise ValueError('Job state for job %s is missing.' % job...
def x(cls, job_config, mapreduce_spec): state = model.MapreduceState.create_new(job_config.job_id) state.mapreduce_spec = mapreduce_spec state.active = True state.active_shards = 0 state.app_id = job_config._app config = datastore_rpc.Configuration(force_writes=job_config._force_writes) state.put(config=config) return state
create_and_save_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config, mapreduce_spec):\n state = model.MapreduceState.create_new(job_config.job_id)\n state.mapreduce_spec = mapreduce_spec\n state.active ...
def x(cls, job_config, mapreduce_spec): params = {'mapreduce_id': job_config.job_id} kickoff_task = taskqueue.Task(url=job_config._base_path + '/kickoffjob_callback/' + job_config.job_id, headers=util. _get_task_headers(job_config.job_id), params=params) if job_config._hooks_cls: hooks = job_config._hooks_cls(mapreduce_spec) try: hooks.enqueue_kickoff_task(kickoff_task, job_config.queue_name) return except NotImplementedError: pass kickoff_task.add(job_config.queue_name, transactional=True)
add_kickoff_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config, mapreduce_spec):\n params = {'mapreduce_id': job_config.job_id}\n kickoff_task = taskqueue.Task(url=job_config._base_path +\n '/k...
def x(self): _json = self.to_json() try: return json.dumps(_json, sort_keys=True, cls=JsonEncoder) except: logging.exception('Could not serialize JSON: %r', _json) raise
to_json_str
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n _json = self.to_json()\n try:\n return json.dumps(_json, sort_keys=True, cls=JsonEncoder)\n except:\n logging.exception('Could ...
def x(cls, json_str): return cls.from_json(json.loads(json_str, cls=JsonDecoder))
from_json_str
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json_str):\n return cls.from_json(json.loads(json_str, cls=JsonDecoder))\n", "role": "user" }, { "content": "from_json_str", "role": "a...
def x(self, model_instance): value = super(JsonProperty, self).get_value_for_datastore(model_instance) if not value: return None json_value = value if not isinstance(value, dict): json_value = value.to_json() if not json_value: return None return datastore_types.Text(json.dumps(json_value, sort_keys=True, cls= JsonEncoder))
get_value_for_datastore
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_instance):\n value = super(JsonProperty, self).get_value_for_datastore(model_instance)\n if not value:\n return None\n json_value =...
def x(self, value): if value is None: return None _json = json.loads(value, cls=JsonDecoder) if self.data_type == dict: return _json return self.data_type.from_json(_json)
make_value_from_datastore
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n if value is None:\n return None\n _json = json.loads(value, cls=JsonDecoder)\n if self.data_type == dict:\n return _json...
def x(self, value): if value is not None and not isinstance(value, self.data_type): raise datastore_errors.BadValueError( 'Property %s must be convertible to a %s instance (%s)' % (self .name, self.data_type, value)) return super(JsonProperty, self).validate(value)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n if value is not None and not isinstance(value, self.data_type):\n raise datastore_errors.BadValueError(\n 'Property %s mus...
def x(self, queue_name, transactional=False): task = self.to_task() task.add(queue_name, transactional)
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, queue_name, transactional=False):\n task = self.to_task()\n task.add(queue_name, transactional)\n", "role": "user" }, { "content": "add...
def x(self): return taskqueue.Task(url=self.url, payload=self._payload, name=self. name, eta=self.eta, countdown=self.countdown, headers=self._headers)
to_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return taskqueue.Task(url=self.url, payload=self._payload, name=self.\n name, eta=self.eta, countdown=self.countdown, headers=self._headers)...
def x(cls, request): if request.headers.get(cls.PAYLOAD_VERSION_HEADER) != cls.PAYLOAD_VERSION: raise DeprecationWarning( 'Task is generated by an older incompatible version of mapreduce. Please kill this job manually' ) return cls._decode_payload(request.body)
decode_payload
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request):\n if request.headers.get(cls.PAYLOAD_VERSION_HEADER) != cls.PAYLOAD_VERSION:\n raise DeprecationWarning(\n 'Task is generat...
def x(self, counter_name, delta): current_value = self.counters.get(counter_name, 0) new_value = current_value + delta self.counters[counter_name] = new_value return new_value
increment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, counter_name, delta):\n current_value = self.counters.get(counter_name, 0)\n new_value = current_value + delta\n self.counters[counter_name] = n...
def x(self, counters_map): for counter_name in counters_map.counters: self.increment(counter_name, counters_map.counters[counter_name])
add_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, counters_map):\n for counter_name in counters_map.counters:\n self.increment(counter_name, counters_map.counters[counter_name])\n", "role":...
def x(self, counters_map): for counter_name in counters_map.counters: self.increment(counter_name, -counters_map.counters[counter_name])
sub_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, counters_map):\n for counter_name in counters_map.counters:\n self.increment(counter_name, -counters_map.counters[counter_name])\n", "role"...
def x(self): result = {'mapper_handler_spec': self.handler_spec, 'mapper_input_reader': self.input_reader_spec, 'mapper_params': self.params, 'mapper_shard_count': self.shard_count} if self.output_writer_spec: result['mapper_output_writer'] = self.output_writer_spec return result
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n result = {'mapper_handler_spec': self.handler_spec,\n 'mapper_input_reader': self.input_reader_spec, 'mapper_params':\n self.params, ...
def x(self): if self.__hooks is None and self.hooks_class_name is not None: hooks_class = util.for_name(self.hooks_class_name) if not isinstance(hooks_class, type): raise ValueError( 'hooks_class_name must refer to a class, got %s' % type( hooks_class).__name__) if not issubclass(hooks_class, hooks.Hooks): raise ValueError( 'hooks_class_name must refer to a hooks.Hooks subclass') self.__hooks = hooks_class(self) return self.__hooks
get_hooks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.__hooks is None and self.hooks_class_name is not None:\n hooks_class = util.for_name(self.hooks_class_name)\n if not isinstan...
def x(self): mapper_spec = self.mapper.to_json() return {'name': self.name, 'mapreduce_id': self.mapreduce_id, 'mapper_spec': mapper_spec, 'params': self.params, 'hooks_class_name': self.hooks_class_name}
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n mapper_spec = self.mapper.to_json()\n return {'name': self.name, 'mapreduce_id': self.mapreduce_id,\n 'mapper_spec': mapper_spec, 'params...
def x(cls, json): mapreduce_spec = cls(json['name'], json['mapreduce_id'], json[ 'mapper_spec'], json.get('params'), json.get('hooks_class_name')) return mapreduce_spec
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json):\n mapreduce_spec = cls(json['name'], json['mapreduce_id'], json[\n 'mapper_spec'], json.get('params'), json.get('hooks_class_name'))\n ...
def x(cls, mr_id): key = 'GAE-MR-spec: %s' % mr_id spec_json = memcache.get(key) if spec_json: return cls.from_json(spec_json) state = MapreduceState.get_by_job_id(mr_id) spec = state.mapreduce_spec spec_json = spec.to_json() memcache.set(key, spec_json) return spec
get_mapreduce_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mr_id):\n key = 'GAE-MR-spec: %s' % mr_id\n spec_json = memcache.get(key)\n if spec_json:\n return cls.from_json(spec_json)\n state = M...
def x(cls, mapreduce_id): return db.Key.from_path(cls.kind(), str(mapreduce_id))
get_key_by_job_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_id):\n return db.Key.from_path(cls.kind(), str(mapreduce_id))\n", "role": "user" }, { "content": "get_key_by_job_id", "role":...
def x(self, shards_processed, shards_status): chart = google_chart_api.BarChart() def x(status_to_filter): return [(count if status == status_to_filter else 0) for count, status in zip(shards_processed, shards_status)] if shards_status: chart.stacked = True chart.AddBars(filter_status('unknown'), color='404040') chart.AddBars(filter_status('success'), color='00ac42') chart.AddBars(filter_status('running'), color='3636a9') chart.AddBars(filter_status('aborted'), color='e29e24') chart.AddBars(filter_status('failed'), color='f6350f') else: chart.AddBars(shards_processed) shard_count = len(shards_processed) if shard_count > 95: pixels_per_shard = 700.0 / shard_count bar_thickness = int(pixels_per_shard * 0.9) chart.style = bar_chart.BarChartStyle(bar_thickness=bar_thickness, bar_gap=0.1, use_fractional_gap_spacing=True) if shards_processed and shard_count <= 95: stride_length = max(1, shard_count / 16) chart.bottom.labels = [] for x in xrange(shard_count): if x % stride_length == 0 or x == shard_count - 1: chart.bottom.labels.append(x) else: chart.bottom.labels.append('') chart.left.labels = ['0', str(max(shards_processed))] chart.left.min = 0 self.chart_width = min(700, max(300, shard_count * 20)) self.chart_url = chart.display.Url(self.chart_width, 200)
set_processed_counts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, shards_processed, shards_status):\n chart = google_chart_api.BarChart()\n\n def x(status_to_filter):\n return [(count if status == status_to...
def x(mapreduce_id=None, gettime=datetime.datetime.now): if not mapreduce_id: mapreduce_id = MapreduceState.new_mapreduce_id() state = MapreduceState(key_name=mapreduce_id, last_poll_time=gettime()) state.set_processed_counts([], []) return state
create_new
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mapreduce_id=None, gettime=datetime.datetime.now):\n if not mapreduce_id:\n mapreduce_id = MapreduceState.new_mapreduce_id()\n state = MapreduceSt...
def x(self, output_writer): self.input_reader = self.initial_input_reader self.slice_id = 0 self.retries += 1 self.output_writer = output_writer self.handler = self.mapreduce_spec.mapper.handler
reset_for_retry
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, output_writer):\n self.input_reader = self.initial_input_reader\n self.slice_id = 0\n self.retries += 1\n self.output_writer = output_writer\...
def x(self, recovery_slice=False): if recovery_slice: self.slice_id += 2 self.input_reader = self.input_reader.from_json(self._input_reader_json ) else: self.slice_id += 1
advance_for_next_slice
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, recovery_slice=False):\n if recovery_slice:\n self.slice_id += 2\n self.input_reader = self.input_reader.from_json(self._input_reader_js...
def x(self): result = {'mapreduce_spec': self.mapreduce_spec.to_json_str(), 'shard_id': self.shard_id, 'slice_id': str(self.slice_id), 'input_reader_state': self.input_reader.to_json_str(), 'initial_input_reader_state': self.initial_input_reader.to_json_str (), 'retries': str(self.retries)} if self.output_writer: result['output_writer_state'] = self.output_writer.to_json_str() serialized_handler = util.try_serialize_handler(self.handler) if serialized_handler: result['serialized_handler'] = serialized_handler return result
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n result = {'mapreduce_spec': self.mapreduce_spec.to_json_str(),\n 'shard_id': self.shard_id, 'slice_id': str(self.slice_id),\n 'input_...
def x(cls, request): mapreduce_spec = MapreduceSpec.from_json_str(request.get('mapreduce_spec')) mapper_spec = mapreduce_spec.mapper input_reader_spec_dict = json.loads(request.get('input_reader_state'), cls=json_util.JsonDecoder) input_reader = mapper_spec.input_reader_class().from_json( input_reader_spec_dict) initial_input_reader_spec_dict = json.loads(request.get( 'initial_input_reader_state'), cls=json_util.JsonDecoder) initial_input_reader = mapper_spec.input_reader_class().from_json( initial_input_reader_spec_dict) output_writer = None if mapper_spec.output_writer_class(): output_writer = mapper_spec.output_writer_class().from_json(json. loads(request.get('output_writer_state', '{}'), cls=json_util. JsonDecoder)) assert isinstance(output_writer, mapper_spec.output_writer_class() ), '%s.from_json returned an instance of wrong class: %s' % ( mapper_spec.output_writer_class(), output_writer.__class__) handler = util.try_deserialize_handler(request.get('serialized_handler')) if not handler: handler = mapreduce_spec.mapper.handler return cls(mapreduce_spec.params['base_path'], mapreduce_spec, str( request.get('shard_id')), int(request.get('slice_id')), input_reader, initial_input_reader, output_writer=output_writer, retries=int(request.get('retries')), handler=handler)
from_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request):\n mapreduce_spec = MapreduceSpec.from_json_str(request.get('mapreduce_spec'))\n mapper_spec = mapreduce_spec.mapper\n input_reader_spec...
def x(self): self.retries += 1 self.last_work_item = '' self.active = True self.result_status = None self.input_finished = False self.counters_map = CountersMap() self.slice_id = 0 self.slice_start_time = None self.slice_request_id = None self.slice_retries = 0 self.acquired_once = False
reset_for_retry
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.retries += 1\n self.last_work_item = ''\n self.active = True\n self.result_status = None\n self.input_finished = False\n self.c...
def x(self, recovery_slice=False): self.slice_start_time = None self.slice_request_id = None self.slice_retries = 0 self.acquired_once = False if recovery_slice: self.slice_id += 2 else: self.slice_id += 1
advance_for_next_slice
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, recovery_slice=False):\n self.slice_start_time = None\n self.slice_request_id = None\n self.slice_retries = 0\n self.acquired_once = False\n ...
def x(self, other_state): for prop in self.properties().values(): setattr(self, prop.name, getattr(other_state, prop.name))
copy_from
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, other_state):\n for prop in self.properties().values():\n setattr(self, prop.name, getattr(other_state, prop.name))\n", "role": "user" },...
def x(cls, mapreduce_state): keys = cls.calculate_keys_by_mapreduce_state(mapreduce_state) i = 0 while i < len(keys): @db.non_transactional def x(i): return db.get(keys[i:i + cls._MAX_STATES_IN_MEMORY]) states = no_tx_get(i) for s in states: i += 1 if s is not None: yield s
find_all_by_mapreduce_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_state):\n keys = cls.calculate_keys_by_mapreduce_state(mapreduce_state)\n i = 0\n while i < len(keys):\n\n @db.non_transactional...
def x(cls, mapreduce_state): if mapreduce_state is None: return [] keys = [] for i in range(mapreduce_state.mapreduce_spec.mapper.shard_count): shard_id = cls.shard_id_from_number(mapreduce_state.key().name(), i) keys.append(cls.get_key_by_shard_id(shard_id)) return keys
calculate_keys_by_mapreduce_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_state):\n if mapreduce_state is None:\n return []\n keys = []\n for i in range(mapreduce_state.mapreduce_spec.mapper.shard_count...
def x(cls, mapreduce_id, shard_number): shard_id = cls.shard_id_from_number(mapreduce_id, shard_number) state = cls(key_name=shard_id, mapreduce_id=mapreduce_id) return state
create_new
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_id, shard_number):\n shard_id = cls.shard_id_from_number(mapreduce_id, shard_number)\n state = cls(key_name=shard_id, mapreduce_id=mapredu...
def x(cls, mapreduce_id): return db.Key.from_path(cls.kind(), '%s:%s' % (mapreduce_id, cls._KEY_NAME) )
get_key_by_job_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_id):\n return db.Key.from_path(cls.kind(), '%s:%s' % (mapreduce_id, cls._KEY_NAME)\n )\n", "role": "user" }, { "content": ...
def x(cls, mapreduce_id, **kwargs): cls(key_name='%s:%s' % (mapreduce_id, cls._KEY_NAME), command=cls.ABORT ).put(**kwargs)
abort
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapreduce_id, **kwargs):\n cls(key_name='%s:%s' % (mapreduce_id, cls._KEY_NAME), command=cls.ABORT\n ).put(**kwargs)\n", "role": "user" },...
def x(cls, job_config): super(DatastoreInputReader, cls).validate(job_config) params = job_config.input_reader_params entity_kind = params[cls.ENTITY_KIND_PARAM] if '.' in entity_kind: logging.warning( '. detected in entity kind %s specified for reader %s.Assuming entity kind contains the dot.' , entity_kind, cls.__name__) if cls.FILTERS_PARAM in params: filters = params[cls.FILTERS_PARAM] for f in filters: if f[1] != '=': raise errors.BadReaderParamsError( 'Only equality filters are supported: %s', f)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config):\n super(DatastoreInputReader, cls).validate(job_config)\n params = job_config.input_reader_params\n entity_kind = params[cls.ENTITY_...