source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(cls, app, namespaces, shard_count, query_spec): key_ranges_by_ns = [] for namespace in namespaces: ranges = cls._split_ns_by_scatter(shard_count, namespace, query_spec.entity_kind, app) random.shuffle(ranges) key_ranges_by_ns.append(ranges) ranges_by_shard = [[] for _ in range(shard_count)] for ranges in key_ranges_by_ns: for i, k_range in enumerate(ranges): if k_range: ranges_by_shard[i].append(k_range) key_ranges_by_shard = [] for ranges in ranges_by_shard: if ranges: key_ranges_by_shard.append(key_ranges.KeyRangesFactory. create_from_list(ranges)) return key_ranges_by_shard
to_key_ranges_by_shard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, app, namespaces, shard_count, query_spec):\n key_ranges_by_ns = []\n for namespace in namespaces:\n ranges = cls._split_ns_by_scatter(shard_c...
def x(cls, shard_count, namespace, raw_entity_kind, app): if shard_count == 1: return [key_range.KeyRange(namespace=namespace, _app=app)] ds_query = datastore.Query(kind=raw_entity_kind, namespace=namespace, _app=app, keys_only=True) ds_query.Order('__scatter__') oversampling_factor = 32 random_keys = ds_query.Get(shard_count * oversampling_factor) if not random_keys: return [key_range.KeyRange(namespace=namespace, _app=app)] + [None] * ( shard_count - 1) random_keys.sort() if len(random_keys) >= shard_count: random_keys = cls._choose_split_points(random_keys, shard_count) k_ranges = [] k_ranges.append(key_range.KeyRange(key_start=None, key_end=random_keys[ 0], direction=key_range.KeyRange.ASC, include_start=False, include_end=False, namespace=namespace, _app=app)) for i in range(0, len(random_keys) - 1): k_ranges.append(key_range.KeyRange(key_start=random_keys[i], key_end=random_keys[i + 1], direction=key_range.KeyRange.ASC, include_start=True, include_end=False, namespace=namespace, _app=app)) k_ranges.append(key_range.KeyRange(key_start=random_keys[-1], key_end= None, direction=key_range.KeyRange.ASC, include_start=True, include_end=False, namespace=namespace, _app=app)) if len(k_ranges) < shard_count: k_ranges += [None] * (shard_count - len(k_ranges)) return k_ranges
split_ns_by_scatter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, shard_count, namespace, raw_entity_kind, app):\n if shard_count == 1:\n return [key_range.KeyRange(namespace=namespace, _app=app)]\n ds_query...
def x(alphabet=NAMESPACE_CHARACTERS, max_length=MAX_NAMESPACE_LENGTH, batch_size=NAMESPACE_BATCH_SIZE): global NAMESPACE_CHARACTERS global MAX_NAMESPACE_LENGTH global MAX_NAMESPACE global _LEX_DISTANCE global NAMESPACE_BATCH_SIZE NAMESPACE_CHARACTERS = alphabet MAX_NAMESPACE_LENGTH = max_length MAX_NAMESPACE = NAMESPACE_CHARACTERS[-1] * MAX_NAMESPACE_LENGTH NAMESPACE_BATCH_SIZE = batch_size _LEX_DISTANCE = [1] for i in range(1, MAX_NAMESPACE_LENGTH): _LEX_DISTANCE.append(_LEX_DISTANCE[i - 1] * len( NAMESPACE_CHARACTERS) + 1) del i
setup_constants
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(alphabet=NAMESPACE_CHARACTERS, max_length=MAX_NAMESPACE_LENGTH,\n batch_size=NAMESPACE_BATCH_SIZE):\n global NAMESPACE_CHARACTERS\n global MAX_NAMESPA...
def x(n, _max_length=None): if _max_length is None: _max_length = MAX_NAMESPACE_LENGTH length = _LEX_DISTANCE[_max_length - 1] if n == 0: return '' n -= 1 return NAMESPACE_CHARACTERS[n / length] + _ord_to_namespace(n % length, _max_length - 1)
ord_to_namespace
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(n, _max_length=None):\n if _max_length is None:\n _max_length = MAX_NAMESPACE_LENGTH\n length = _LEX_DISTANCE[_max_length - 1]\n if n == 0:\n ...
def x(namespace): n = 0 for i, c in enumerate(namespace): n += _LEX_DISTANCE[MAX_NAMESPACE_LENGTH - i - 1 ] * NAMESPACE_CHARACTERS.index(c) + 1 return n
namespace_to_ord
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(namespace):\n n = 0\n for i, c in enumerate(namespace):\n n += _LEX_DISTANCE[MAX_NAMESPACE_LENGTH - i - 1\n ] * NAMESPACE_CHARACTERS.in...
def x(namespace, app): if namespace: return db.Key.from_path(metadata.Namespace.KIND_NAME, namespace, _app=app) else: return db.Key.from_path(metadata.Namespace.KIND_NAME, metadata. Namespace.EMPTY_NAMESPACE_ID, _app=app)
key_for_namespace
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(namespace, app):\n if namespace:\n return db.Key.from_path(metadata.Namespace.KIND_NAME, namespace,\n _app=app)\n else:\n return...
def x(app, limit): ns_query = datastore.Query('__namespace__', keys_only=True, _app=app) return list(ns_query.Run(limit=limit, batch_size=limit))
get_namespace_keys
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(app, limit):\n ns_query = datastore.Query('__namespace__', keys_only=True, _app=app)\n return list(ns_query.Run(limit=limit, batch_size=limit))\n", "...
def x(self): if self.is_single_namespace: return [self] mid_point = (_namespace_to_ord(self.namespace_start) + _namespace_to_ord(self.namespace_end)) // 2 return [NamespaceRange(self.namespace_start, _ord_to_namespace( mid_point), _app=self.app), NamespaceRange(_ord_to_namespace( mid_point + 1), self.namespace_end, _app=self.app)]
split_range
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.is_single_namespace:\n return [self]\n mid_point = (_namespace_to_ord(self.namespace_start) +\n _namespace_to_ord(self.nam...
def x(self, after_namespace): namespace_start = _ord_to_namespace(_namespace_to_ord(after_namespace) + 1) return NamespaceRange(namespace_start, self.namespace_end, _app=self.app)
with_start_after
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, after_namespace):\n namespace_start = _ord_to_namespace(_namespace_to_ord(after_namespace) + 1)\n return NamespaceRange(namespace_start, self.names...
def x(self, cursor=None): filters = {} filters['__key__ >= '] = _key_for_namespace(self.namespace_start, self.app) filters['__key__ <= '] = _key_for_namespace(self.namespace_end, self.app) return datastore.Query('__namespace__', filters=filters, keys_only=True, cursor=cursor, _app=self.app)
make_datastore_query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, cursor=None):\n filters = {}\n filters['__key__ >= '] = _key_for_namespace(self.namespace_start, self.app)\n filters['__key__ <= '] = _key_for_n...
def x(self): namespaces_after_key = list(self.make_datastore_query().Run(limit=1)) if not namespaces_after_key: return None namespace_after_key = namespaces_after_key[0].name() or '' return NamespaceRange(namespace_after_key, self.namespace_end, _app= self.app)
normalized_start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n namespaces_after_key = list(self.make_datastore_query().Run(limit=1))\n if not namespaces_after_key:\n return None\n namespace_after_k...
def x(self): obj_dict = dict(namespace_start=self.namespace_start, namespace_end= self.namespace_end) if self.app is not None: obj_dict['app'] = self.app return obj_dict
to_json_object
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n obj_dict = dict(namespace_start=self.namespace_start, namespace_end=\n self.namespace_end)\n if self.app is not None:\n obj_dict['...
def x(cls, n, contiguous, can_query=itertools.chain(itertools.repeat(True, 50), itertools.repeat(False)).next, _app=None): if n < 1: raise ValueError('n must be >= 1') ranges = None if can_query(): if not contiguous: ns_keys = get_namespace_keys(_app, n + 1) if not ns_keys: return [] else: if len(ns_keys) <= n: ns_range = [] for ns_key in ns_keys: ns_range.append(NamespaceRange(ns_key.name() or '', ns_key.name() or '', _app=_app)) return sorted(ns_range, key=lambda ns_range: ns_range. namespace_start) ranges = [NamespaceRange(ns_keys[0].name() or '', _app=_app)] else: ns_range = NamespaceRange(_app=_app).normalized_start() if ns_range is None: return [NamespaceRange(_app=_app)] ranges = [ns_range] else: ranges = [NamespaceRange(_app=_app)] singles = [] while ranges and len(ranges) + len(singles) < n: namespace_range = ranges.pop(0) if namespace_range.is_single_namespace: singles.append(namespace_range) else: left, right = namespace_range.split_range() if can_query(): right = right.normalized_start() if right is not None: ranges.append(right) ranges.append(left) ns_ranges = sorted(singles + ranges, key=lambda ns_range: ns_range. namespace_start) if contiguous: if not ns_ranges: return [NamespaceRange(_app=_app)] continuous_ns_ranges = [] for i in range(len(ns_ranges)): if i == 0: namespace_start = MIN_NAMESPACE else: namespace_start = ns_ranges[i].namespace_start if i == len(ns_ranges) - 1: namespace_end = MAX_NAMESPACE else: namespace_end = _ord_to_namespace(_namespace_to_ord( ns_ranges[i + 1].namespace_start) - 1) continuous_ns_ranges.append(NamespaceRange(namespace_start, namespace_end, _app=_app)) return continuous_ns_ranges else: return ns_ranges
split
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, n, contiguous, can_query=itertools.chain(itertools.repeat(True, \n 50), itertools.repeat(False)).next, _app=None):\n if n < 1:\n raise ValueE...
def x(self, data): data_length = len(data) if self._size + data_length > self._flush_size: self.flush() if not self._exclusive and data_length > _FILE_POOL_MAX_SIZE: raise errors.Error('Too big input %s (%s).' % (data_length, _FILE_POOL_MAX_SIZE)) else: self._buffer.append(data) self._size += data_length if self._size > self._flush_size: self.flush()
append
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n data_length = len(data)\n if self._size + data_length > self._flush_size:\n self.flush()\n if not self._exclusive and data_lengt...
def x(self): buf = cStringIO.StringIO() with records.RecordsWriter(buf) as w: for record in self._buffer: w.write(record) w._pad_block() str_buf = buf.getvalue() buf.close() if not self._exclusive and len(str_buf) > _FILE_POOL_MAX_SIZE: raise errors.Error( "Buffer too big. Can't write more than %s bytes in one request: risk of writes interleaving. Got: %s" % (_FILE_POOL_MAX_SIZE, len(str_buf))) start_time = time.time() self._write(str_buf) if self._ctx: operation.counters.Increment(COUNTER_IO_WRITE_BYTES, len(str_buf))(self ._ctx) operation.counters.Increment(COUNTER_IO_WRITE_MSEC, int((time.time( ) - start_time) * 1000))(self._ctx) self._buffer = [] self._size = 0 gc.collect()
flush
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n buf = cStringIO.StringIO()\n with records.RecordsWriter(buf) as w:\n for record in self._buffer:\n w.write(record)\n w....
def x(self, str_buf): self._filehandle.write(str_buf) self._buf_size += len(str_buf)
write
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, str_buf):\n self._filehandle.write(str_buf)\n self._buf_size += len(str_buf)\n", "role": "user" }, { "content": "write", "role": "a...
def x(self, force=False): super(GCSRecordsPool, self).flush() if force: extra_padding = self._buf_size % self._GCS_BLOCK_SIZE if extra_padding > 0: self._write('\x00' * (self._GCS_BLOCK_SIZE - extra_padding)) self._filehandle.flush()
flush
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n super(GCSRecordsPool, self).flush()\n if force:\n extra_padding = self._buf_size % self._GCS_BLOCK_SIZE\n if extra_pa...
def x(cls, writer_spec): if cls.TMP_BUCKET_NAME_PARAM in writer_spec: return writer_spec[cls.TMP_BUCKET_NAME_PARAM] return cls._get_gcs_bucket(writer_spec)
get_tmp_gcs_bucket
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, writer_spec):\n if cls.TMP_BUCKET_NAME_PARAM in writer_spec:\n return writer_spec[cls.TMP_BUCKET_NAME_PARAM]\n return cls._get_gcs_bucket(wri...
def x(cls, writer_spec): if cls.TMP_BUCKET_NAME_PARAM in writer_spec: return writer_spec.get(cls._TMP_ACCOUNT_ID_PARAM, None) return cls._get_account_id(writer_spec)
get_tmp_account_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, writer_spec):\n if cls.TMP_BUCKET_NAME_PARAM in writer_spec:\n return writer_spec.get(cls._TMP_ACCOUNT_ID_PARAM, None)\n return cls._get_acco...
def x(cls, writer_spec, filename_suffix, use_tmp_bucket=False): if use_tmp_bucket: bucket = cls._get_tmp_gcs_bucket(writer_spec) account_id = cls._get_tmp_account_id(writer_spec) else: bucket = cls._get_gcs_bucket(writer_spec) account_id = cls._get_account_id(writer_spec) filename = '/%s/%s' % (bucket, filename_suffix) content_type = writer_spec.get(cls.CONTENT_TYPE_PARAM, None) options = {} if cls.ACL_PARAM in writer_spec: options['x-goog-acl'] = writer_spec.get(cls.ACL_PARAM) return cloudstorage.open(filename, mode='w', content_type=content_type, options=options, _account_id=account_id)
open_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, writer_spec, filename_suffix, use_tmp_bucket=False):\n if use_tmp_bucket:\n bucket = cls._get_tmp_gcs_bucket(writer_spec)\n account_id = ...
def x(self, data): start_time = time.time() self._get_write_buffer().write(data) ctx = context.get() operation.counters.Increment(COUNTER_IO_WRITE_BYTES, len(data))(ctx) operation.counters.Increment(COUNTER_IO_WRITE_MSEC, int((time.time() - start_time) * 1000))(ctx)
write
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n start_time = time.time()\n self._get_write_buffer().write(data)\n ctx = context.get()\n operation.counters.Increment(COUNTER_IO_WRIT...
def x(cls, mapper_spec): writer_spec = cls.get_params(mapper_spec, allow_old=False) if writer_spec.get(cls._NO_DUPLICATE, False) not in (True, False): raise errors.BadWriterParamsError('No duplicate must a boolean.') super(_GoogleCloudStorageOutputWriter, cls).validate(mapper_spec)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mapper_spec):\n writer_spec = cls.get_params(mapper_spec, allow_old=False)\n if writer_spec.get(cls._NO_DUPLICATE, False) not in (True, False):\n ...
def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None): writer_spec = cls.get_params(mr_spec.mapper, allow_old=False) seg_index = None if writer_spec.get(cls._NO_DUPLICATE, False): seg_index = 0 key = cls._generate_filename(writer_spec, mr_spec.name, mr_spec. mapreduce_id, shard_number, shard_attempt, seg_index) return cls._create(writer_spec, key)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None):\n writer_spec = cls.get_params(mr_spec.mapper, allow_old=False)\n seg_index = None\n i...
def x(cls, writer_spec, filename_suffix): writer = cls._open_file(writer_spec, filename_suffix) return cls(writer, writer_spec=writer_spec)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, writer_spec, filename_suffix):\n writer = cls._open_file(writer_spec, filename_suffix)\n return cls(writer, writer_spec=writer_spec)\n", "role":...
def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None): writer_spec = cls.get_params(mr_spec.mapper, allow_old=False) key = cls._generate_filename(writer_spec, mr_spec.name, mr_spec. mapreduce_id, shard_number, shard_attempt) status = _ConsistentStatus() status.writer_spec = writer_spec status.mainfile = cls._open_file(writer_spec, key) status.mapreduce_id = mr_spec.mapreduce_id status.shard = shard_number return cls(status)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, mr_spec, shard_number, shard_attempt, _writer_state=None):\n writer_spec = cls.get_params(mr_spec.mapper, allow_old=False)\n key = cls._generate_fil...
def x(self, mainfile, tmpfile, writer_spec): if mainfile.closed: return account_id = self._get_tmp_account_id(writer_spec) f = cloudstorage_api.open(tmpfile, _account_id=account_id) data = f.read(self._REWRITE_BLOCK_SIZE) while data: mainfile.write(data) data = f.read(self._REWRITE_BLOCK_SIZE) f.close() mainfile.flush()
rewrite_tmpfile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mainfile, tmpfile, writer_spec):\n if mainfile.closed:\n return\n account_id = self._get_tmp_account_id(writer_spec)\n f = cloudstorage_a...
def x(cls, status): tmpl = string.Template(cls._TMPFILE_PATTERN) filename = tmpl.substitute(id=status.mapreduce_id, shard=status.shard, random=random.getrandbits(cls._RAND_BITS)) return cls._open_file(status.writer_spec, filename, use_tmp_bucket=True)
create_tmpfile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, status):\n tmpl = string.Template(cls._TMPFILE_PATTERN)\n filename = tmpl.substitute(id=status.mapreduce_id, shard=status.shard,\n random=ran...
def x(self, writer_spec, exclude_list=()): tmpl = string.Template(self._TMPFILE_PREFIX) prefix = tmpl.substitute(id=self.status.mapreduce_id, shard=self.status .shard) bucket = self._get_tmp_gcs_bucket(writer_spec) account_id = self._get_tmp_account_id(writer_spec) for f in cloudstorage.listbucket('/%s/%s' % (bucket, prefix), _account_id=account_id): if f.filename not in exclude_list: self._remove_tmpfile(f.filename, self.status.writer_spec)
try_to_clean_garbage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, writer_spec, exclude_list=()):\n tmpl = string.Template(self._TMPFILE_PREFIX)\n prefix = tmpl.substitute(id=self.status.mapreduce_id, shard=self.st...
def x(cls, state): return cls(state[cls.COUNT], state[cls.STRING_LENGTH])
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, state):\n return cls(state[cls.COUNT], state[cls.STRING_LENGTH])\n", "role": "user" }, { "content": "from_json", "role": "assistant" ...
def x(cls, job_config): params = job_config.input_reader_params count = params[cls.COUNT] string_length = params.get(cls.STRING_LENGTH, cls._DEFAULT_STRING_LENGTH) shard_count = job_config.shard_count count_per_shard = count // shard_count mr_input_readers = [cls(count_per_shard, string_length) for _ in range( shard_count)] left = count - count_per_shard * shard_count if left > 0: mr_input_readers.append(cls(left, string_length)) return mr_input_readers
split_input
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config):\n params = job_config.input_reader_params\n count = params[cls.COUNT]\n string_length = params.get(cls.STRING_LENGTH, cls._DEFAULT_S...
def x(cls, job_config): super(SampleInputReader, cls).validate(job_config) params = job_config.input_reader_params if cls.COUNT not in params: raise errors.BadReaderParamsError('Must specify %s' % cls.COUNT) if not isinstance(params[cls.COUNT], int): raise errors.BadReaderParamsError('%s should be an int but is %s' % (cls.COUNT, type(params[cls.COUNT]))) if params[cls.COUNT] <= 0: raise errors.BadReaderParamsError('%s should be a positive int') if cls.STRING_LENGTH in params and not (isinstance(params[cls. STRING_LENGTH], int) and params[cls.STRING_LENGTH] > 0): raise errors.BadReaderParamsError( '%s should be a positive int but is %s' % (cls.STRING_LENGTH, params[cls.STRING_LENGTH]))
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config):\n super(SampleInputReader, cls).validate(job_config)\n params = job_config.input_reader_params\n if cls.COUNT not in params:\n ...
def x(max_length): weights = [1] for i in range(1, max_length): weights.append(weights[i - 1] * len(_ALPHABET) + 1) weights.reverse() return weights
get_weights
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(max_length):\n weights = [1]\n for i in range(1, max_length):\n weights.append(weights[i - 1] * len(_ALPHABET) + 1)\n weights.reverse()\n re...
def x(content, weights): ordinal = 0 for i, c in enumerate(content): ordinal += weights[i] * _ALPHABET.index(c) + 1 return ordinal
str_to_ord
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(content, weights):\n ordinal = 0\n for i, c in enumerate(content):\n ordinal += weights[i] * _ALPHABET.index(c) + 1\n return ordinal\n", "r...
def x(ordinal, weights): chars = [] for weight in weights: if ordinal == 0: return ''.join(chars) ordinal -= 1 index, ordinal = divmod(ordinal, weight) chars.append(_ALPHABET[index]) return ''.join(chars)
ord_to_str
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ordinal, weights):\n chars = []\n for weight in weights:\n if ordinal == 0:\n return ''.join(chars)\n ordinal -= 1\n inde...
def x(cls, filters, model_class): if not filters: return None, None, None range_property = None start_val = None end_val = None start_filter = None end_filter = None for f in filters: prop, op, val = f if op in ['>', '>=', '<', '<=']: if range_property and range_property != prop: raise errors.BadReaderParamsError( 'Range on only one property is supported.') range_property = prop if val is None: raise errors.BadReaderParamsError( "Range can't be None in filter %s", f) if op in ['>', '>=']: if start_val is not None: raise errors.BadReaderParamsError( 'Operation %s is specified more than once.', op) start_val = val start_filter = f else: if end_val is not None: raise errors.BadReaderParamsError( 'Operation %s is specified more than once.', op) end_val = val end_filter = f elif op != '=': raise errors.BadReaderParamsError( 'Only < <= > >= = are supported as operation. Got %s', op) if not range_property: return None, None, None if start_val is None or end_val is None: raise errors.BadReaderParamsError( 'Filter should contains a complete range on property %s', range_property) if issubclass(model_class, db.Model): property_obj = model_class.properties()[range_property] else: property_obj = model_class._properties[range_property] supported_properties = _DISCRETE_PROPERTY_SPLIT_FUNCTIONS.keys( ) + _CONTINUOUS_PROPERTY_SPLIT_FUNCTIONS.keys() if not isinstance(property_obj, tuple(supported_properties)): raise errors.BadReaderParamsError( 'Filtered property %s is not supported by sharding.', range_property) if not start_val < end_val: raise errors.BadReaderParamsError( 'Start value %s should be smaller than end value %s', start_val, end_val) return property_obj, start_filter, end_filter
get_range_from_filters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, filters, model_class):\n if not filters:\n return None, None, None\n range_property = None\n start_val = None\n end_val = None\n sta...
def x(self, n): new_range_filters = [] name = self.start[0] prop_cls = self.prop.__class__ if prop_cls in _DISCRETE_PROPERTY_SPLIT_FUNCTIONS: splitpoints = _DISCRETE_PROPERTY_SPLIT_FUNCTIONS[prop_cls](self. start[2], self.end[2], n, self.start[1] == '>=', self.end[1] == '<=') start_filter = name, '>=', splitpoints[0] for p in splitpoints[1:]: end_filter = name, '<', p new_range_filters.append([start_filter, end_filter]) start_filter = name, '>=', p else: splitpoints = _CONTINUOUS_PROPERTY_SPLIT_FUNCTIONS[prop_cls](self. start[2], self.end[2], n) start_filter = self.start for p in splitpoints: end_filter = name, '<', p new_range_filters.append([start_filter, end_filter]) start_filter = name, '>=', p new_range_filters.append([start_filter, self.end]) for f in new_range_filters: f.extend(self._equality_filters) return [self.__class__(f, self.model_class_path) for f in new_range_filters ]
split
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, n):\n new_range_filters = []\n name = self.start[0]\n prop_cls = self.prop.__class__\n if prop_cls in _DISCRETE_PROPERTY_SPLIT_FUNCTIONS:\n ...
def x(self, ns): if issubclass(self.model_class, db.Model): query = db.Query(self.model_class, namespace=ns) for f in self.filters: query.filter('%s %s' % (f[0], f[1]), f[2]) else: query = self.model_class.query(namespace=ns) for f in self.filters: query = query.filter(ndb.FilterNode(*f)) return query
make_query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ns):\n if issubclass(self.model_class, db.Model):\n query = db.Query(self.model_class, namespace=ns)\n for f in self.filters:\n ...
def x(cls, job_config): if job_config.output_writer_cls != cls: raise errors.BadWriterParamsError( 'Expect output writer class %r, got %r.' % (cls, job_config. output_writer_cls))
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, job_config):\n if job_config.output_writer_cls != cls:\n raise errors.BadWriterParamsError(\n 'Expect output writer class %r, got %r....
def x(cls, shard_ctx, iterator): outs = tuple(iterator) shard_ctx._state.writer_state['outs'] = outs
commit_output
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, shard_ctx, iterator):\n outs = tuple(iterator)\n shard_ctx._state.writer_state['outs'] = outs\n", "role": "user" }, { "content": "commit...
def x(): pipeline_handlers_map = [] if pipeline: pipeline_handlers_map = pipeline.create_handlers_map(prefix= '.*/pipeline') return pipeline_handlers_map + [('.*/worker_callback.*', handlers. MapperWorkerCallbackHandler), ('.*/controller_callback.*', handlers .ControllerCallbackHandler), ('.*/kickoffjob_callback.*', handlers. KickOffJobHandler), ('.*/finalizejob_callback.*', handlers. FinalizeJobHandler), ('.*/command/start_job', handlers. StartJobHandler), ('.*/command/cleanup_job', handlers. CleanUpJobHandler), ('.*/command/abort_job', handlers. AbortJobHandler), ('.*/command/list_configs', status. ListConfigsHandler), ('.*/command/list_jobs', status. ListJobsHandler), ('.*/command/get_job_detail', status. GetJobDetailHandler), (STATIC_RE, status.ResourceHandler), ('.*', RedirectHandler)]
create_handlers_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n pipeline_handlers_map = []\n if pipeline:\n pipeline_handlers_map = pipeline.create_handlers_map(prefix=\n '.*/pipeline')\n ret...
def x(cls, json): if json['name'] in _KEYRANGES_CLASSES: return _KEYRANGES_CLASSES[json['name']].from_json(json) raise ValueError('Invalid json %s', json)
from_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, json):\n if json['name'] in _KEYRANGES_CLASSES:\n return _KEYRANGES_CLASSES[json['name']].from_json(json)\n raise ValueError('Invalid json %s...
def x(s): s = re.sub('\\s+', ' ', s) s = re.sub('[\\\\.\\\\?\\\\!]', '\n', s) return s.split('\n')
split_into_sentences
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n s = re.sub('\\\\s+', ' ', s)\n s = re.sub('[\\\\\\\\.\\\\\\\\?\\\\\\\\!]', '\\n', s)\n return s.split('\\n')\n", "role": "user" }, { ...
def x(s): s = re.sub('\\W+', ' ', s) s = re.sub('[_0-9]+', ' ', s) return s.split()
split_into_words
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n s = re.sub('\\\\W+', ' ', s)\n s = re.sub('[_0-9]+', ' ', s)\n return s.split()\n", "role": "user" }, { "content": "split_into_words...
def x(data): entry, text_fn = data text = text_fn() logging.debug('Got %s', entry.filename) for s in split_into_sentences(text): for w in split_into_words(s.lower()): yield w, entry.filename
index_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data):\n entry, text_fn = data\n text = text_fn()\n logging.debug('Got %s', entry.filename)\n for s in split_into_sentences(text):\n for w i...
def x(data): entry, text_fn = data text = text_fn() filename = entry.filename logging.debug('Got %s', filename) for s in split_into_sentences(text): words = split_into_words(s.lower()) if len(words) < PHRASE_LENGTH: yield ':'.join(words), filename continue for i in range(0, len(words) - PHRASE_LENGTH): yield ':'.join(words[i:i + PHRASE_LENGTH]), filename
phrases_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data):\n entry, text_fn = data\n text = text_fn()\n filename = entry.filename\n logging.debug('Got %s', filename)\n for s in split_into_sentence...
def x(key, values): if len(values) < 10: return counts = {} for filename in values: counts[filename] = counts.get(filename, 0) + 1 words = re.sub(':', ' ', key) threshold = len(values) / 2 for filename, count in counts.items(): if count > threshold: yield '%s:%s\n' % (words, filename)
phrases_reduce
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key, values):\n if len(values) < 10:\n return\n counts = {}\n for filename in values:\n counts[filename] = counts.get(filename, 0) + 1\n...
def x(username, date, blob_key): sep = FileMetadata.__SEP return str(username + sep + str(date) + sep + blob_key)
getKeyName
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(username, date, blob_key):\n sep = FileMetadata.__SEP\n return str(username + sep + str(date) + sep + blob_key)\n", "role": "user" }, { "cont...
def x(cls, spec): dec = MontyDecoder() def x(dotpath): modname, classname = dotpath.rsplit('.', 1) mod = __import__(modname, globals(), locals(), [classname], 0) return getattr(mod, classname) def x(d): decoded = {} for k, v in d.items(): if k.startswith('$'): if isinstance(v, list): v = [os.path.expandvars(i) for i in v] elif isinstance(v, dict): v = {k2: os.path.expandvars(v2) for k2, v2 in v.items()} else: v = os.path.expandvars(v) decoded[k.strip('$')] = dec.process_decoded(v) return decoded jobs = [] common_params = process_params(spec.get('jobs_common_params', {})) for d in spec['jobs']: cls_ = load_class(d['jb']) params = process_params(d.get('params', {})) params.update(common_params) jobs.append(cls_(**params)) handlers = [] for d in spec.get('handlers', []): cls_ = load_class(d['hdlr']) params = process_params(d.get('params', {})) handlers.append(cls_(**params)) validators = [] for d in spec.get('validators', []): cls_ = load_class(d['vldr']) params = process_params(d.get('params', {})) validators.append(cls_(**params)) custodian_params = process_params(spec.get('custodian_params', {})) return cls(jobs=jobs, handlers=handlers, validators=validators, ** custodian_params)
from_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, spec):\n dec = MontyDecoder()\n\n def x(dotpath):\n modname, classname = dotpath.rsplit('.', 1)\n mod = __import__(modname, globals(),...
def x(self): cwd = os.getcwd() with ScratchDir(self.scratch_dir, create_symbolic_link=True, copy_to_current_on_exit=True, copy_from_current_on_enter=True ) as temp_dir: self.total_errors = 0 start = datetime.datetime.now() logger.info('Run started at {} in {}.'.format(start, temp_dir)) v = sys.version.replace('\n', ' ') logger.info('Custodian running on Python version {}'.format(v)) logger.info('Hostname: {}, Cluster: {}'.format(* get_execution_host_info())) try: for job_n, job in islice(enumerate(self.jobs, 1), self.restart, None): self._run_job(job_n, job) dumpfn(self.run_log, Custodian.LOG_FILE, cls=MontyEncoder, indent=4) if self.checkpoint: self.restart = job_n Custodian._save_checkpoint(cwd, job_n) except CustodianError as ex: logger.error(ex.message) if ex.raises: raise finally: logger.info('Logging to {}...'.format(Custodian.LOG_FILE)) dumpfn(self.run_log, Custodian.LOG_FILE, cls=MontyEncoder, indent=4 ) end = datetime.datetime.now() logger.info('Run ended at {}.'.format(end)) run_time = end - start logger.info('Run completed. Total time taken = {}.'.format( run_time)) if self.gzipped_output: gzip_dir('.') Custodian._delete_checkpoints(cwd) return self.run_log
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cwd = os.getcwd()\n with ScratchDir(self.scratch_dir, create_symbolic_link=True,\n copy_to_current_on_exit=True, copy_from_current_on_ent...
def x(self, job_n, job): self.run_log.append({'job': job.as_dict(), 'corrections': [], 'handler': None, 'validator': None, 'max_errors': False, 'max_errors_per_job': False, 'max_errors_per_handler': False, 'nonzero_return_code': False}) self.errors_current_job = 0 for h in self.handlers: h.n_applied_corrections = 0 job.setup() attempt = 0 while (self.total_errors < self.max_errors and self.errors_current_job < self.max_errors_per_job): attempt += 1 logger.info( 'Starting job no. {} ({}) attempt no. {}. Total errors and errors in job thus far = {}, {}.' .format(job_n, job.name, attempt, self.total_errors, self. errors_current_job)) p = job.run() has_error = False zero_return_code = True if isinstance(p, subprocess.Popen): if self.monitors: n = 0 while True: n += 1 time.sleep(self.polling_time_step) if p.poll() is not None: break terminate = self.terminate_func or p.terminate if n % self.monitor_freq == 0: has_error = self._do_check(self.monitors, terminate) if terminate is not None and terminate != p.terminate: time.sleep(self.polling_time_step) else: p.wait() if (self.terminate_func is not None and self.terminate_func != p.terminate): self.terminate_func() time.sleep(self.polling_time_step) zero_return_code = p.returncode == 0 logger.info('{}.run has completed. Checking remaining handlers'. format(job.name)) if has_error: self._do_check([h for h in self.handlers if not h.is_monitor]) else: has_error = self._do_check(self.handlers) if has_error: job.terminate() if not has_error: for v in self.validators: if v.check(): self.run_log[-1]['validator'] = v s = 'Validation failed: {}'.format(v) raise ValidationError(s, True, v) if not zero_return_code: if self.terminate_on_nonzero_returncode: self.run_log[-1]['nonzero_return_code'] = True s = 'Job return code is %d. Terminating...' % p.returncode logger.info(s) raise ReturnCodeError(s, True) else: warnings.warn( 'subprocess returned a non-zero return code. Check outputs carefully...' ) job.postprocess() return for x in self.run_log[-1]['corrections']: if not x['actions'] and x['handler'].raises_runtime_error: self.run_log[-1]['handler'] = x['handler'] s = 'Unrecoverable error for handler: {}'.format(x['handler']) raise NonRecoverableError(s, True, x['handler']) for x in self.run_log[-1]['corrections']: if not x['actions']: self.run_log[-1]['handler'] = x['handler'] s = 'Unrecoverable error for handler: %s' % x['handler'] raise NonRecoverableError(s, False, x['handler']) if self.errors_current_job >= self.max_errors_per_job: self.run_log[-1]['max_errors_per_job'] = True msg = 'Max errors per job reached: {}.'.format(self.max_errors_per_job) logger.info(msg) raise MaxCorrectionsPerJobError(msg, True, self.max_errors_per_job, job ) else: self.run_log[-1]['max_errors'] = True msg = 'Max errors reached: {}.'.format(self.max_errors) logger.info(msg) raise MaxCorrectionsError(msg, True, self.max_errors)
run_job
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, job_n, job):\n self.run_log.append({'job': job.as_dict(), 'corrections': [], 'handler':\n None, 'validator': None, 'max_errors': False, 'max_er...
def x(self): start = datetime.datetime.now() try: cwd = os.getcwd() v = sys.version.replace('\n', ' ') logger.info('Custodian started in singleshot mode at {} in {}.'. format(start, cwd)) logger.info('Custodian running on Python version {}'.format(v)) if os.path.exists(Custodian.LOG_FILE): self.run_log = loadfn(Custodian.LOG_FILE, cls=MontyDecoder) if len(self.run_log) == 0: job_n = 0 job = self.jobs[job_n] logger.info('Setting up job no. 1 ({}) '.format(job.name)) job.setup() self.run_log.append({'job': job.as_dict(), 'corrections': [], 'job_n': job_n}) return len(self.jobs) else: job_n = self.run_log[-1]['job_n'] job = self.jobs[job_n] if len(self.run_log[-1]['corrections']) > 0: logger.info('Reran {}.run due to fixable errors'.format(job .name)) logger.info('Checking error handlers for {}.run'.format(job.name)) if self._do_check(self.handlers): logger.info('Failed validation based on error handlers') for x in self.run_log[-1]['corrections']: if not x['actions'] and x['handler'].raises_runtime_error: self.run_log[-1]['handler'] = x['handler'] s = ( 'Unrecoverable error for handler: {}. Raising RuntimeError' .format(x['handler'])) raise NonRecoverableError(s, True, x['handler']) logger.info('Corrected input based on error handlers') return len(self.jobs) - job_n logger.info('Checking validator for {}.run'.format(job.name)) for v in self.validators: if v.check(): self.run_log[-1]['validator'] = v logger.info('Failed validation based on validator') s = 'Validation failed: {}'.format(v) raise ValidationError(s, True, v) logger.info('Postprocessing for {}.run'.format(job.name)) job.postprocess() if len(self.jobs) == job_n + 1: self.finished = True return 0 job_n += 1 job = self.jobs[job_n] self.run_log.append({'job': job.as_dict(), 'corrections': [], 'job_n': job_n}) job.setup() return len(self.jobs) - job_n except CustodianError as ex: logger.error(ex.message) if ex.raises: raise finally: logger.info('Logging to {}...'.format(Custodian.LOG_FILE)) dumpfn(self.run_log, Custodian.LOG_FILE, cls=MontyEncoder, indent=4) end = datetime.datetime.now() logger.info('Run ended at {}.'.format(end)) run_time = end - start logger.info('Run completed. Total time taken = {}.'.format(run_time)) if self.finished and self.gzipped_output: gzip_dir('.')
run_interrupted
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n start = datetime.datetime.now()\n try:\n cwd = os.getcwd()\n v = sys.version.replace('\\n', ' ')\n logger.info('Custodian s...
def x(self, handlers, terminate_func=None): corrections = [] for h in handlers: try: if h.check(): if (h.max_num_corrections is not None and h. n_applied_corrections >= h.max_num_corrections): msg = ( 'Maximum number of corrections {} reached for handler {}' .format(h.max_num_corrections, h)) if h.raise_on_max: self.run_log[-1]['handler'] = h self.run_log[-1]['max_errors_per_handler'] = True raise MaxCorrectionsPerHandlerError(msg, True, h. max_num_corrections, h) else: logger.warning(msg + ' Correction not applied.') continue if terminate_func is not None and h.is_terminating: logger.info('Terminating job') terminate_func() terminate_func = None d = h.correct() d['handler'] = h logger.error('\n' + pformat(d, indent=2, width=-1)) corrections.append(d) h.n_applied_corrections += 1 except Exception: if not self.skip_over_errors: raise else: import traceback logger.error('Bad handler %s ' % h) logger.error(traceback.format_exc()) corrections.append({'errors': ['Bad handler %s ' % h], 'actions': []}) self.total_errors += len(corrections) self.errors_current_job += len(corrections) self.run_log[-1]['corrections'].extend(corrections) dumpfn(self.run_log, Custodian.LOG_FILE, cls=MontyEncoder, indent=4) return len(corrections) > 0
do_check
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, handlers, terminate_func=None):\n corrections = []\n for h in handlers:\n try:\n if h.check():\n if (h.max_num_cor...
def x(self, actions): modified = [] for a in actions: if 'dict' in a: k = a['dict'] modified.append(k) self.feffinp[k] = self.modify_object(a['action'], self.feffinp[k]) elif 'file' in a: self.modify(a['action'], a['file']) else: raise ValueError('Unrecognized format: {}'.format(a)) if modified: feff = self.feffinp feff_input = '\n\n'.join(str(feff[k]) for k in ['HEADER', 'PARAMETERS', 'POTENTIALS', 'ATOMS'] if k in feff) for k, v in six.iteritems(feff): with open(os.path.join('.', k), 'w') as f: f.write(str(v)) with open(os.path.join('.', 'feff.inp'), 'w') as f: f.write(feff_input)
apply_actions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, actions):\n modified = []\n for a in actions:\n if 'dict' in a:\n k = a['dict']\n modified.append(k)\n self...
def x(filename, settings): if len(settings) != 1: raise ValueError( "Settings must only contain one item with key 'content'.") for k, v in settings.items(): if k == 'content': with open(filename, 'w') as f: f.write(v)
file_create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, settings):\n if len(settings) != 1:\n raise ValueError(\n \"Settings must only contain one item with key 'content'.\")\n for ...
def x(filename, settings): if len(settings) != 1: raise ValueError("Settings must only contain one item with key 'dest'." ) for k, v in settings.items(): if k == 'dest': shutil.move(filename, v)
file_move
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, settings):\n if len(settings) != 1:\n raise ValueError(\"Settings must only contain one item with key 'dest'.\"\n )\n for k, ...
def x(filename, settings): if len(settings) != 1: raise ValueError("Settings must only contain one item with key 'mode'." ) for k, v in settings.items(): if k == 'mode' and v == 'actual': try: os.remove(filename) except OSError: pass elif k == 'mode' and v == 'simulated': print('Simulated removal of {}'.format(filename))
file_delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, settings):\n if len(settings) != 1:\n raise ValueError(\"Settings must only contain one item with key 'mode'.\"\n )\n for k, ...
def x(filename, settings): for k, v in settings.items(): if k.startswith('dest'): shutil.copyfile(filename, v)
file_copy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, settings):\n for k, v in settings.items():\n if k.startswith('dest'):\n shutil.copyfile(filename, v)\n", "role": "user" },...
def x(filename, settings): for k, v in settings.items(): if k == 'mode': os.chmod(filename, v) if k == 'owners': os.chown(filename, v)
file_modify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, settings):\n for k, v in settings.items():\n if k == 'mode':\n os.chmod(filename, v)\n if k == 'owners':\n os....
def x(self): decompress_dir('.') if self.backup: for f in FEFF_INPUT_FILES: shutil.copy(f, '{}.orig'.format(f)) for f in FEFF_BACKUP_FILES: if os.path.isfile(f): shutil.copy(f, '{}.orig'.format(f))
setup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n decompress_dir('.')\n if self.backup:\n for f in FEFF_INPUT_FILES:\n shutil.copy(f, '{}.orig'.format(f))\n for f in FEF...
def x(self): with open(self.output_file, 'w') as f_std, open(self.stderr_file, 'w', buffering=1) as f_err: p = subprocess.Popen(self.feff_cmd, stdout=f_std, stderr=f_err, shell=True) return p
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n with open(self.output_file, 'w') as f_std, open(self.stderr_file, 'w',\n buffering=1) as f_err:\n p = subprocess.Popen(self.feff_cmd,...
def x(self, modification, obj): for action, settings in modification.items(): if action in self.supported_actions: self.supported_actions[action].__call__(obj, settings) elif self.strict: raise ValueError('{} is not a supported action!'.format(action))
modify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, modification, obj):\n for action, settings in modification.items():\n if action in self.supported_actions:\n self.supported_actions[...
def x(self, modification, obj): d = obj.as_dict() self.modify(modification, d) return obj.from_dict(d)
modify_object
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, modification, obj):\n d = obj.as_dict()\n self.modify(modification, d)\n return obj.from_dict(d)\n", "role": "user" }, { "content": ...
def x(self, actions): modified = [] for a in actions: if 'dict' in a: k = a['dict'] modified.append(k) self.vi[k] = self.modify_object(a['action'], self.vi[k]) elif 'file' in a: self.modify(a['action'], a['file']) else: raise ValueError('Unrecognized format: {}'.format(a)) for f in modified: self.vi[f].write_file(f)
apply_actions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, actions):\n modified = []\n for a in actions:\n if 'dict' in a:\n k = a['dict']\n modified.append(k)\n self...
def x(filenames, prefix='error'): num = max([0] + [int(f.split('.')[1]) for f in glob('{}.*.tar.gz'. format(prefix))]) filename = '{}.{}.tar.gz'.format(prefix, num + 1) logging.info('Backing up run to {}.'.format(filename)) with tarfile.open(filename, 'w:gz') as tar: for fname in filenames: for f in glob(fname): tar.add(f)
backup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filenames, prefix='error'):\n num = max([0] + [int(f.split('.')[1]) for f in glob('{}.*.tar.gz'.\n format(prefix))])\n filename = '{}.{}.tar.gz'.f...
def x(): host = os.environ.get('HOSTNAME', None) cluster = os.environ.get('SGE_O_HOST', None) if host is None: try: import socket host = host or socket.gethostname() except: pass return host or 'unknown', cluster or 'unknown'
get_execution_host_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n host = os.environ.get('HOSTNAME', None)\n cluster = os.environ.get('SGE_O_HOST', None)\n if host is None:\n try:\n import socke...
def x(self): qclog = open(self.qclog_file, 'w') p = subprocess.Popen(self.current_command, stdout=qclog) return p
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n qclog = open(self.qclog_file, 'w')\n p = subprocess.Popen(self.current_command, stdout=qclog)\n return p\n", "role": "user" }, { ...
def x(cls, qchem_command, multimode='openmp', input_file='mol.qin', output_file='mol.qout', qclog_file='mol.qclog', max_iterations=10, max_molecule_perturb_scale=0.3, check_connectivity=True, **QCJob_kwargs): min_molecule_perturb_scale = 0.1 scale_grid = 10 perturb_scale_grid = (max_molecule_perturb_scale - min_molecule_perturb_scale) / scale_grid if not os.path.exists(input_file): raise AssertionError('Input file must be present!') orig_opt_input = QCInput.from_file(input_file) orig_opt_rem = copy.deepcopy(orig_opt_input.rem) orig_freq_rem = copy.deepcopy(orig_opt_input.rem) orig_freq_rem['job_type'] = 'freq' first = True reversed_direction = False num_neg_freqs = [] for ii in range(max_iterations): yield QCJob(qchem_command=qchem_command, multimode=multimode, input_file=input_file, output_file=output_file, qclog_file= qclog_file, suffix='.opt_' + str(ii), backup=first, **QCJob_kwargs) first = False opt_outdata = QCOutput(output_file + '.opt_' + str(ii)).data if opt_outdata['structure_change' ] == 'unconnected_fragments' and not opt_outdata['completion']: print( 'Unstable molecule broke into unconnected fragments which failed to optimize! Exiting...' ) break else: freq_QCInput = QCInput(molecule=opt_outdata.get( 'molecule_from_optimized_geometry'), rem=orig_freq_rem, opt =orig_opt_input.opt, pcm=orig_opt_input.pcm, solvent= orig_opt_input.solvent) freq_QCInput.write_file(input_file) yield QCJob(qchem_command=qchem_command, multimode=multimode, input_file=input_file, output_file=output_file, qclog_file= qclog_file, suffix='.freq_' + str(ii), backup=first, ** QCJob_kwargs) outdata = QCOutput(output_file + '.freq_' + str(ii)).data errors = outdata.get('errors') if len(errors) != 0: raise AssertionError( 'No errors should be encountered while flattening frequencies!' ) if outdata.get('frequencies')[0] > 0.0: print('All frequencies positive!') break else: num_neg_freqs += [sum(1 for freq in outdata.get( 'frequencies') if freq < 0)] if len(num_neg_freqs) > 1: if num_neg_freqs[-1] == num_neg_freqs[-2 ] and not reversed_direction: reversed_direction = True elif num_neg_freqs[-1] == num_neg_freqs[-2 ] and reversed_direction: if len(num_neg_freqs) < 3: raise AssertionError( 'ERROR: This should only be possible after at least three frequency flattening iterations! Exiting...' ) else: raise Exception( 'ERROR: Reversing the perturbation direction still could not flatten any frequencies. Exiting...' ) elif num_neg_freqs[-1] != num_neg_freqs[-2 ] and reversed_direction: reversed_direction = False negative_freq_vecs = outdata.get('frequency_mode_vectors')[0] structure_successfully_perturbed = False for molecule_perturb_scale in np.arange( max_molecule_perturb_scale, min_molecule_perturb_scale, -perturb_scale_grid): new_coords = perturb_coordinates(old_coords=outdata.get ('initial_geometry'), negative_freq_vecs= negative_freq_vecs, molecule_perturb_scale= molecule_perturb_scale, reversed_direction= reversed_direction) new_molecule = Molecule(species=outdata.get('species'), coords=new_coords, charge=outdata.get('charge'), spin_multiplicity=outdata.get('multiplicity')) if check_connectivity: old_molgraph = MoleculeGraph.with_local_env_strategy( outdata.get('initial_molecule'), OpenBabelNN(), reorder=False, extend_structure=False) new_molgraph = MoleculeGraph.with_local_env_strategy( new_molecule, OpenBabelNN(), reorder=False, extend_structure=False) if old_molgraph.isomorphic_to(new_molgraph): structure_successfully_perturbed = True break if not structure_successfully_perturbed: raise Exception( 'ERROR: Unable to perturb coordinates to remove negative frequency without changing the connectivity! Exiting...' ) new_opt_QCInput = QCInput(molecule=new_molecule, rem= orig_opt_rem, opt=orig_opt_input.opt, pcm= orig_opt_input.pcm, solvent=orig_opt_input.solvent) new_opt_QCInput.write_file(input_file)
opt_with_frequency_flattener
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, qchem_command, multimode='openmp', input_file='mol.qin',\n output_file='mol.qout', qclog_file='mol.qclog', max_iterations=10,\n max_molecule_perturb...
def x(self): decompress_dir('.') if self.backup: for f in VASP_INPUT_FILES: shutil.copy(f, '{}.orig'.format(f)) if self.auto_npar: try: incar = Incar.from_file('INCAR') if not (incar.get('LHFCALC') or incar.get('LRPA') or incar.get( 'LEPSILON')): if incar.get('IBRION') in [5, 6, 7, 8]: del incar['NPAR'] else: import multiprocessing ncores = os.environ.get('NSLOTS' ) or multiprocessing.cpu_count() ncores = int(ncores) for npar in range(int(math.sqrt(ncores)), ncores): if ncores % npar == 0: incar['NPAR'] = npar break incar.write_file('INCAR') except: pass if self.auto_continue: if os.path.exists('continue.json'): actions = loadfn('continue.json').get('actions') logger.info('Continuing previous VaspJob. Actions: {}'.format( actions)) backup(VASP_BACKUP_FILES, prefix='prev_run') VaspModder().apply_actions(actions) else: if self.auto_continue is True: actions = [{'file': 'CONTCAR', 'action': {'_file_copy': { 'dest': 'POSCAR'}}}, {'dict': 'INCAR', 'action': { '_set': {'ISTART': 1}}}] else: actions = self.auto_continue dumpfn({'actions': actions}, 'continue.json') if self.settings_override is not None: VaspModder().apply_actions(self.settings_override)
setup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n decompress_dir('.')\n if self.backup:\n for f in VASP_INPUT_FILES:\n shutil.copy(f, '{}.orig'.format(f))\n if self.auto_npa...
def x(self): cmd = list(self.vasp_cmd) if self.auto_gamma: vi = VaspInput.from_directory('.') kpts = vi['KPOINTS'] if kpts.style == Kpoints.supported_modes.Gamma and tuple(kpts.kpts[0] ) == (1, 1, 1): if self.gamma_vasp_cmd is not None and which(self. gamma_vasp_cmd[-1]): cmd = self.gamma_vasp_cmd elif which(cmd[-1] + '.gamma'): cmd[-1] += '.gamma' logger.info('Running {}'.format(' '.join(cmd))) with open(self.output_file, 'w') as f_std, open(self.stderr_file, 'w', buffering=1) as f_err: p = subprocess.Popen(cmd, stdout=f_std, stderr=f_err) return p
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cmd = list(self.vasp_cmd)\n if self.auto_gamma:\n vi = VaspInput.from_directory('.')\n kpts = vi['KPOINTS']\n if kpts.style...
def x(self): for f in (VASP_OUTPUT_FILES + [self.output_file]): if os.path.exists(f): if self.final and self.suffix != '': shutil.move(f, '{}{}'.format(f, self.suffix)) elif self.suffix != '': shutil.copy(f, '{}{}'.format(f, self.suffix)) if self.copy_magmom and not self.final: try: outcar = Outcar('OUTCAR') magmom = [m['tot'] for m in outcar.magnetization] incar = Incar.from_file('INCAR') incar['MAGMOM'] = magmom incar.write_file('INCAR') except: logger.error('MAGMOM copy from OUTCAR to INCAR failed') if os.path.exists('continue.json'): os.remove('continue.json')
postprocess
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for f in (VASP_OUTPUT_FILES + [self.output_file]):\n if os.path.exists(f):\n if self.final and self.suffix != '':\n ...
def x(cls, vasp_cmd, auto_npar=True, ediffg=-0.05, half_kpts_first_relax= False, auto_continue=False): incar_update = {'ISTART': 1} if ediffg: incar_update['EDIFFG'] = ediffg settings_overide_1 = None settings_overide_2 = [{'dict': 'INCAR', 'action': {'_set': incar_update }}, {'file': 'CONTCAR', 'action': {'_file_copy': {'dest': 'POSCAR'}}}] if half_kpts_first_relax and os.path.exists('KPOINTS') and os.path.exists( 'POSCAR'): kpts = Kpoints.from_file('KPOINTS') orig_kpts_dict = kpts.as_dict() kpts.kpts = np.round(np.maximum(np.array(kpts.kpts) / 2, 1)).astype(int ).tolist() low_kpts_dict = kpts.as_dict() settings_overide_1 = [{'dict': 'KPOINTS', 'action': {'_set': low_kpts_dict}}] settings_overide_2.append({'dict': 'KPOINTS', 'action': {'_set': orig_kpts_dict}}) return [VaspJob(vasp_cmd, final=False, suffix='.relax1', auto_npar= auto_npar, auto_continue=auto_continue, settings_override= settings_overide_1), VaspJob(vasp_cmd, final=True, backup=False, suffix='.relax2', auto_npar=auto_npar, auto_continue=auto_continue, settings_override=settings_overide_2)]
double_relaxation_run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, vasp_cmd, auto_npar=True, ediffg=-0.05, half_kpts_first_relax=\n False, auto_continue=False):\n incar_update = {'ISTART': 1}\n if ediffg:\n ...
def x(cls, vasp_cmd, auto_npar=True, ediffg=-0.05, half_kpts_first_relax= False, auto_continue=False): incar = Incar.from_file('INCAR') metaGGA = incar.get('METAGGA', 'SCAN') pre_opt_setings = [{'dict': 'INCAR', 'action': {'_set': {'METAGGA': None, 'LWAVE': True, 'NSW': 0}}}] jobs = [VaspJob(vasp_cmd, auto_npar=auto_npar, final=False, suffix= '.precondition', settings_override=pre_opt_setings)] jobs.extend(VaspJob.double_relaxation_run(vasp_cmd, auto_npar=auto_npar, ediffg=ediffg, half_kpts_first_relax=half_kpts_first_relax)) jobs[1].backup = False post_opt_settings = [{'dict': 'INCAR', 'action': {'_set': {'METAGGA': metaGGA, 'ISTART': 1, 'NSW': incar.get('NSW', 99), 'LWAVE': incar. get('LWAVE', False)}}}, {'file': 'CONTCAR', 'action': {'_file_copy': {'dest': 'POSCAR'}}}] if jobs[1].settings_override: post_opt_settings = jobs[1].settings_override + post_opt_settings jobs[1].settings_override = post_opt_settings return jobs
metagga_opt_run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, vasp_cmd, auto_npar=True, ediffg=-0.05, half_kpts_first_relax=\n False, auto_continue=False):\n incar = Incar.from_file('INCAR')\n metaGGA = inca...
def x(cls, vasp_cmd, vol_change_tol=0.02, max_steps=10, ediffg=-0.05, half_kpts_first_relax=False, **vasp_job_kwargs): for i in range(max_steps): if i == 0: settings = None backup = True if half_kpts_first_relax and os.path.exists('KPOINTS' ) and os.path.exists('POSCAR'): kpts = Kpoints.from_file('KPOINTS') orig_kpts_dict = kpts.as_dict() kpts.kpts = np.maximum(np.array(kpts.kpts) / 2, 1).tolist() low_kpts_dict = kpts.as_dict() settings = [{'dict': 'KPOINTS', 'action': {'_set': low_kpts_dict}}] else: backup = False initial = Poscar.from_file('POSCAR').structure final = Poscar.from_file('CONTCAR').structure vol_change = (final.volume - initial.volume) / initial.volume logger.info('Vol change = %.1f %%!' % (vol_change * 100)) if abs(vol_change) < vol_change_tol: logger.info('Stopping optimization!') break else: incar_update = {'ISTART': 1} if ediffg: incar_update['EDIFFG'] = ediffg settings = [{'dict': 'INCAR', 'action': {'_set': incar_update}}, {'file': 'CONTCAR', 'action': { '_file_copy': {'dest': 'POSCAR'}}}] if i == 1 and half_kpts_first_relax: settings.append({'dict': 'KPOINTS', 'action': {'_set': orig_kpts_dict}}) logger.info('Generating job = %d!' % (i + 1)) yield VaspJob(vasp_cmd, final=False, backup=backup, suffix= '.relax%d' % (i + 1), settings_override=settings, **vasp_job_kwargs )
full_opt_run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, vasp_cmd, vol_change_tol=0.02, max_steps=10, ediffg=-0.05,\n half_kpts_first_relax=False, **vasp_job_kwargs):\n for i in range(max_steps):\n ...
def x(cls, vasp_cmd, lattice_direction, initial_strain, atom_relax=True, max_steps=20, algo='bfgs', **vasp_job_kwargs): nsw = 99 if atom_relax else 0 incar = Incar.from_file('INCAR') if incar.get('EDIFFG') and incar.get('EDIFFG') > 0: etol = incar['EDIFFG'] else: etol = incar.get('EDIFF', 0.0001) * 10 if lattice_direction == 'a': lattice_index = 0 elif lattice_direction == 'b': lattice_index = 1 else: lattice_index = 2 energies = {} for i in range(max_steps): if i == 0: settings = [{'dict': 'INCAR', 'action': {'_set': {'ISIF': 2, 'NSW': nsw}}}] structure = Poscar.from_file('POSCAR').structure x = structure.lattice.abc[lattice_index] backup = True else: backup = False v = Vasprun('vasprun.xml') structure = v.final_structure energy = v.final_energy lattice = structure.lattice x = lattice.abc[lattice_index] energies[x] = energy if i == 1: x *= 1 + initial_strain else: min_x = min(energies.keys(), key=lambda e: energies[e]) sorted_x = sorted(energies.keys()) ind = sorted_x.index(min_x) if ind == 0: other = ind + 1 elif ind == len(sorted_x) - 1: other = ind - 1 else: other = ind + 1 if energies[sorted_x[ind + 1]] < energies[ sorted_x[ind - 1]] else ind - 1 if abs(energies[min_x] - energies[sorted_x[other]]) < etol: logger.info('Stopping optimization! Final %s = %f' % ( lattice_direction, min_x)) break if ind == 0 and len(sorted_x) > 2: x = sorted_x[0] - abs(sorted_x[1] - sorted_x[0]) logger.info( 'Lowest energy lies below bounds. Setting %s = %f.' % (lattice_direction, x)) elif ind == len(sorted_x) - 1 and len(sorted_x) > 2: x = sorted_x[-1] + abs(sorted_x[-1] - sorted_x[-2]) logger.info( 'Lowest energy lies above bounds. Setting %s = %f.' % (lattice_direction, x)) elif algo.lower() == 'bfgs' and len(sorted_x) >= 4: try: x1 = list(energies.keys()) y1 = [energies[j] for j in x1] z1 = np.polyfit(x1, y1, 2) pp = np.poly1d(z1) from scipy.optimize import minimize result = minimize(pp, min_x, bounds=[(sorted_x[0], sorted_x[-1])]) if not result.success or result.x[0] < 0: raise ValueError('Negative lattice constant!') x = result.x[0] logger.info('BFGS minimized %s = %f.' % ( lattice_direction, x)) except ValueError as ex: logger.info(str(ex)) x = (min_x + sorted_x[other]) / 2 logger.info('Falling back on bisection %s = %f.' % (lattice_direction, x)) else: x = (min_x + sorted_x[other]) / 2 logger.info('Bisection %s = %f.' % (lattice_direction, x)) lattice = lattice.matrix lattice[lattice_index] = lattice[lattice_index] / np.linalg.norm( lattice[lattice_index]) * x s = Structure(lattice, structure.species, structure.frac_coords) fname = 'POSCAR.%f' % x s.to(filename=fname) incar_update = {'ISTART': 1, 'NSW': nsw, 'ISIF': 2} settings = [{'dict': 'INCAR', 'action': {'_set': incar_update}}, {'file': fname, 'action': {'_file_copy': {'dest': 'POSCAR'}}}] logger.info('Generating job = %d with parameter %f!' % (i + 1, x)) yield VaspJob(vasp_cmd, final=False, backup=backup, suffix= '.static.%f' % x, settings_override=settings, **vasp_job_kwargs) with open('EOS.txt', 'wt') as f: f.write('# %s energy\n' % lattice_direction) for k in sorted(energies.keys()): f.write('%f %f\n' % (k, energies[k]))
constrained_opt_run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, vasp_cmd, lattice_direction, initial_strain, atom_relax=True,\n max_steps=20, algo='bfgs', **vasp_job_kwargs):\n nsw = 99 if atom_relax else 0\n ...
def x(self): neb_dirs = self.neb_dirs if self.backup: for f in VASP_NEB_INPUT_FILES: shutil.copy(f, '{}.orig'.format(f)) for path in neb_dirs: poscar = os.path.join(path, 'POSCAR') shutil.copy(poscar, '{}.orig'.format(poscar)) if self.half_kpts and os.path.exists('KPOINTS'): kpts = Kpoints.from_file('KPOINTS') kpts.kpts = np.maximum(np.array(kpts.kpts) / 2, 1) kpts.kpts = kpts.kpts.astype(int).tolist() if tuple(kpts.kpts[0]) == (1, 1, 1): kpt_dic = kpts.as_dict() kpt_dic['generation_style'] = 'Gamma' kpts = Kpoints.from_dict(kpt_dic) kpts.write_file('KPOINTS') if self.auto_npar: try: incar = Incar.from_file('INCAR') import multiprocessing ncores = os.environ.get('NSLOTS') or multiprocessing.cpu_count() ncores = int(ncores) for npar in range(int(math.sqrt(ncores)), ncores): if ncores % npar == 0: incar['NPAR'] = npar break incar.write_file('INCAR') except: pass if self.auto_continue and os.path.exists('STOPCAR') and not os.access( 'STOPCAR', os.W_OK): os.chmod('STOPCAR', 420) os.remove('STOPCAR') for path in self.neb_sub: contcar = os.path.join(path, 'CONTCAR') poscar = os.path.join(path, 'POSCAR') shutil.copy(contcar, poscar) if self.settings_override is not None: VaspModder().apply_actions(self.settings_override)
setup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n neb_dirs = self.neb_dirs\n if self.backup:\n for f in VASP_NEB_INPUT_FILES:\n shutil.copy(f, '{}.orig'.format(f))\n for...
def x(self): for path in self.neb_dirs: for f in VASP_NEB_OUTPUT_SUB_FILES: f = os.path.join(path, f) if os.path.exists(f): if self.final and self.suffix != '': shutil.move(f, '{}{}'.format(f, self.suffix)) elif self.suffix != '': shutil.copy(f, '{}{}'.format(f, self.suffix)) for f in (VASP_NEB_OUTPUT_FILES + [self.output_file]): if os.path.exists(f): if self.final and self.suffix != '': shutil.move(f, '{}{}'.format(f, self.suffix)) elif self.suffix != '': shutil.copy(f, '{}{}'.format(f, self.suffix))
postprocess
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for path in self.neb_dirs:\n for f in VASP_NEB_OUTPUT_SUB_FILES:\n f = os.path.join(path, f)\n if os.path.exists(f):\n...
def x(self): if self.backup: shutil.copy(self.input_file, '{}.orig'.format(self.input_file))
setup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.backup:\n shutil.copy(self.input_file, '{}.orig'.format(self.input_file))\n", "role": "user" }, { "content": "setup", ...
def x(self): with zopen(self.output_file, 'w') as fout: return subprocess.Popen(self.nwchem_cmd + [self.input_file], stdout =fout)
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n with zopen(self.output_file, 'w') as fout:\n return subprocess.Popen(self.nwchem_cmd + [self.input_file], stdout\n =fout)\n", ...
def x(x): x = float(x) if x < 0.0 or x > 1.0: raise ArgumentTypeError('{} not in range [0.0, 1.0]'.format(x)) return x
valid_GC
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x):\n x = float(x)\n if x < 0.0 or x > 1.0:\n raise ArgumentTypeError('{} not in range [0.0, 1.0]'.format(x))\n return x\n", "role": "user"...
def x(fq, args): if args.quality: quality_check = ave_qual else: quality_check = silent_quality_check minlen = args.length + int(args.headcrop or 0) - int(args.tailcrop or 0) for rec in SeqIO.parse(fq, 'fastq'): if args.GC_filter: gc = (rec.seq.upper().count('C') + rec.seq.upper().count('G') ) / len(rec) else: gc = 0.5 if quality_check(rec.letter_annotations['phred_quality'] ) > args.quality and minlen <= len(rec ) <= args.maxlength and args.minGC <= gc <= args.maxGC: print(rec[args.headcrop:args.tailcrop].format('fastq'), end='')
filter_stream
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fq, args):\n if args.quality:\n quality_check = ave_qual\n else:\n quality_check = silent_quality_check\n minlen = args.length + int(arg...
def x(fq, args): data = {entry[0]: entry[1] for entry in process_summary(summaryfile= args.summary, threads='NA', readtype=args.readtype, barcoded=False) [['readIDs', 'quals']].itertuples(index=False)} try: for record in SeqIO.parse(fq, 'fastq'): if data[record.id] > args.quality and args.length <= len(record ) <= args.maxlength: print(record[args.headcrop:args.tailcrop].format('fastq'), end='') except KeyError: logging.error( 'mismatch between summary and fastq: {} was not found in the summary file.' .format(record.id)) sys.exit( """ ERROR: mismatch between sequencing_summary and fastq file: {} was not found in the summary file. Quitting.""" .format(record.id))
filter_using_summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fq, args):\n data = {entry[0]: entry[1] for entry in process_summary(summaryfile=\n args.summary, threads='NA', readtype=args.readtype, barcoded=Fals...
def x(func): def x(obj, *args, **kw): conn = ACCESS_KEYS if not (conn and conn.get('master_key')): message = '%s requires the master key' % func.__name__ raise core.ParseError(message) func(obj, *args, **kw) return ret
master_key_required
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func):\n\n def x(obj, *args, **kw):\n conn = ACCESS_KEYS\n if not (conn and conn.get('master_key')):\n message = '%s requires the m...
def x(cls, uri, http_verb, extra_headers=None, batch=False, _body=None, **kw): if batch: urlsplitter = urlparse(API_ROOT).netloc ret = {'method': http_verb, 'path': uri.split(urlsplitter, 1)[1]} if kw: ret['body'] = kw return ret if not ('app_id' in ACCESS_KEYS and 'rest_key' in ACCESS_KEYS): raise core.ParseError('Missing connection credentials') app_id = ACCESS_KEYS.get('app_id') rest_key = ACCESS_KEYS.get('rest_key') master_key = ACCESS_KEYS.get('master_key') url = uri if uri.startswith(API_ROOT) else cls.ENDPOINT_ROOT + uri if _body is None: data = kw and json.dumps(kw, default=date_handler) or '{}' else: data = _body if http_verb == 'GET' and data: url += '?%s' % urlencode(kw) data = None elif cls.__name__ == 'File': data = data else: data = data.encode('utf-8') headers = {'Content-type': 'application/json', 'X-Parse-Application-Id': app_id, 'X-Parse-REST-API-Key': rest_key} headers.update(extra_headers or {}) if cls.__name__ == 'File': request = Request(url.encode('utf-8'), data, headers) else: request = Request(url, data, headers) if ACCESS_KEYS.get('session_token'): request.add_header('X-Parse-Session-Token', ACCESS_KEYS.get( 'session_token')) elif master_key: request.add_header('X-Parse-Master-Key', master_key) request.get_method = lambda : http_verb try: response = urlopen(request, timeout=CONNECTION_TIMEOUT) except HTTPError as e: exc = {(400): core.ResourceRequestBadRequest, (401): core. ResourceRequestLoginRequired, (403): core. ResourceRequestForbidden, (404): core.ResourceRequestNotFound}.get( e.code, core.ParseError) raise exc(e.read()) return json.loads(response.read().decode('utf-8'))
execute
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, uri, http_verb, extra_headers=None, batch=False, _body=None, **kw):\n if batch:\n urlsplitter = urlparse(API_ROOT).netloc\n ret = {'metho...
def x(self, methods): methods = list(methods) if not methods: return queries, callbacks = list(zip(*[m(batch=True) for m in methods])) responses = self.execute('', 'POST', requests=queries) batched_errors = [] for callback, response in zip(callbacks, responses): if 'success' in response: callback(response['success']) else: batched_errors.append(response['error']) if batched_errors: raise core.ParseBatchError(batched_errors)
batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, methods):\n methods = list(methods)\n if not methods:\n return\n queries, callbacks = list(zip(*[m(batch=True) for m in methods]))\n r...
def x(cls, installation_id, channels_to_add=set(), channels_to_remove=set(), **kw): installation_url = cls._get_installation_url(installation_id) current_config = cls.GET(installation_url) new_channels = list(set(current_config['channels']).union( channels_to_add).difference(channels_to_remove)) cls.PUT(installation_url, channels=new_channels)
update_channels
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, installation_id, channels_to_add=set(), channels_to_remove=set(),\n **kw):\n installation_url = cls._get_installation_url(installation_id)\n curr...
def x(self, count=False): if self._result_cache is not None: return len(self._result_cache) if count else self._result_cache """ Return a list of objects matching query, or if count == True return only the number of objects matching. """ options = dict(self._options) if self._where: options['where'] = json.dumps(self._where) if self._select_related: options['include'] = ','.join(self._select_related) if count: return self._manager._count(**options) self._result_cache = self._manager._fetch(**options) return self._result_cache
fetch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, count=False):\n if self._result_cache is not None:\n return len(self._result_cache) if count else self._result_cache\n \"\"\"\n Retur...
def x(name=None): def x(cls): ParseType.type_mapping[name or cls.__name__] = cls return cls return wrapped
complex_type
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name=None):\n\n def x(cls):\n ParseType.type_mapping[name or cls.__name__] = cls\n return cls\n return wrapped\n", "role": "user" }, ...
def x(cls, class_name): class_name = str(class_name.lstrip('_')) types = ParseResource.__subclasses__() while types: t = types.pop() if t.__name__ == class_name: return t types.extend(t.__subclasses__()) else: return type(class_name, (Object,), {})
factory
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, class_name):\n class_name = str(class_name.lstrip('_'))\n types = ParseResource.__subclasses__()\n while types:\n t = types.pop()\n ...
def x(cls): root = '/'.join([API_ROOT, 'schemas', cls.__name__]) schema = cls.GET(root) return schema
schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls):\n root = '/'.join([API_ROOT, 'schemas', cls.__name__])\n schema = cls.GET(root)\n return schema\n", "role": "user" }, { "content": "...
def x(cls, key): root = '/'.join([API_ROOT, 'schemas', cls.__name__]) payload = {'className': cls.__name__, 'fields': {key: {'__op': 'Delete'}}} cls.PUT(root, **payload)
schema_delete_field
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, key):\n root = '/'.join([API_ROOT, 'schemas', cls.__name__])\n payload = {'className': cls.__name__, 'fields': {key: {'__op': 'Delete'}}}\n cls.P...
def x(self, key, amount=1): payload = {key: {'__op': 'Increment', 'amount': amount}} self.__class__.PUT(self._absolute_url, **payload) self.__dict__[key] += amount
increment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, amount=1):\n payload = {key: {'__op': 'Increment', 'amount': amount}}\n self.__class__.PUT(self._absolute_url, **payload)\n self.__dict__[k...
def x(self, key): payload = {key: {'__op': 'Delete'}} self.__class__.PUT(self._absolute_url, **payload) del self.__dict__[key]
remove
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key):\n payload = {key: {'__op': 'Delete'}}\n self.__class__.PUT(self._absolute_url, **payload)\n del self.__dict__[key]\n", "role": "user" ...
def x(func): def x(obj, *args, **kw): if not hasattr(obj, 'sessionToken'): message = '%s requires a logged-in session' % func.__name__ raise ResourceRequestLoginRequired(message) return func(obj, *args, **kw) return ret
login_required
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func):\n\n def x(obj, *args, **kw):\n if not hasattr(obj, 'sessionToken'):\n message = '%s requires a logged-in session' % func.__name__\n...
def x(email): url = '/'.join([API_ROOT, 'requestPasswordReset']) try: User.POST(url, email=email) return True except ParseError: return False
request_password_reset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(email):\n url = '/'.join([API_ROOT, 'requestPasswordReset'])\n try:\n User.POST(url, email=email)\n return True\n except ParseError:\n ...
def x(d): res = {} for k, v in iteritems(d): if isinstance(v, string_types) and DATETIME_ISO_FORMAT.match(v): v = dateutil.parser.parse(v) res[k] = v return res
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d):\n res = {}\n for k, v in iteritems(d):\n if isinstance(v, string_types) and DATETIME_ISO_FORMAT.match(v):\n v = dateutil.parser.par...
def x(obj, path, **kw): open_kw = {'mode': 'w'} if PY3: open_kw['encoding'] = 'utf-8' if kw.get('indent') and kw.get('separators') is None: kw['separators'] = ',', ': ' with open(str(path), **open_kw) as fp: return json.dump(obj, fp, **kw)
dump
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj, path, **kw):\n open_kw = {'mode': 'w'}\n if PY3:\n open_kw['encoding'] = 'utf-8'\n if kw.get('indent') and kw.get('separators') is None:\n...
def x(path, **kw): _kw = {} if PY3: _kw['encoding'] = 'utf-8' with open(str(path), **_kw) as fp: return json.load(fp, **kw)
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, **kw):\n _kw = {}\n if PY3:\n _kw['encoding'] = 'utf-8'\n with open(str(path), **_kw) as fp:\n return json.load(fp, **kw)\n", ...
def x(text, brackets=None): res = [] for c, type_ in _tokens(text, brackets=brackets): if type_ == TextType.text: res.append(c) return ''.join(res).strip()
strip_brackets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, brackets=None):\n res = []\n for c, type_ in _tokens(text, brackets=brackets):\n if type_ == TextType.text:\n res.append(c)\n ...
def x(text, separators=WHITESPACE, brackets=None): res, chunk = [], [] for c, type_ in _tokens(text, brackets=brackets): if type_ == TextType.text and c in separators: res.append(''.join(chunk).strip()) chunk = [] else: chunk.append(c) res.append(''.join(chunk).strip()) return nfilter(res)
split_text_with_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, separators=WHITESPACE, brackets=None):\n res, chunk = [], []\n for c, type_ in _tokens(text, brackets=brackets):\n if type_ == TextType.text...
def x(text, separators=re.compile('\\s'), brackets=None, strip=False): if not isinstance(separators, PATTERN_TYPE): separators = re.compile('[{0}]'.format(''.join('\\{0}'.format(c) for c in separators))) return nfilter(s.strip() if strip else s for s in separators.split( strip_brackets(text, brackets=brackets)))
split_text
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, separators=re.compile('\\\\s'), brackets=None, strip=False):\n if not isinstance(separators, PATTERN_TYPE):\n separators = re.compile('[{0}]'.f...