repo
stringlengths
7
54
path
stringlengths
4
192
url
stringlengths
87
284
code
stringlengths
78
104k
code_tokens
list
docstring
stringlengths
1
46.9k
docstring_tokens
list
language
stringclasses
1 value
partition
stringclasses
3 values
SBRG/ssbio
ssbio/pipeline/atlas2.py
https://github.com/SBRG/ssbio/blob/e9449e64ffc1a1f5ad07e5849aa12a650095f8a2/ssbio/pipeline/atlas2.py#L722-L785
def calculate_residue_counts_perstrain(protein_pickle_path, outdir, pdbflex_keys_file, wt_pid_cutoff=None, force_rerun=False): """Writes out a feather file for a PROTEIN counting amino acid occurences for ALL STRAINS along with SUBSEQUENCES""" from collections import defaultdict from ssbio.protein.sequence....
[ "def", "calculate_residue_counts_perstrain", "(", "protein_pickle_path", ",", "outdir", ",", "pdbflex_keys_file", ",", "wt_pid_cutoff", "=", "None", ",", "force_rerun", "=", "False", ")", ":", "from", "collections", "import", "defaultdict", "from", "ssbio", ".", "pr...
Writes out a feather file for a PROTEIN counting amino acid occurences for ALL STRAINS along with SUBSEQUENCES
[ "Writes", "out", "a", "feather", "file", "for", "a", "PROTEIN", "counting", "amino", "acid", "occurences", "for", "ALL", "STRAINS", "along", "with", "SUBSEQUENCES" ]
python
train
tornadoweb/tornado
tornado/web.py
https://github.com/tornadoweb/tornado/blob/b8b481770bcdb333a69afde5cce7eaa449128326/tornado/web.py#L951-L971
def render_linked_css(self, css_files: Iterable[str]) -> str: """Default method used to render the final css links for the rendered webpage. Override this method in a sub-classed controller to change the output. """ paths = [] unique_paths = set() # type: Set[str] ...
[ "def", "render_linked_css", "(", "self", ",", "css_files", ":", "Iterable", "[", "str", "]", ")", "->", "str", ":", "paths", "=", "[", "]", "unique_paths", "=", "set", "(", ")", "# type: Set[str]", "for", "path", "in", "css_files", ":", "if", "not", "i...
Default method used to render the final css links for the rendered webpage. Override this method in a sub-classed controller to change the output.
[ "Default", "method", "used", "to", "render", "the", "final", "css", "links", "for", "the", "rendered", "webpage", "." ]
python
train
jasonrbriggs/stomp.py
stomp/protocol.py
https://github.com/jasonrbriggs/stomp.py/blob/643843c5fbf25fd24339dd0e69a9411c3d8b94c7/stomp/protocol.py#L445-L454
def _escape_headers(self, headers): """ :param dict(str,str) headers: """ for key, val in headers.items(): try: val = val.replace('\\', '\\\\').replace('\n', '\\n').replace(':', '\\c').replace('\r', '\\r') except: pass h...
[ "def", "_escape_headers", "(", "self", ",", "headers", ")", ":", "for", "key", ",", "val", "in", "headers", ".", "items", "(", ")", ":", "try", ":", "val", "=", "val", ".", "replace", "(", "'\\\\'", ",", "'\\\\\\\\'", ")", ".", "replace", "(", "'\\...
:param dict(str,str) headers:
[ ":", "param", "dict", "(", "str", "str", ")", "headers", ":" ]
python
train
biocore/deblur
deblur/workflow.py
https://github.com/biocore/deblur/blob/4b4badccdbac8fe9d8f8b3f1349f3700e31b5d7b/deblur/workflow.py#L366-L493
def remove_artifacts_seqs(seqs_fp, ref_fp, working_dir, ref_db_fp, negate=False, threads=1, verbose=False, sim_thresh=None, ...
[ "def", "remove_artifacts_seqs", "(", "seqs_fp", ",", "ref_fp", ",", "working_dir", ",", "ref_db_fp", ",", "negate", "=", "False", ",", "threads", "=", "1", ",", "verbose", "=", "False", ",", "sim_thresh", "=", "None", ",", "coverage_thresh", "=", "None", "...
Remove artifacts from FASTA file using SortMeRNA. Parameters ---------- seqs_fp: string file path to FASTA input sequence file ref_fp: tuple file path(s) to FASTA database file working_dir: string working directory path ref_db_fp: tuple file path(s) to indexed FA...
[ "Remove", "artifacts", "from", "FASTA", "file", "using", "SortMeRNA", "." ]
python
train
humilis/humilis-lambdautils
lambdautils/state.py
https://github.com/humilis/humilis-lambdautils/blob/58f75eb5ace23523c283708d56a9193181ea7e8e/lambdautils/state.py#L137-L143
def get_secret(key, *args, **kwargs): """Retrieves a secret.""" env_value = os.environ.get(key.replace('.', '_').upper()) if not env_value: # Backwards compatibility: the deprecated secrets vault return _get_secret_from_vault(key, *args, **kwargs) return env_value
[ "def", "get_secret", "(", "key", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "env_value", "=", "os", ".", "environ", ".", "get", "(", "key", ".", "replace", "(", "'.'", ",", "'_'", ")", ".", "upper", "(", ")", ")", "if", "not", "env_va...
Retrieves a secret.
[ "Retrieves", "a", "secret", "." ]
python
train
wiredrive/wtframework
wtframework/wtf/web/web_utils.py
https://github.com/wiredrive/wtframework/blob/ef7f86c4d4cf7fb17745fd627b3cc4a41f4c0216/wtframework/wtf/web/web_utils.py#L57-L79
def get_base_url(webdriver): """ Get the current base URL. Args: webdriver: Selenium WebDriver instance. Returns: str - base URL. Usage:: driver.get("http://www.google.com/?q=hello+world") WebUtils.get_base_url(driver) ...
[ "def", "get_base_url", "(", "webdriver", ")", ":", "current_url", "=", "webdriver", ".", "current_url", "try", ":", "return", "re", ".", "findall", "(", "\"^[^/]+//[^/$]+\"", ",", "current_url", ")", "[", "0", "]", "except", ":", "raise", "RuntimeError", "("...
Get the current base URL. Args: webdriver: Selenium WebDriver instance. Returns: str - base URL. Usage:: driver.get("http://www.google.com/?q=hello+world") WebUtils.get_base_url(driver) #returns 'http://www.google.com'
[ "Get", "the", "current", "base", "URL", "." ]
python
train
aws/aws-iot-device-sdk-python
AWSIoTPythonSDK/MQTTLib.py
https://github.com/aws/aws-iot-device-sdk-python/blob/f0aa2ce34b21dd2e44f4fb7e1d058656aaf2fc62/AWSIoTPythonSDK/MQTTLib.py#L671-L707
def subscribeAsync(self, topic, QoS, ackCallback=None, messageCallback=None): """ **Description** Subscribe to the desired topic and register a message callback with SUBACK callback. **Syntax** .. code:: python # Subscribe to "myTopic" with QoS0, custom SUBACK callb...
[ "def", "subscribeAsync", "(", "self", ",", "topic", ",", "QoS", ",", "ackCallback", "=", "None", ",", "messageCallback", "=", "None", ")", ":", "return", "self", ".", "_mqtt_core", ".", "subscribe_async", "(", "topic", ",", "QoS", ",", "ackCallback", ",", ...
**Description** Subscribe to the desired topic and register a message callback with SUBACK callback. **Syntax** .. code:: python # Subscribe to "myTopic" with QoS0, custom SUBACK callback and a message callback myAWSIoTMQTTClient.subscribe("myTopic", 0, ackCallback=mySuba...
[ "**", "Description", "**" ]
python
train
DataONEorg/d1_python
lib_common/src/d1_common/checksum.py
https://github.com/DataONEorg/d1_python/blob/3ac4d4f3ca052d3e8641a6a329cab526c8ddcb0d/lib_common/src/d1_common/checksum.py#L289-L301
def format_checksum(checksum_pyxb): """Create string representation of a PyXB Checksum object. Args: PyXB Checksum object Returns: str : Combined hexadecimal value and algorithm name. """ return '{}/{}'.format( checksum_pyxb.algorithm.upper().replace('-', ''), checksum_pyxb.va...
[ "def", "format_checksum", "(", "checksum_pyxb", ")", ":", "return", "'{}/{}'", ".", "format", "(", "checksum_pyxb", ".", "algorithm", ".", "upper", "(", ")", ".", "replace", "(", "'-'", ",", "''", ")", ",", "checksum_pyxb", ".", "value", "(", ")", ".", ...
Create string representation of a PyXB Checksum object. Args: PyXB Checksum object Returns: str : Combined hexadecimal value and algorithm name.
[ "Create", "string", "representation", "of", "a", "PyXB", "Checksum", "object", "." ]
python
train
opendatateam/udata
udata/sentry.py
https://github.com/opendatateam/udata/blob/f016585af94b0ff6bd73738c700324adc8ba7f8f/udata/sentry.py#L24-L31
def public_dsn(dsn): '''Transform a standard Sentry DSN into a public one''' m = RE_DSN.match(dsn) if not m: log.error('Unable to parse Sentry DSN') public = '{scheme}://{client_id}@{domain}/{site_id}'.format( **m.groupdict()) return public
[ "def", "public_dsn", "(", "dsn", ")", ":", "m", "=", "RE_DSN", ".", "match", "(", "dsn", ")", "if", "not", "m", ":", "log", ".", "error", "(", "'Unable to parse Sentry DSN'", ")", "public", "=", "'{scheme}://{client_id}@{domain}/{site_id}'", ".", "format", "...
Transform a standard Sentry DSN into a public one
[ "Transform", "a", "standard", "Sentry", "DSN", "into", "a", "public", "one" ]
python
train
mcocdawc/chemcoord
version.py
https://github.com/mcocdawc/chemcoord/blob/95561ce387c142227c38fb14a1d182179aef8f5f/version.py#L136-L160
def get_version(pep440=False): """Tracks the version number. pep440: bool When True, this function returns a version string suitable for a release as defined by PEP 440. When False, the githash (if available) will be appended to the version string. The file VERSION holds the versio...
[ "def", "get_version", "(", "pep440", "=", "False", ")", ":", "git_version", "=", "format_git_describe", "(", "call_git_describe", "(", ")", ",", "pep440", "=", "pep440", ")", "if", "git_version", "is", "None", ":", "# not a git repository", "return", "read_relea...
Tracks the version number. pep440: bool When True, this function returns a version string suitable for a release as defined by PEP 440. When False, the githash (if available) will be appended to the version string. The file VERSION holds the version information. If this is not a git ...
[ "Tracks", "the", "version", "number", "." ]
python
train
librosa/librosa
librosa/core/audio.py
https://github.com/librosa/librosa/blob/180e8e6eb8f958fa6b20b8cba389f7945d508247/librosa/core/audio.py#L1020-L1118
def chirp(fmin, fmax, sr=22050, length=None, duration=None, linear=False, phi=None): """Returns a chirp signal that goes from frequency `fmin` to frequency `fmax` Parameters ---------- fmin : float > 0 initial frequency fmax : float > 0 final frequency sr : number > 0 ...
[ "def", "chirp", "(", "fmin", ",", "fmax", ",", "sr", "=", "22050", ",", "length", "=", "None", ",", "duration", "=", "None", ",", "linear", "=", "False", ",", "phi", "=", "None", ")", ":", "if", "fmin", "is", "None", "or", "fmax", "is", "None", ...
Returns a chirp signal that goes from frequency `fmin` to frequency `fmax` Parameters ---------- fmin : float > 0 initial frequency fmax : float > 0 final frequency sr : number > 0 desired sampling rate of the output signal length : int > 0 desired number of s...
[ "Returns", "a", "chirp", "signal", "that", "goes", "from", "frequency", "fmin", "to", "frequency", "fmax" ]
python
test
google/budou
budou/chunk.py
https://github.com/google/budou/blob/101224e6523186851f38ee57a6b2e7bdbd826de2/budou/chunk.py#L210-L220
def swap(self, old_chunks, new_chunk): """Swaps old consecutive chunks with new chunk. Args: old_chunks (:obj:`budou.chunk.ChunkList`): List of consecutive Chunks to be removed. new_chunk (:obj:`budou.chunk.Chunk`): A Chunk to be inserted. """ indexes = [self.index(chunk) for chun...
[ "def", "swap", "(", "self", ",", "old_chunks", ",", "new_chunk", ")", ":", "indexes", "=", "[", "self", ".", "index", "(", "chunk", ")", "for", "chunk", "in", "old_chunks", "]", "del", "self", "[", "indexes", "[", "0", "]", ":", "indexes", "[", "-"...
Swaps old consecutive chunks with new chunk. Args: old_chunks (:obj:`budou.chunk.ChunkList`): List of consecutive Chunks to be removed. new_chunk (:obj:`budou.chunk.Chunk`): A Chunk to be inserted.
[ "Swaps", "old", "consecutive", "chunks", "with", "new", "chunk", "." ]
python
train
BernardFW/bernard
src/bernard/platforms/facebook/layers.py
https://github.com/BernardFW/bernard/blob/9c55703e5ffe5717c9fa39793df59dbfa5b4c5ab/src/bernard/platforms/facebook/layers.py#L289-L297
def is_sharable(self): """ Can only be sharable if marked as such and no child element is blocking sharing due to security reasons. """ return bool( self.sharable and all(x.is_sharable() for x in self.elements) )
[ "def", "is_sharable", "(", "self", ")", ":", "return", "bool", "(", "self", ".", "sharable", "and", "all", "(", "x", ".", "is_sharable", "(", ")", "for", "x", "in", "self", ".", "elements", ")", ")" ]
Can only be sharable if marked as such and no child element is blocking sharing due to security reasons.
[ "Can", "only", "be", "sharable", "if", "marked", "as", "such", "and", "no", "child", "element", "is", "blocking", "sharing", "due", "to", "security", "reasons", "." ]
python
train
svenkreiss/databench
databench/analysis.py
https://github.com/svenkreiss/databench/blob/99d4adad494b60a42af6b8bfba94dd0c41ba0786/databench/analysis.py#L189-L205
def emit(self, signal, message='__nomessagetoken__'): """Emit a signal to the frontend. :param str signal: name of the signal :param message: message to send :returns: return value from frontend emit function :rtype: tornado.concurrent.Future """ # call pre-emit ...
[ "def", "emit", "(", "self", ",", "signal", ",", "message", "=", "'__nomessagetoken__'", ")", ":", "# call pre-emit hooks", "if", "signal", "==", "'log'", ":", "self", ".", "log_backend", ".", "info", "(", "message", ")", "elif", "signal", "==", "'warn'", "...
Emit a signal to the frontend. :param str signal: name of the signal :param message: message to send :returns: return value from frontend emit function :rtype: tornado.concurrent.Future
[ "Emit", "a", "signal", "to", "the", "frontend", "." ]
python
train
armet/python-armet
armet/query/parser.py
https://github.com/armet/python-armet/blob/d61eca9082256cb1e7f7f3c7f2fbc4b697157de7/armet/query/parser.py#L295-L366
def parse_segment(text): "we expect foo=bar" if not len(text): return NoopQuerySegment() q = QuerySegment() # First we need to split the segment into key/value pairs. This is done # by attempting to split the sequence for each equality comparison. Then # discard any that did not spl...
[ "def", "parse_segment", "(", "text", ")", ":", "if", "not", "len", "(", "text", ")", ":", "return", "NoopQuerySegment", "(", ")", "q", "=", "QuerySegment", "(", ")", "# First we need to split the segment into key/value pairs. This is done", "# by attempting to split th...
we expect foo=bar
[ "we", "expect", "foo", "=", "bar" ]
python
valid
google/openhtf
openhtf/util/console_output.py
https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/console_output.py#L65-L68
def _printed_len(some_string): """Compute the visible length of the string when printed.""" return len([x for x in ANSI_ESC_RE.sub('', some_string) if x in string.printable])
[ "def", "_printed_len", "(", "some_string", ")", ":", "return", "len", "(", "[", "x", "for", "x", "in", "ANSI_ESC_RE", ".", "sub", "(", "''", ",", "some_string", ")", "if", "x", "in", "string", ".", "printable", "]", ")" ]
Compute the visible length of the string when printed.
[ "Compute", "the", "visible", "length", "of", "the", "string", "when", "printed", "." ]
python
train
smdabdoub/phylotoast
bin/biom_phyla_summary.py
https://github.com/smdabdoub/phylotoast/blob/0b74ef171e6a84761710548501dfac71285a58a3/bin/biom_phyla_summary.py#L27-L46
def summarize_taxa(biom): """ Given an abundance table, group the counts by every taxonomic level. """ tamtcounts = defaultdict(int) tot_seqs = 0.0 for row, col, amt in biom['data']: tot_seqs += amt rtax = biom['rows'][row]['metadata']['taxonomy'] for i, t in enumera...
[ "def", "summarize_taxa", "(", "biom", ")", ":", "tamtcounts", "=", "defaultdict", "(", "int", ")", "tot_seqs", "=", "0.0", "for", "row", ",", "col", ",", "amt", "in", "biom", "[", "'data'", "]", ":", "tot_seqs", "+=", "amt", "rtax", "=", "biom", "[",...
Given an abundance table, group the counts by every taxonomic level.
[ "Given", "an", "abundance", "table", "group", "the", "counts", "by", "every", "taxonomic", "level", "." ]
python
train
django-danceschool/django-danceschool
danceschool/payments/square/cms_plugins.py
https://github.com/django-danceschool/django-danceschool/blob/bb08cbf39017a812a5a94bdb4ea34170bf1a30ba/danceschool/payments/square/cms_plugins.py#L18-L26
def render(self, context, instance, placeholder): ''' Add the cart-specific context to this form ''' context = super(SquareCheckoutFormPlugin, self).render(context, instance, placeholder) context.update({ 'squareApplicationId': getattr(settings,'SQUARE_APPLICATION_ID',''), ...
[ "def", "render", "(", "self", ",", "context", ",", "instance", ",", "placeholder", ")", ":", "context", "=", "super", "(", "SquareCheckoutFormPlugin", ",", "self", ")", ".", "render", "(", "context", ",", "instance", ",", "placeholder", ")", "context", "."...
Add the cart-specific context to this form
[ "Add", "the", "cart", "-", "specific", "context", "to", "this", "form" ]
python
train
kisom/pypcapfile
pcapfile/protocols/linklayer/wifi.py
https://github.com/kisom/pypcapfile/blob/67520cfbb6c2e9ab3e7c181a8012ddc56ec5cad8/pcapfile/protocols/linklayer/wifi.py#L958-L972
def strip_rx_flags(self, idx): """strip(2 byte) radiotap.rxflags :idx: int :return: int idx :return: collections.namedtuple """ rx_flags = collections.namedtuple('rx_flags', ['reserved', 'badplcp']) idx = Radiotap.align(idx, 2) flags, = struct...
[ "def", "strip_rx_flags", "(", "self", ",", "idx", ")", ":", "rx_flags", "=", "collections", ".", "namedtuple", "(", "'rx_flags'", ",", "[", "'reserved'", ",", "'badplcp'", "]", ")", "idx", "=", "Radiotap", ".", "align", "(", "idx", ",", "2", ")", "flag...
strip(2 byte) radiotap.rxflags :idx: int :return: int idx :return: collections.namedtuple
[ "strip", "(", "2", "byte", ")", "radiotap", ".", "rxflags", ":", "idx", ":", "int", ":", "return", ":", "int", "idx", ":", "return", ":", "collections", ".", "namedtuple" ]
python
valid
astropy/photutils
photutils/psf/utils.py
https://github.com/astropy/photutils/blob/cc9bb4534ab76bac98cb5f374a348a2573d10401/photutils/psf/utils.py#L15-L106
def prepare_psf_model(psfmodel, xname=None, yname=None, fluxname=None, renormalize_psf=True): """ Convert a 2D PSF model to one suitable for use with `BasicPSFPhotometry` or its subclasses. The resulting model may be a composite model, but should have only the x, y, and flux r...
[ "def", "prepare_psf_model", "(", "psfmodel", ",", "xname", "=", "None", ",", "yname", "=", "None", ",", "fluxname", "=", "None", ",", "renormalize_psf", "=", "True", ")", ":", "if", "xname", "is", "None", ":", "xinmod", "=", "models", ".", "Shift", "("...
Convert a 2D PSF model to one suitable for use with `BasicPSFPhotometry` or its subclasses. The resulting model may be a composite model, but should have only the x, y, and flux related parameters un-fixed. Parameters ---------- psfmodel : a 2D model The model to assume as representati...
[ "Convert", "a", "2D", "PSF", "model", "to", "one", "suitable", "for", "use", "with", "BasicPSFPhotometry", "or", "its", "subclasses", "." ]
python
train
vertical-knowledge/ripozo-sqlalchemy
ripozo_sqlalchemy/alchemymanager.py
https://github.com/vertical-knowledge/ripozo-sqlalchemy/blob/4bcc57ec6db1b39b84b50553bb264e4950ce4ec2/ripozo_sqlalchemy/alchemymanager.py#L172-L214
def retrieve_list(self, session, filters, *args, **kwargs): """ Retrieves a list of the model for this manager. It is restricted by the filters provided. :param Session session: The SQLAlchemy session to use :param dict filters: The filters to restrict the returned m...
[ "def", "retrieve_list", "(", "self", ",", "session", ",", "filters", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "query", "=", "self", ".", "queryset", "(", "session", ")", "translator", "=", "IntegerField", "(", "'tmp'", ")", "pagination_count"...
Retrieves a list of the model for this manager. It is restricted by the filters provided. :param Session session: The SQLAlchemy session to use :param dict filters: The filters to restrict the returned models on :return: A tuple of the list of dictionary representation ...
[ "Retrieves", "a", "list", "of", "the", "model", "for", "this", "manager", ".", "It", "is", "restricted", "by", "the", "filters", "provided", "." ]
python
train
sunt05/SuPy
src/supy/supy_util.py
https://github.com/sunt05/SuPy/blob/47178bd5aee50a059414e3e504940662fbfae0dc/src/supy/supy_util.py#L756-L808
def gen_req_sfc(lat_x, lon_x, start, end, grid=[0.125, 0.125], scale=0): '''generate a dict of reqs kwargs for (lat_x,lon_x) spanning [start, end] Parameters ---------- lat_x : [type] [description] lon_x : [type] [description] start : [type] [description] end : [type...
[ "def", "gen_req_sfc", "(", "lat_x", ",", "lon_x", ",", "start", ",", "end", ",", "grid", "=", "[", "0.125", ",", "0.125", "]", ",", "scale", "=", "0", ")", ":", "# scale is a factor to rescale grid size", "size", "=", "grid", "[", "0", "]", "*", "scale...
generate a dict of reqs kwargs for (lat_x,lon_x) spanning [start, end] Parameters ---------- lat_x : [type] [description] lon_x : [type] [description] start : [type] [description] end : [type] [description] grid : list, optional [description] (the def...
[ "generate", "a", "dict", "of", "reqs", "kwargs", "for", "(", "lat_x", "lon_x", ")", "spanning", "[", "start", "end", "]" ]
python
train
brocade/pynos
pynos/versions/ver_7/ver_7_1_0/yang/brocade_maps_ext.py
https://github.com/brocade/pynos/blob/bd8a34e98f322de3fc06750827d8bbc3a0c00380/pynos/versions/ver_7/ver_7_1_0/yang/brocade_maps_ext.py#L93-L105
def maps_get_default_rules_output_rules_groupname(self, **kwargs): """Auto Generated Code """ config = ET.Element("config") maps_get_default_rules = ET.Element("maps_get_default_rules") config = maps_get_default_rules output = ET.SubElement(maps_get_default_rules, "output...
[ "def", "maps_get_default_rules_output_rules_groupname", "(", "self", ",", "*", "*", "kwargs", ")", ":", "config", "=", "ET", ".", "Element", "(", "\"config\"", ")", "maps_get_default_rules", "=", "ET", ".", "Element", "(", "\"maps_get_default_rules\"", ")", "confi...
Auto Generated Code
[ "Auto", "Generated", "Code" ]
python
train
olitheolix/qtmacs
qtmacs/extensions/qtmacsscintilla_macros.py
https://github.com/olitheolix/qtmacs/blob/36253b082b82590f183fe154b053eb3a1e741be2/qtmacs/extensions/qtmacsscintilla_macros.py#L1588-L1594
def qteToBeKilled(self): """ Remove all selections and install the original lexer. """ self.clearHighlighting() self.qteWidget.qteSetLexer(self.originalLexer) self.qteText.textChanged.disconnect(self.qteTextChanged)
[ "def", "qteToBeKilled", "(", "self", ")", ":", "self", ".", "clearHighlighting", "(", ")", "self", ".", "qteWidget", ".", "qteSetLexer", "(", "self", ".", "originalLexer", ")", "self", ".", "qteText", ".", "textChanged", ".", "disconnect", "(", "self", "."...
Remove all selections and install the original lexer.
[ "Remove", "all", "selections", "and", "install", "the", "original", "lexer", "." ]
python
train
astropy/astropy-helpers
astropy_helpers/utils.py
https://github.com/astropy/astropy-helpers/blob/f5a27d3f84a98ea0eebb85e0cf3e7214c6bc0d09/astropy_helpers/utils.py#L37-L44
def _get_platlib_dir(cmd): """ Given a build command, return the name of the appropriate platform-specific build subdirectory directory (e.g. build/lib.linux-x86_64-2.7) """ plat_specifier = '.{0}-{1}'.format(cmd.plat_name, sys.version[0:3]) return os.path.join(cmd.build_base, 'lib' + plat_spec...
[ "def", "_get_platlib_dir", "(", "cmd", ")", ":", "plat_specifier", "=", "'.{0}-{1}'", ".", "format", "(", "cmd", ".", "plat_name", ",", "sys", ".", "version", "[", "0", ":", "3", "]", ")", "return", "os", ".", "path", ".", "join", "(", "cmd", ".", ...
Given a build command, return the name of the appropriate platform-specific build subdirectory directory (e.g. build/lib.linux-x86_64-2.7)
[ "Given", "a", "build", "command", "return", "the", "name", "of", "the", "appropriate", "platform", "-", "specific", "build", "subdirectory", "directory", "(", "e", ".", "g", ".", "build", "/", "lib", ".", "linux", "-", "x86_64", "-", "2", ".", "7", ")"...
python
train
clusterpoint/python-client-api
pycps/converters.py
https://github.com/clusterpoint/python-client-api/blob/fabf9bd8355aa54ba08fd6649e48f16e2c35eacd/pycps/converters.py#L85-L138
def dict_to_etree(source, root_tag=None): """ Recursively load dict/list representation of an XML tree into an etree representation. Args: source -- A dictionary representing an XML document where identical children tags are countained in a list. Keyword args: ...
[ "def", "dict_to_etree", "(", "source", ",", "root_tag", "=", "None", ")", ":", "def", "dict_to_etree_recursive", "(", "source", ",", "parent", ")", ":", "if", "hasattr", "(", "source", ",", "'keys'", ")", ":", "for", "key", ",", "value", "in", "source", ...
Recursively load dict/list representation of an XML tree into an etree representation. Args: source -- A dictionary representing an XML document where identical children tags are countained in a list. Keyword args: root_tag -- A parent tag in which to wrap t...
[ "Recursively", "load", "dict", "/", "list", "representation", "of", "an", "XML", "tree", "into", "an", "etree", "representation", "." ]
python
train
tadashi-aikawa/owlmixin
owlmixin/transformers.py
https://github.com/tadashi-aikawa/owlmixin/blob/7c4a042c3008abddc56a8e8e55ae930d276071f5/owlmixin/transformers.py#L230-L266
def to_pretty_json(self, ignore_none: bool=True, ignore_empty: bool=False) -> str: """From instance to pretty json string :param ignore_none: Properties which is None are excluded if True :param ignore_empty: Properties which is empty are excluded if True :return: Json string U...
[ "def", "to_pretty_json", "(", "self", ",", "ignore_none", ":", "bool", "=", "True", ",", "ignore_empty", ":", "bool", "=", "False", ")", "->", "str", ":", "return", "self", ".", "to_json", "(", "4", ",", "ignore_none", ",", "ignore_empty", ")" ]
From instance to pretty json string :param ignore_none: Properties which is None are excluded if True :param ignore_empty: Properties which is empty are excluded if True :return: Json string Usage: >>> from owlmixin.samples import Human >>> human = Human.from_d...
[ "From", "instance", "to", "pretty", "json", "string" ]
python
train
JukeboxPipeline/jukebox-core
src/jukeboxcore/reftrack.py
https://github.com/JukeboxPipeline/jukebox-core/blob/bac2280ca49940355270e4b69400ce9976ab2e6f/src/jukeboxcore/reftrack.py#L2107-L2121
def import_taskfile(self, refobj, taskfileinfo): """Import the given taskfileinfo and update the refobj This will call :meth:`ReftypeInterface.import_taskfile`. :param refobj: the refobject :type refobj: refobject :param taskfileinfo: the taskfileinfo to reference :type...
[ "def", "import_taskfile", "(", "self", ",", "refobj", ",", "taskfileinfo", ")", ":", "inter", "=", "self", ".", "get_typ_interface", "(", "self", ".", "get_typ", "(", "refobj", ")", ")", "inter", ".", "import_taskfile", "(", "refobj", ",", "taskfileinfo", ...
Import the given taskfileinfo and update the refobj This will call :meth:`ReftypeInterface.import_taskfile`. :param refobj: the refobject :type refobj: refobject :param taskfileinfo: the taskfileinfo to reference :type taskfileinfo: :class:`jukeboxcore.filesys.TaskFileInfo` ...
[ "Import", "the", "given", "taskfileinfo", "and", "update", "the", "refobj" ]
python
train
okpy/ok-client
client/utils/auth.py
https://github.com/okpy/ok-client/blob/517f57dd76284af40ba9766e42d9222b644afd9c/client/utils/auth.py#L72-L87
def pick_free_port(hostname=REDIRECT_HOST, port=0): """ Try to bind a port. Default=0 selects a free port. """ import socket s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) try: s.bind((hostname, port)) # port=0 finds an open port except OSError as e: log.warning("Could not bi...
[ "def", "pick_free_port", "(", "hostname", "=", "REDIRECT_HOST", ",", "port", "=", "0", ")", ":", "import", "socket", "s", "=", "socket", ".", "socket", "(", "socket", ".", "AF_INET", ",", "socket", ".", "SOCK_STREAM", ")", "try", ":", "s", ".", "bind",...
Try to bind a port. Default=0 selects a free port.
[ "Try", "to", "bind", "a", "port", ".", "Default", "=", "0", "selects", "a", "free", "port", "." ]
python
train
HdrHistogram/HdrHistogram_py
hdrh/log.py
https://github.com/HdrHistogram/HdrHistogram_py/blob/cb99981b0564a62e1aa02bd764efa6445923f8f7/hdrh/log.py#L51-L92
def output_interval_histogram(self, histogram, start_time_stamp_sec=0, end_time_stamp_sec=0, max_value_unit_ratio=1000000.0): '''Output an interval histogram, with the given ti...
[ "def", "output_interval_histogram", "(", "self", ",", "histogram", ",", "start_time_stamp_sec", "=", "0", ",", "end_time_stamp_sec", "=", "0", ",", "max_value_unit_ratio", "=", "1000000.0", ")", ":", "if", "not", "start_time_stamp_sec", ":", "start_time_stamp_sec", ...
Output an interval histogram, with the given timestamp and a configurable maxValueUnitRatio. (note that the specified timestamp will be used, and the timestamp in the actual histogram will be ignored). The max value reported with the interval line will be scaled by the given max_...
[ "Output", "an", "interval", "histogram", "with", "the", "given", "timestamp", "and", "a", "configurable", "maxValueUnitRatio", ".", "(", "note", "that", "the", "specified", "timestamp", "will", "be", "used", "and", "the", "timestamp", "in", "the", "actual", "h...
python
train
hubo1016/vlcp
vlcp/event/matchtree.py
https://github.com/hubo1016/vlcp/blob/239055229ec93a99cc7e15208075724ccf543bd1/vlcp/event/matchtree.py#L67-L81
def insert(self, matcher, obj): ''' Insert a new matcher :param matcher: an EventMatcher :param obj: object to return ''' current = self.subtree(matcher, True) #current.matchers[(obj, matcher)] = None if current._use_dict: cur...
[ "def", "insert", "(", "self", ",", "matcher", ",", "obj", ")", ":", "current", "=", "self", ".", "subtree", "(", "matcher", ",", "True", ")", "#current.matchers[(obj, matcher)] = None", "if", "current", ".", "_use_dict", ":", "current", ".", "matchers_dict", ...
Insert a new matcher :param matcher: an EventMatcher :param obj: object to return
[ "Insert", "a", "new", "matcher", ":", "param", "matcher", ":", "an", "EventMatcher", ":", "param", "obj", ":", "object", "to", "return" ]
python
train
vtraag/leidenalg
setup.py
https://github.com/vtraag/leidenalg/blob/a9e15116973a81048edf02ef7cf800d54debe1cc/setup.py#L71-L75
def ensure_dir_does_not_exist(*args): """Ensures that the given directory does not exist.""" path = os.path.join(*args) if os.path.isdir(path): shutil.rmtree(path)
[ "def", "ensure_dir_does_not_exist", "(", "*", "args", ")", ":", "path", "=", "os", ".", "path", ".", "join", "(", "*", "args", ")", "if", "os", ".", "path", ".", "isdir", "(", "path", ")", ":", "shutil", ".", "rmtree", "(", "path", ")" ]
Ensures that the given directory does not exist.
[ "Ensures", "that", "the", "given", "directory", "does", "not", "exist", "." ]
python
train
MacHu-GWU/pathlib_mate-project
pathlib_mate/mate_path_filters.py
https://github.com/MacHu-GWU/pathlib_mate-project/blob/f9fb99dd7cc9ea05d1bec8b9ce8f659e8d97b0f1/pathlib_mate/mate_path_filters.py#L156-L170
def select_by_ext(self, ext, recursive=True): """ Select file path by extension. :param ext: **中文文档** 选择与预定义的若干个扩展名匹配的文件。 """ ext = [ext.strip().lower() for ext in ensure_list(ext)] def filters(p): return p.suffix.lower() in ext return self.se...
[ "def", "select_by_ext", "(", "self", ",", "ext", ",", "recursive", "=", "True", ")", ":", "ext", "=", "[", "ext", ".", "strip", "(", ")", ".", "lower", "(", ")", "for", "ext", "in", "ensure_list", "(", "ext", ")", "]", "def", "filters", "(", "p",...
Select file path by extension. :param ext: **中文文档** 选择与预定义的若干个扩展名匹配的文件。
[ "Select", "file", "path", "by", "extension", "." ]
python
valid
Adarnof/adarnauth-esi
esi/managers.py
https://github.com/Adarnof/adarnauth-esi/blob/f6618a31efbfeedeb96316ab9b82ecadda776ac1/esi/managers.py#L105-L112
def equivalent_to(self, token): """ Gets all tokens which match the character and scopes of a reference token :param token: :class:`esi.models.Token` :return: :class:`esi.managers.TokenQueryset` """ return self.filter(character_id=token.character_id).require_scopes_exact(...
[ "def", "equivalent_to", "(", "self", ",", "token", ")", ":", "return", "self", ".", "filter", "(", "character_id", "=", "token", ".", "character_id", ")", ".", "require_scopes_exact", "(", "token", ".", "scopes", ".", "all", "(", ")", ")", ".", "filter",...
Gets all tokens which match the character and scopes of a reference token :param token: :class:`esi.models.Token` :return: :class:`esi.managers.TokenQueryset`
[ "Gets", "all", "tokens", "which", "match", "the", "character", "and", "scopes", "of", "a", "reference", "token", ":", "param", "token", ":", ":", "class", ":", "esi", ".", "models", ".", "Token", ":", "return", ":", ":", "class", ":", "esi", ".", "ma...
python
train
log2timeline/dfvfs
dfvfs/analyzer/tsk_analyzer_helper.py
https://github.com/log2timeline/dfvfs/blob/2b3ccd115f9901d89f383397d4a1376a873c83c4/dfvfs/analyzer/tsk_analyzer_helper.py#L20-L57
def GetFormatSpecification(self): """Retrieves the format specification. Returns: FormatSpecification: format specification or None if the format cannot be defined by a specification object. """ format_specification = specification.FormatSpecification( self.type_indicator) ...
[ "def", "GetFormatSpecification", "(", "self", ")", ":", "format_specification", "=", "specification", ".", "FormatSpecification", "(", "self", ".", "type_indicator", ")", "# FAT volume header signature.", "format_specification", ".", "AddNewSignature", "(", "b'\\x55\\xaa'",...
Retrieves the format specification. Returns: FormatSpecification: format specification or None if the format cannot be defined by a specification object.
[ "Retrieves", "the", "format", "specification", "." ]
python
train
mitsei/dlkit
dlkit/json_/relationship/sessions.py
https://github.com/mitsei/dlkit/blob/445f968a175d61c8d92c0f617a3c17dc1dc7c584/dlkit/json_/relationship/sessions.py#L2065-L2085
def get_root_families(self): """Gets the root families in the family hierarchy. A node with no parents is an orphan. While all family ``Ids`` are known to the hierarchy, an orphan does not appear in the hierarchy unless explicitly added as a root node or child of another node. ...
[ "def", "get_root_families", "(", "self", ")", ":", "# Implemented from template for", "# osid.resource.BinHierarchySession.get_root_bins", "if", "self", ".", "_catalog_session", "is", "not", "None", ":", "return", "self", ".", "_catalog_session", ".", "get_root_catalogs", ...
Gets the root families in the family hierarchy. A node with no parents is an orphan. While all family ``Ids`` are known to the hierarchy, an orphan does not appear in the hierarchy unless explicitly added as a root node or child of another node. return: (osid.relationship.Famil...
[ "Gets", "the", "root", "families", "in", "the", "family", "hierarchy", "." ]
python
train
ArchiveTeam/wpull
wpull/network/connection.py
https://github.com/ArchiveTeam/wpull/blob/ddf051aa3322479325ba20aa778cb2cb97606bf5/wpull/network/connection.py#L248-L259
def read(self, amount: int=-1) -> bytes: '''Read data.''' assert self._state == ConnectionState.created, \ 'Expect conn created. Got {}.'.format(self._state) data = yield from \ self.run_network_operation( self.reader.read(amount), close_t...
[ "def", "read", "(", "self", ",", "amount", ":", "int", "=", "-", "1", ")", "->", "bytes", ":", "assert", "self", ".", "_state", "==", "ConnectionState", ".", "created", ",", "'Expect conn created. Got {}.'", ".", "format", "(", "self", ".", "_state", ")"...
Read data.
[ "Read", "data", "." ]
python
train
NASA-AMMOS/AIT-Core
ait/core/geom.py
https://github.com/NASA-AMMOS/AIT-Core/blob/9d85bd9c738e7a6a6fbdff672bea708238b02a3a/ait/core/geom.py#L700-L705
def segments (self): """Return the Line segments that comprise this Polygon.""" for n in xrange(len(self.vertices) - 1): yield Line(self.vertices[n], self.vertices[n + 1]) yield Line(self.vertices[-1], self.vertices[0])
[ "def", "segments", "(", "self", ")", ":", "for", "n", "in", "xrange", "(", "len", "(", "self", ".", "vertices", ")", "-", "1", ")", ":", "yield", "Line", "(", "self", ".", "vertices", "[", "n", "]", ",", "self", ".", "vertices", "[", "n", "+", ...
Return the Line segments that comprise this Polygon.
[ "Return", "the", "Line", "segments", "that", "comprise", "this", "Polygon", "." ]
python
train
RJT1990/pyflux
pyflux/garch/segarch.py
https://github.com/RJT1990/pyflux/blob/297f2afc2095acd97c12e827dd500e8ea5da0c0f/pyflux/garch/segarch.py#L155-L217
def _mb_model(self, beta, mini_batch): """ Creates the structure of the model Parameters ---------- beta : np.array Contains untransformed starting values for latent variables mini_batch : int Mini batch size for the data sampling Returns ...
[ "def", "_mb_model", "(", "self", ",", "beta", ",", "mini_batch", ")", ":", "# Transform latent variables", "parm", "=", "np", ".", "array", "(", "[", "self", ".", "latent_variables", ".", "z_list", "[", "k", "]", ".", "prior", ".", "transform", "(", "bet...
Creates the structure of the model Parameters ---------- beta : np.array Contains untransformed starting values for latent variables mini_batch : int Mini batch size for the data sampling Returns ---------- lambda : np.array ...
[ "Creates", "the", "structure", "of", "the", "model" ]
python
train
vertexproject/synapse
synapse/lib/gis.py
https://github.com/vertexproject/synapse/blob/22e67c5a8f6d7caddbcf34b39ab1bd2d6c4a6e0b/synapse/lib/gis.py#L13-L25
def latlong(text): ''' Chop a latlong string and return (float,float). Does not perform validation on the coordinates. Args: text (str): A longitude,latitude string. Returns: (float,float): A longitude, latitude float tuple. ''' nlat, nlon = text.split(',') return (flo...
[ "def", "latlong", "(", "text", ")", ":", "nlat", ",", "nlon", "=", "text", ".", "split", "(", "','", ")", "return", "(", "float", "(", "nlat", ")", ",", "float", "(", "nlon", ")", ")" ]
Chop a latlong string and return (float,float). Does not perform validation on the coordinates. Args: text (str): A longitude,latitude string. Returns: (float,float): A longitude, latitude float tuple.
[ "Chop", "a", "latlong", "string", "and", "return", "(", "float", "float", ")", ".", "Does", "not", "perform", "validation", "on", "the", "coordinates", "." ]
python
train
spyder-ide/spyder
spyder/utils/programs.py
https://github.com/spyder-ide/spyder/blob/f76836ce1b924bcc4efd3f74f2960d26a4e528e0/spyder/utils/programs.py#L251-L270
def get_python_args(fname, python_args, interact, debug, end_args): """Construct Python interpreter arguments""" p_args = [] if python_args is not None: p_args += python_args.split() if interact: p_args.append('-i') if debug: p_args.extend(['-m', 'pdb']) if fname...
[ "def", "get_python_args", "(", "fname", ",", "python_args", ",", "interact", ",", "debug", ",", "end_args", ")", ":", "p_args", "=", "[", "]", "if", "python_args", "is", "not", "None", ":", "p_args", "+=", "python_args", ".", "split", "(", ")", "if", "...
Construct Python interpreter arguments
[ "Construct", "Python", "interpreter", "arguments" ]
python
train
dossier/dossier.models
dossier/models/etl/interface.py
https://github.com/dossier/dossier.models/blob/c9e282f690eab72963926329efe1600709e48b13/dossier/models/etl/interface.py#L152-L231
def html_to_fc(html=None, clean_html=None, clean_visible=None, encoding=None, url=None, timestamp=None, other_features=None): '''`html` is expected to be a raw string received over the wire from a remote webserver, and `encoding`, if provided, is used to decode it. Typically, encoding comes ...
[ "def", "html_to_fc", "(", "html", "=", "None", ",", "clean_html", "=", "None", ",", "clean_visible", "=", "None", ",", "encoding", "=", "None", ",", "url", "=", "None", ",", "timestamp", "=", "None", ",", "other_features", "=", "None", ")", ":", "def",...
`html` is expected to be a raw string received over the wire from a remote webserver, and `encoding`, if provided, is used to decode it. Typically, encoding comes from the Content-Type header field. The :func:`~streamcorpus_pipeline._clean_html.make_clean_html` function handles character encodings.
[ "html", "is", "expected", "to", "be", "a", "raw", "string", "received", "over", "the", "wire", "from", "a", "remote", "webserver", "and", "encoding", "if", "provided", "is", "used", "to", "decode", "it", ".", "Typically", "encoding", "comes", "from", "the"...
python
train
mardix/Mocha
mocha/extras/jinja_helpers.py
https://github.com/mardix/Mocha/blob/bce481cb31a0972061dd99bc548701411dcb9de3/mocha/extras/jinja_helpers.py#L45-L72
def img_src(url, class_="", responsive=False, lazy_load=False, id_=""): """ Create an image src {{ xyz.jpg | img_src }} :param url: :param class_: :param responsive: :param lazy_load: :param id_: :return: """ if not url.startswith("http://") and not url.startswith("https://...
[ "def", "img_src", "(", "url", ",", "class_", "=", "\"\"", ",", "responsive", "=", "False", ",", "lazy_load", "=", "False", ",", "id_", "=", "\"\"", ")", ":", "if", "not", "url", ".", "startswith", "(", "\"http://\"", ")", "and", "not", "url", ".", ...
Create an image src {{ xyz.jpg | img_src }} :param url: :param class_: :param responsive: :param lazy_load: :param id_: :return:
[ "Create", "an", "image", "src" ]
python
train
bkg/greenwich
greenwich/raster.py
https://github.com/bkg/greenwich/blob/57ec644dadfe43ce0ecf2cfd32a2de71e0c8c141/greenwich/raster.py#L474-L488
def get_offset(self, envelope): """Returns a 4-tuple pixel window (x_offset, y_offset, x_size, y_size). Arguments: envelope -- coordinate extent tuple or Envelope """ if isinstance(envelope, collections.Sequence): envelope = Envelope(envelope) if not (self.en...
[ "def", "get_offset", "(", "self", ",", "envelope", ")", ":", "if", "isinstance", "(", "envelope", ",", "collections", ".", "Sequence", ")", ":", "envelope", "=", "Envelope", "(", "envelope", ")", "if", "not", "(", "self", ".", "envelope", ".", "contains"...
Returns a 4-tuple pixel window (x_offset, y_offset, x_size, y_size). Arguments: envelope -- coordinate extent tuple or Envelope
[ "Returns", "a", "4", "-", "tuple", "pixel", "window", "(", "x_offset", "y_offset", "x_size", "y_size", ")", "." ]
python
test
spyder-ide/spyder
spyder/plugins/console/widgets/shell.py
https://github.com/spyder-ide/spyder/blob/f76836ce1b924bcc4efd3f74f2960d26a4e528e0/spyder/plugins/console/widgets/shell.py#L858-L875
def show_completion_list(self, completions, completion_text=""): """Display the possible completions""" if not completions: return if not isinstance(completions[0], tuple): completions = [(c, '') for c in completions] if len(completions) == 1 and completions...
[ "def", "show_completion_list", "(", "self", ",", "completions", ",", "completion_text", "=", "\"\"", ")", ":", "if", "not", "completions", ":", "return", "if", "not", "isinstance", "(", "completions", "[", "0", "]", ",", "tuple", ")", ":", "completions", "...
Display the possible completions
[ "Display", "the", "possible", "completions" ]
python
train
assamite/creamas
creamas/core/simulation.py
https://github.com/assamite/creamas/blob/54dc3e31c97a3f938e58272f8ab80b6bcafeff58/creamas/core/simulation.py#L254-L270
def next(self): """Trigger next agent to :py:meth:`~creamas.core.CreativeAgent.act` in the current step. """ # all agents acted, init next step t = time.time() if len(self._agents_to_act) == 0: self._init_step() addr = self._agents_to_act.pop(0) ...
[ "def", "next", "(", "self", ")", ":", "# all agents acted, init next step", "t", "=", "time", ".", "time", "(", ")", "if", "len", "(", "self", ".", "_agents_to_act", ")", "==", "0", ":", "self", ".", "_init_step", "(", ")", "addr", "=", "self", ".", ...
Trigger next agent to :py:meth:`~creamas.core.CreativeAgent.act` in the current step.
[ "Trigger", "next", "agent", "to", ":", "py", ":", "meth", ":", "~creamas", ".", "core", ".", "CreativeAgent", ".", "act", "in", "the", "current", "step", "." ]
python
train
nephila/python-taiga
taiga/models/models.py
https://github.com/nephila/python-taiga/blob/5b471d6b8b59e5d410162a6f1c2f0d4188445a56/taiga/models/models.py#L617-L627
def attach(self, attached_file, **attrs): """ Attach a file to the :class:`Task` :param attached_file: file path to attach :param attrs: optional attributes for the attached file """ return TaskAttachments(self.requester).create( self.project, self.id, ...
[ "def", "attach", "(", "self", ",", "attached_file", ",", "*", "*", "attrs", ")", ":", "return", "TaskAttachments", "(", "self", ".", "requester", ")", ".", "create", "(", "self", ".", "project", ",", "self", ".", "id", ",", "attached_file", ",", "*", ...
Attach a file to the :class:`Task` :param attached_file: file path to attach :param attrs: optional attributes for the attached file
[ "Attach", "a", "file", "to", "the", ":", "class", ":", "Task" ]
python
train
onnx/onnx
onnx/helper.py
https://github.com/onnx/onnx/blob/2f7dc10f03a072526d94b6820cedbf2a1ec5a2c4/onnx/helper.py#L538-L551
def strip_doc_string(proto): # type: (google.protobuf.message.Message) -> None """ Empties `doc_string` field on any nested protobuf messages """ assert isinstance(proto, google.protobuf.message.Message) for descriptor in proto.DESCRIPTOR.fields: if descriptor.name == 'doc_string': ...
[ "def", "strip_doc_string", "(", "proto", ")", ":", "# type: (google.protobuf.message.Message) -> None", "assert", "isinstance", "(", "proto", ",", "google", ".", "protobuf", ".", "message", ".", "Message", ")", "for", "descriptor", "in", "proto", ".", "DESCRIPTOR", ...
Empties `doc_string` field on any nested protobuf messages
[ "Empties", "doc_string", "field", "on", "any", "nested", "protobuf", "messages" ]
python
train
aestrivex/bctpy
bct/utils/visualization.py
https://github.com/aestrivex/bctpy/blob/4cb0e759eb4a038750b07e23bd29958c400684b8/bct/utils/visualization.py#L476-L599
def reorder_matrix(m1, cost='line', verbose=False, H=1e4, Texp=10, T0=1e-3, Hbrk=10): ''' This function rearranges the nodes in matrix M1 such that the matrix elements are squeezed along the main diagonal. The function uses a version of simulated annealing. Parameters ---------- M1 : NxN n...
[ "def", "reorder_matrix", "(", "m1", ",", "cost", "=", "'line'", ",", "verbose", "=", "False", ",", "H", "=", "1e4", ",", "Texp", "=", "10", ",", "T0", "=", "1e-3", ",", "Hbrk", "=", "10", ")", ":", "from", "scipy", "import", "linalg", ",", "stats...
This function rearranges the nodes in matrix M1 such that the matrix elements are squeezed along the main diagonal. The function uses a version of simulated annealing. Parameters ---------- M1 : NxN np.ndarray connection matrix weighted/binary directed/undirected cost : str 'li...
[ "This", "function", "rearranges", "the", "nodes", "in", "matrix", "M1", "such", "that", "the", "matrix", "elements", "are", "squeezed", "along", "the", "main", "diagonal", ".", "The", "function", "uses", "a", "version", "of", "simulated", "annealing", "." ]
python
train
chrisspen/burlap
burlap/mysql.py
https://github.com/chrisspen/burlap/blob/a92b0a8e5206850bb777c74af8421ea8b33779bd/burlap/mysql.py#L582-L591
def database_exists(name, **kwargs): """ Check if a MySQL database exists. """ with settings(hide('running', 'stdout', 'stderr', 'warnings'), warn_only=True): res = query("SHOW DATABASES LIKE '%(name)s';" % { 'name': name }, **kwargs) return res.succeeded and (res == nam...
[ "def", "database_exists", "(", "name", ",", "*", "*", "kwargs", ")", ":", "with", "settings", "(", "hide", "(", "'running'", ",", "'stdout'", ",", "'stderr'", ",", "'warnings'", ")", ",", "warn_only", "=", "True", ")", ":", "res", "=", "query", "(", ...
Check if a MySQL database exists.
[ "Check", "if", "a", "MySQL", "database", "exists", "." ]
python
valid
pypa/pipenv
pipenv/patched/notpip/_internal/vcs/git.py
https://github.com/pypa/pipenv/blob/cae8d76c210b9777e90aab76e9c4b0e53bb19cde/pipenv/patched/notpip/_internal/vcs/git.py#L82-L100
def get_current_branch(self, location): """ Return the current branch, or None if HEAD isn't at a branch (e.g. detached HEAD). """ # git-symbolic-ref exits with empty stdout if "HEAD" is a detached # HEAD rather than a symbolic ref. In addition, the -q causes the ...
[ "def", "get_current_branch", "(", "self", ",", "location", ")", ":", "# git-symbolic-ref exits with empty stdout if \"HEAD\" is a detached", "# HEAD rather than a symbolic ref. In addition, the -q causes the", "# command to exit with status code 1 instead of 128 in this case", "# and to suppr...
Return the current branch, or None if HEAD isn't at a branch (e.g. detached HEAD).
[ "Return", "the", "current", "branch", "or", "None", "if", "HEAD", "isn", "t", "at", "a", "branch", "(", "e", ".", "g", ".", "detached", "HEAD", ")", "." ]
python
train
log2timeline/plaso
plaso/cli/image_export_tool.py
https://github.com/log2timeline/plaso/blob/9c564698d2da3ffbe23607a3c54c0582ea18a6cc/plaso/cli/image_export_tool.py#L415-L427
def _ParseNamesString(self, names_string): """Parses the name string. Args: names_string (str): comma separated filenames to filter. """ if not names_string: return names_string = names_string.lower() names = [name.strip() for name in names_string.split(',')] file_entry_filter ...
[ "def", "_ParseNamesString", "(", "self", ",", "names_string", ")", ":", "if", "not", "names_string", ":", "return", "names_string", "=", "names_string", ".", "lower", "(", ")", "names", "=", "[", "name", ".", "strip", "(", ")", "for", "name", "in", "name...
Parses the name string. Args: names_string (str): comma separated filenames to filter.
[ "Parses", "the", "name", "string", "." ]
python
train
chaoss/grimoirelab-sigils
src/migration/utils.py
https://github.com/chaoss/grimoirelab-sigils/blob/33d395195acb316287143a535a2c6e4009bf0528/src/migration/utils.py#L40-L58
def replace(pretty, old_str, new_str): """ Replace strings giving some info on where the replacement was done """ out_str = '' line_number = 1 changes = 0 for line in pretty.splitlines(keepends=True): new_line = line.replace(old_str, new_str) if line.find(old_str) != -1: ...
[ "def", "replace", "(", "pretty", ",", "old_str", ",", "new_str", ")", ":", "out_str", "=", "''", "line_number", "=", "1", "changes", "=", "0", "for", "line", "in", "pretty", ".", "splitlines", "(", "keepends", "=", "True", ")", ":", "new_line", "=", ...
Replace strings giving some info on where the replacement was done
[ "Replace", "strings", "giving", "some", "info", "on", "where", "the", "replacement", "was", "done" ]
python
train
google/grr
grr/core/grr_response_core/lib/parsers/linux_file_parser.py
https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/core/grr_response_core/lib/parsers/linux_file_parser.py#L663-L669
def AddPassword(self, fileset): """Add the passwd entries to the shadow store.""" passwd = fileset.get("/etc/passwd") if passwd: self._ParseFile(passwd, self.ParsePasswdEntry) else: logging.debug("No /etc/passwd file.")
[ "def", "AddPassword", "(", "self", ",", "fileset", ")", ":", "passwd", "=", "fileset", ".", "get", "(", "\"/etc/passwd\"", ")", "if", "passwd", ":", "self", ".", "_ParseFile", "(", "passwd", ",", "self", ".", "ParsePasswdEntry", ")", "else", ":", "loggin...
Add the passwd entries to the shadow store.
[ "Add", "the", "passwd", "entries", "to", "the", "shadow", "store", "." ]
python
train
openstack/networking-cisco
networking_cisco/neutronclient/hostingdevicescheduler.py
https://github.com/openstack/networking-cisco/blob/aa58a30aec25b86f9aa5952b0863045975debfa9/networking_cisco/neutronclient/hostingdevicescheduler.py#L70-L75
def associate_hosting_device_with_config_agent( self, client, config_agent_id, body): """Associates a hosting_device with a config agent.""" return client.post((ConfigAgentHandlingHostingDevice.resource_path + CFG_AGENT_HOSTING_DEVICES) % config_agent_id, ...
[ "def", "associate_hosting_device_with_config_agent", "(", "self", ",", "client", ",", "config_agent_id", ",", "body", ")", ":", "return", "client", ".", "post", "(", "(", "ConfigAgentHandlingHostingDevice", ".", "resource_path", "+", "CFG_AGENT_HOSTING_DEVICES", ")", ...
Associates a hosting_device with a config agent.
[ "Associates", "a", "hosting_device", "with", "a", "config", "agent", "." ]
python
train
datakortet/dkfileutils
tasks.py
https://github.com/datakortet/dkfileutils/blob/924098d6e2edf88ad9b3ffdec9c74530f80a7d77/tasks.py#L92-L123
def build(ctx, less=False, docs=False, js=False, force=False): """Build everything and collectstatic. """ specified = any([less, docs, js]) buildall = not specified if buildall or less: less_fname = ctx.pkg.source_less / ctx.pkg.name + '.less' if less_fname.exists(): les...
[ "def", "build", "(", "ctx", ",", "less", "=", "False", ",", "docs", "=", "False", ",", "js", "=", "False", ",", "force", "=", "False", ")", ":", "specified", "=", "any", "(", "[", "less", ",", "docs", ",", "js", "]", ")", "buildall", "=", "not"...
Build everything and collectstatic.
[ "Build", "everything", "and", "collectstatic", "." ]
python
train
PSPC-SPAC-buyandsell/von_anchor
von_anchor/wallet/wallet.py
https://github.com/PSPC-SPAC-buyandsell/von_anchor/blob/78ac1de67be42a676274f4bf71fe12f66e72f309/von_anchor/wallet/wallet.py#L1136-L1159
async def sign(self, message: bytes, verkey: str = None) -> bytes: """ Derive signing key and Sign message; return signature. Raise WalletState if wallet is closed. Raise AbsentMessage for missing message, or WalletState if wallet is closed. :param message: Content to sign, as bytes ...
[ "async", "def", "sign", "(", "self", ",", "message", ":", "bytes", ",", "verkey", ":", "str", "=", "None", ")", "->", "bytes", ":", "LOGGER", ".", "debug", "(", "'Wallet.sign >>> message: %s, verkey: %s'", ",", "message", ",", "verkey", ")", "if", "not", ...
Derive signing key and Sign message; return signature. Raise WalletState if wallet is closed. Raise AbsentMessage for missing message, or WalletState if wallet is closed. :param message: Content to sign, as bytes :param verkey: verification key corresponding to private signing key (default anch...
[ "Derive", "signing", "key", "and", "Sign", "message", ";", "return", "signature", ".", "Raise", "WalletState", "if", "wallet", "is", "closed", ".", "Raise", "AbsentMessage", "for", "missing", "message", "or", "WalletState", "if", "wallet", "is", "closed", "." ...
python
train
dhocker/udmx-pyusb
pyudmx/pyudmx.py
https://github.com/dhocker/udmx-pyusb/blob/ee7d10604ecd83857154ed6739793de3b7bd5fc1/pyudmx/pyudmx.py#L43-L67
def open(self, vendor_id: int = 0x16c0, product_id: int = 0x5dc, bus: int = None, address: int = None) -> bool: """ Open the first device that matches the search criteria. Th default parameters are set up for the likely most common case of a single uDMX interface. However, for the case o...
[ "def", "open", "(", "self", ",", "vendor_id", ":", "int", "=", "0x16c0", ",", "product_id", ":", "int", "=", "0x5dc", ",", "bus", ":", "int", "=", "None", ",", "address", ":", "int", "=", "None", ")", "->", "bool", ":", "kwargs", "=", "{", "}", ...
Open the first device that matches the search criteria. Th default parameters are set up for the likely most common case of a single uDMX interface. However, for the case of multiple uDMX interfaces, you can use the bus and address paramters to further specifiy the uDMX interface to be o...
[ "Open", "the", "first", "device", "that", "matches", "the", "search", "criteria", ".", "Th", "default", "parameters", "are", "set", "up", "for", "the", "likely", "most", "common", "case", "of", "a", "single", "uDMX", "interface", ".", "However", "for", "th...
python
train
dshean/pygeotools
pygeotools/lib/geolib.py
https://github.com/dshean/pygeotools/blob/5ac745717c0098d01eb293ff1fe32fd7358c76ab/pygeotools/lib/geolib.py#L2198-L2252
def get_dem_mosaic_cmd(fn_list, o, fn_list_txt=None, tr=None, t_srs=None, t_projwin=None, georef_tile_size=None, threads=None, tile=None, stat=None): """ Create ASP dem_mosaic command Useful for spawning many single-threaded mosaicing processes """ cmd = ['dem_mosaic',] if o is None: o ...
[ "def", "get_dem_mosaic_cmd", "(", "fn_list", ",", "o", ",", "fn_list_txt", "=", "None", ",", "tr", "=", "None", ",", "t_srs", "=", "None", ",", "t_projwin", "=", "None", ",", "georef_tile_size", "=", "None", ",", "threads", "=", "None", ",", "tile", "=...
Create ASP dem_mosaic command Useful for spawning many single-threaded mosaicing processes
[ "Create", "ASP", "dem_mosaic", "command", "Useful", "for", "spawning", "many", "single", "-", "threaded", "mosaicing", "processes" ]
python
train
jobovy/galpy
galpy/orbit/FullOrbit.py
https://github.com/jobovy/galpy/blob/9c5b9fe65d58835624dffe432be282060918ee08/galpy/orbit/FullOrbit.py#L759-L836
def _fit_orbit_mlogl(new_vxvv,vxvv,vxvv_err,pot,radec,lb, customsky,lb_to_customsky,pmllpmbb_to_customsky, tmockAA, ro,vo,obs): """The log likelihood for fitting an orbit""" #Use this _parse_args routine, which does forward and backward integration ...
[ "def", "_fit_orbit_mlogl", "(", "new_vxvv", ",", "vxvv", ",", "vxvv_err", ",", "pot", ",", "radec", ",", "lb", ",", "customsky", ",", "lb_to_customsky", ",", "pmllpmbb_to_customsky", ",", "tmockAA", ",", "ro", ",", "vo", ",", "obs", ")", ":", "#Use this _p...
The log likelihood for fitting an orbit
[ "The", "log", "likelihood", "for", "fitting", "an", "orbit" ]
python
train
tamasgal/km3pipe
km3pipe/math.py
https://github.com/tamasgal/km3pipe/blob/7a9b59ac899a28775b5bdc5d391d9a5340d08040/km3pipe/math.py#L425-L442
def qrot(vector, quaternion): """Rotate a 3D vector using quaternion algebra. Implemented by Vladimir Kulikovskiy. Parameters ---------- vector: np.array quaternion: np.array Returns ------- np.array """ t = 2 * np.cross(quaternion[1:], vector) v_rot = vector + quater...
[ "def", "qrot", "(", "vector", ",", "quaternion", ")", ":", "t", "=", "2", "*", "np", ".", "cross", "(", "quaternion", "[", "1", ":", "]", ",", "vector", ")", "v_rot", "=", "vector", "+", "quaternion", "[", "0", "]", "*", "t", "+", "np", ".", ...
Rotate a 3D vector using quaternion algebra. Implemented by Vladimir Kulikovskiy. Parameters ---------- vector: np.array quaternion: np.array Returns ------- np.array
[ "Rotate", "a", "3D", "vector", "using", "quaternion", "algebra", "." ]
python
train
core/uricore
uricore/wkz_wsgi.py
https://github.com/core/uricore/blob/dc5ef4be7bd93da4c39e5c1cbd1ae4f3ad3f1f2a/uricore/wkz_wsgi.py#L172-L178
def make_limited_stream(stream, limit): """Makes a stream limited.""" if not isinstance(stream, LimitedStream): if limit is None: raise TypeError('stream not limited and no limit provided.') stream = LimitedStream(stream, limit) return stream
[ "def", "make_limited_stream", "(", "stream", ",", "limit", ")", ":", "if", "not", "isinstance", "(", "stream", ",", "LimitedStream", ")", ":", "if", "limit", "is", "None", ":", "raise", "TypeError", "(", "'stream not limited and no limit provided.'", ")", "strea...
Makes a stream limited.
[ "Makes", "a", "stream", "limited", "." ]
python
train
quantopian/zipline
zipline/data/bcolz_daily_bars.py
https://github.com/quantopian/zipline/blob/77ad15e6dc4c1cbcdc133653bac8a63fc704f7fe/zipline/data/bcolz_daily_bars.py#L645-L676
def sid_day_index(self, sid, day): """ Parameters ---------- sid : int The asset identifier. day : datetime64-like Midnight of the day for which data is requested. Returns ------- int Index into the data tape for the gi...
[ "def", "sid_day_index", "(", "self", ",", "sid", ",", "day", ")", ":", "try", ":", "day_loc", "=", "self", ".", "sessions", ".", "get_loc", "(", "day", ")", "except", "Exception", ":", "raise", "NoDataOnDate", "(", "\"day={0} is outside of calendar={1}\"", "...
Parameters ---------- sid : int The asset identifier. day : datetime64-like Midnight of the day for which data is requested. Returns ------- int Index into the data tape for the given sid and day. Raises a NoDataOnDate exce...
[ "Parameters", "----------", "sid", ":", "int", "The", "asset", "identifier", ".", "day", ":", "datetime64", "-", "like", "Midnight", "of", "the", "day", "for", "which", "data", "is", "requested", "." ]
python
train
saltstack/salt
salt/modules/bamboohr.py
https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/bamboohr.py#L247-L288
def _query(action=None, command=None, args=None, method='GET', data=None): ''' Make a web call to BambooHR The password can be any random text, so we chose Salty text. ''' subdomain = __opts__.get('bamboohr', {}).get('subdomain', None) path = 'https:/...
[ "def", "_query", "(", "action", "=", "None", ",", "command", "=", "None", ",", "args", "=", "None", ",", "method", "=", "'GET'", ",", "data", "=", "None", ")", ":", "subdomain", "=", "__opts__", ".", "get", "(", "'bamboohr'", ",", "{", "}", ")", ...
Make a web call to BambooHR The password can be any random text, so we chose Salty text.
[ "Make", "a", "web", "call", "to", "BambooHR" ]
python
train
dw/mitogen
ansible_mitogen/target.py
https://github.com/dw/mitogen/blob/a7fdb55e1300a7e0a5e404b09eb730cf9a525da7/ansible_mitogen/target.py#L554-L574
def run_module_async(kwargs, job_id, timeout_secs, started_sender, econtext): """ Execute a module with its run status and result written to a file, terminating on the process on completion. This function must run in a child forked using :func:`create_fork_child`. @param mitogen.core.Sender started...
[ "def", "run_module_async", "(", "kwargs", ",", "job_id", ",", "timeout_secs", ",", "started_sender", ",", "econtext", ")", ":", "arunner", "=", "AsyncRunner", "(", "job_id", ",", "timeout_secs", ",", "started_sender", ",", "econtext", ",", "kwargs", ")", "arun...
Execute a module with its run status and result written to a file, terminating on the process on completion. This function must run in a child forked using :func:`create_fork_child`. @param mitogen.core.Sender started_sender: A sender that will receive :data:`True` once the job has reached a ...
[ "Execute", "a", "module", "with", "its", "run", "status", "and", "result", "written", "to", "a", "file", "terminating", "on", "the", "process", "on", "completion", ".", "This", "function", "must", "run", "in", "a", "child", "forked", "using", ":", "func", ...
python
train
saltstack/salt
salt/client/ssh/state.py
https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/client/ssh/state.py#L170-L259
def prep_trans_tar(file_client, chunks, file_refs, pillar=None, id_=None, roster_grains=None): ''' Generate the execution package from the saltenv file refs and a low state data structure ''' gendir = tempfile.mkdtemp() trans_tar = salt.utils.files.mkstemp() lowfn = os.path.join(gendir, 'low...
[ "def", "prep_trans_tar", "(", "file_client", ",", "chunks", ",", "file_refs", ",", "pillar", "=", "None", ",", "id_", "=", "None", ",", "roster_grains", "=", "None", ")", ":", "gendir", "=", "tempfile", ".", "mkdtemp", "(", ")", "trans_tar", "=", "salt",...
Generate the execution package from the saltenv file refs and a low state data structure
[ "Generate", "the", "execution", "package", "from", "the", "saltenv", "file", "refs", "and", "a", "low", "state", "data", "structure" ]
python
train
Unidata/MetPy
metpy/interpolate/grid.py
https://github.com/Unidata/MetPy/blob/16f68a94919b9a82dcf9cada2169cf039129e67b/metpy/interpolate/grid.py#L111-L133
def get_boundary_coords(x, y, spatial_pad=0): r"""Return bounding box based on given x and y coordinates assuming northern hemisphere. x: numeric x coordinates. y: numeric y coordinates. spatial_pad: numeric Number of meters to add to the x and y dimensions to reduce edg...
[ "def", "get_boundary_coords", "(", "x", ",", "y", ",", "spatial_pad", "=", "0", ")", ":", "west", "=", "np", ".", "min", "(", "x", ")", "-", "spatial_pad", "east", "=", "np", ".", "max", "(", "x", ")", "+", "spatial_pad", "north", "=", "np", ".",...
r"""Return bounding box based on given x and y coordinates assuming northern hemisphere. x: numeric x coordinates. y: numeric y coordinates. spatial_pad: numeric Number of meters to add to the x and y dimensions to reduce edge effects. Returns ------- bbox: dict...
[ "r", "Return", "bounding", "box", "based", "on", "given", "x", "and", "y", "coordinates", "assuming", "northern", "hemisphere", "." ]
python
train
joke2k/faker
faker/providers/address/__init__.py
https://github.com/joke2k/faker/blob/965824b61132e52d92d1a6ce470396dbbe01c96c/faker/providers/address/__init__.py#L52-L57
def street_name(self): """ :example 'Crist Parks' """ pattern = self.random_element(self.street_name_formats) return self.generator.parse(pattern)
[ "def", "street_name", "(", "self", ")", ":", "pattern", "=", "self", ".", "random_element", "(", "self", ".", "street_name_formats", ")", "return", "self", ".", "generator", ".", "parse", "(", "pattern", ")" ]
:example 'Crist Parks'
[ ":", "example", "Crist", "Parks" ]
python
train
PaulHancock/Aegean
AegeanTools/wcs_helpers.py
https://github.com/PaulHancock/Aegean/blob/185d2b4a51b48441a1df747efc9a5271c79399fd/AegeanTools/wcs_helpers.py#L100-L118
def from_file(cls, filename, beam=None): """ Create a new WCSHelper class from a given fits file. Parameters ---------- filename : string The file to be read beam : :class:`AegeanTools.fits_image.Beam` or None The synthesized beam. If the supplie...
[ "def", "from_file", "(", "cls", ",", "filename", ",", "beam", "=", "None", ")", ":", "header", "=", "fits", ".", "getheader", "(", "filename", ")", "return", "cls", ".", "from_header", "(", "header", ",", "beam", ")" ]
Create a new WCSHelper class from a given fits file. Parameters ---------- filename : string The file to be read beam : :class:`AegeanTools.fits_image.Beam` or None The synthesized beam. If the supplied beam is None then one is constructed form the header. ...
[ "Create", "a", "new", "WCSHelper", "class", "from", "a", "given", "fits", "file", "." ]
python
train
MycroftAI/mycroft-precise
precise/scripts/train_generated.py
https://github.com/MycroftAI/mycroft-precise/blob/e17cebdd171906dbd8a16e282d8a7966fba2eeba/precise/scripts/train_generated.py#L217-L224
def generate_samples(self): """Generate training samples (network inputs and outputs)""" filenames = glob_all(self.args.random_data_folder, '*.wav') shuffle(filenames) while True: for fn in filenames: for x, y in self.vectors_from_fn(fn): y...
[ "def", "generate_samples", "(", "self", ")", ":", "filenames", "=", "glob_all", "(", "self", ".", "args", ".", "random_data_folder", ",", "'*.wav'", ")", "shuffle", "(", "filenames", ")", "while", "True", ":", "for", "fn", "in", "filenames", ":", "for", ...
Generate training samples (network inputs and outputs)
[ "Generate", "training", "samples", "(", "network", "inputs", "and", "outputs", ")" ]
python
train
sentinel-hub/eo-learn
coregistration/eolearn/coregistration/coregistration_utilities.py
https://github.com/sentinel-hub/eo-learn/blob/b8c390b9f553c561612fe9eb64e720611633a035/coregistration/eolearn/coregistration/coregistration_utilities.py#L56-L62
def random_partition(n, n_data): """return n random rows of data (and also the other len(data)-n rows)""" all_idxs = np.arange(n_data) np.random.shuffle(all_idxs) idxs1 = all_idxs[:n] idxs2 = all_idxs[n:] return idxs1, idxs2
[ "def", "random_partition", "(", "n", ",", "n_data", ")", ":", "all_idxs", "=", "np", ".", "arange", "(", "n_data", ")", "np", ".", "random", ".", "shuffle", "(", "all_idxs", ")", "idxs1", "=", "all_idxs", "[", ":", "n", "]", "idxs2", "=", "all_idxs",...
return n random rows of data (and also the other len(data)-n rows)
[ "return", "n", "random", "rows", "of", "data", "(", "and", "also", "the", "other", "len", "(", "data", ")", "-", "n", "rows", ")" ]
python
train
calmjs/calmjs.parse
src/calmjs/parse/rules.py
https://github.com/calmjs/calmjs.parse/blob/369f0ee346c5a84c4d5c35a7733a0e63b02eac59/src/calmjs/parse/rules.py#L67-L124
def minify(drop_semi=True): """ Rules for minifying output. Arguments: drop_semi Drop semicolons whenever possible. Note that if Dedent and OptionalNewline has a handler defined, it will stop final break statements from being resolved due to reliance on normalized reso...
[ "def", "minify", "(", "drop_semi", "=", "True", ")", ":", "layout_handlers", "=", "{", "OpenBlock", ":", "layout_handler_openbrace", ",", "CloseBlock", ":", "layout_handler_closebrace", ",", "EndStatement", ":", "layout_handler_semicolon", ",", "Space", ":", "layout...
Rules for minifying output. Arguments: drop_semi Drop semicolons whenever possible. Note that if Dedent and OptionalNewline has a handler defined, it will stop final break statements from being resolved due to reliance on normalized resolution.
[ "Rules", "for", "minifying", "output", "." ]
python
train
Contraz/demosys-py
demosys/loaders/scene/gltf.py
https://github.com/Contraz/demosys-py/blob/6466128a3029c4d09631420ccce73024025bd5b6/demosys/loaders/scene/gltf.py#L452-L455
def get_bbox(self, primitive): """Get the bounding box for the mesh""" accessor = primitive.attributes.get('POSITION') return accessor.min, accessor.max
[ "def", "get_bbox", "(", "self", ",", "primitive", ")", ":", "accessor", "=", "primitive", ".", "attributes", ".", "get", "(", "'POSITION'", ")", "return", "accessor", ".", "min", ",", "accessor", ".", "max" ]
Get the bounding box for the mesh
[ "Get", "the", "bounding", "box", "for", "the", "mesh" ]
python
valid
mcieslik-mctp/papy
src/numap/NuMap.py
https://github.com/mcieslik-mctp/papy/blob/708e50827b5db46bbea081982cb74b9b0e464064/src/numap/NuMap.py#L184-L248
def _pool_put(pool_semaphore, tasks, put_to_pool_in, pool_size, id_self, \ is_stopping): """ (internal) Intended to be run in a seperate thread. Feeds tasks into to the pool whenever semaphore permits. Finishes if self._stopping is set. """ log.debug(...
[ "def", "_pool_put", "(", "pool_semaphore", ",", "tasks", ",", "put_to_pool_in", ",", "pool_size", ",", "id_self", ",", "is_stopping", ")", ":", "log", ".", "debug", "(", "'NuMap(%s) started pool_putter.'", "%", "id_self", ")", "last_tasks", "=", "{", "}", "for...
(internal) Intended to be run in a seperate thread. Feeds tasks into to the pool whenever semaphore permits. Finishes if self._stopping is set.
[ "(", "internal", ")", "Intended", "to", "be", "run", "in", "a", "seperate", "thread", ".", "Feeds", "tasks", "into", "to", "the", "pool", "whenever", "semaphore", "permits", ".", "Finishes", "if", "self", ".", "_stopping", "is", "set", "." ]
python
train
amzn/ion-python
amazon/ion/reader_text.py
https://github.com/amzn/ion-python/blob/0b21fa3ba7755f55f745e4aa970d86343b82449d/amazon/ion/reader_text.py#L369-L379
def read_data_event(self, whence, complete=False, can_flush=False): """Creates a transition to a co-routine for retrieving data as bytes. Args: whence (Coroutine): The co-routine to return to after the data is satisfied. complete (Optional[bool]): True if STREAM_END should be em...
[ "def", "read_data_event", "(", "self", ",", "whence", ",", "complete", "=", "False", ",", "can_flush", "=", "False", ")", ":", "return", "Transition", "(", "None", ",", "_read_data_handler", "(", "whence", ",", "self", ",", "complete", ",", "can_flush", ")...
Creates a transition to a co-routine for retrieving data as bytes. Args: whence (Coroutine): The co-routine to return to after the data is satisfied. complete (Optional[bool]): True if STREAM_END should be emitted if no bytes are read or available; False if INCOMPLETE sh...
[ "Creates", "a", "transition", "to", "a", "co", "-", "routine", "for", "retrieving", "data", "as", "bytes", "." ]
python
train
sentinel-hub/sentinelhub-py
sentinelhub/ogc.py
https://github.com/sentinel-hub/sentinelhub-py/blob/08a83b7f1e289187159a643336995d8369860fea/sentinelhub/ogc.py#L171-L209
def _get_common_url_parameters(request): """ Returns parameters common dictionary for WMS, WCS and FIS request. :param request: OGC-type request with specified bounding box, cloud coverage for specific product. :type request: OgcRequest or GeopediaRequest :return: dictionary with param...
[ "def", "_get_common_url_parameters", "(", "request", ")", ":", "params", "=", "{", "'SERVICE'", ":", "request", ".", "service_type", ".", "value", "}", "if", "hasattr", "(", "request", ",", "'maxcc'", ")", ":", "params", "[", "'MAXCC'", "]", "=", "100.0", ...
Returns parameters common dictionary for WMS, WCS and FIS request. :param request: OGC-type request with specified bounding box, cloud coverage for specific product. :type request: OgcRequest or GeopediaRequest :return: dictionary with parameters :rtype: dict
[ "Returns", "parameters", "common", "dictionary", "for", "WMS", "WCS", "and", "FIS", "request", "." ]
python
train
greenbone/ospd
ospd/ospd.py
https://github.com/greenbone/ospd/blob/cef773166b15a19c17764721d3fe404fa0e107bf/ospd/ospd.py#L618-L629
def handle_stop_scan_command(self, scan_et): """ Handles <stop_scan> command. @return: Response string for <stop_scan> command. """ scan_id = scan_et.attrib.get('scan_id') if scan_id is None or scan_id == '': raise OSPDError('No scan_id attribute', 'stop_scan') ...
[ "def", "handle_stop_scan_command", "(", "self", ",", "scan_et", ")", ":", "scan_id", "=", "scan_et", ".", "attrib", ".", "get", "(", "'scan_id'", ")", "if", "scan_id", "is", "None", "or", "scan_id", "==", "''", ":", "raise", "OSPDError", "(", "'No scan_id ...
Handles <stop_scan> command. @return: Response string for <stop_scan> command.
[ "Handles", "<stop_scan", ">", "command", "." ]
python
train
exosite-labs/pyonep
pyonep/onep.py
https://github.com/exosite-labs/pyonep/blob/d27b621b00688a542e0adcc01f3e3354c05238a1/pyonep/onep.py#L58-L64
def add(self, auth, method, args, notimeout=False): """Append a deferred request for a particular auth/CIK.""" authstr = self._authstr(auth) self._requests.setdefault(authstr, []).append((method, args)) self._notimeouts.setdefault(authstr, False) if notimeout: self._n...
[ "def", "add", "(", "self", ",", "auth", ",", "method", ",", "args", ",", "notimeout", "=", "False", ")", ":", "authstr", "=", "self", ".", "_authstr", "(", "auth", ")", "self", ".", "_requests", ".", "setdefault", "(", "authstr", ",", "[", "]", ")"...
Append a deferred request for a particular auth/CIK.
[ "Append", "a", "deferred", "request", "for", "a", "particular", "auth", "/", "CIK", "." ]
python
train
pybel/pybel
src/pybel/struct/mutation/induction_expansion.py
https://github.com/pybel/pybel/blob/c8a7a1bdae4c475fa2a8c77f3a9a5f6d79556ca0/src/pybel/struct/mutation/induction_expansion.py#L39-L49
def get_multi_causal_downstream(graph, nbunch: Union[BaseEntity, Iterable[BaseEntity]]): """Get the union of all of the 2-level deep causal downstream subgraphs from the nbunch. :param pybel.BELGraph graph: A BEL graph :param nbunch: A BEL node or list of BEL nodes :return: A subgraph of the original B...
[ "def", "get_multi_causal_downstream", "(", "graph", ",", "nbunch", ":", "Union", "[", "BaseEntity", ",", "Iterable", "[", "BaseEntity", "]", "]", ")", ":", "result", "=", "get_downstream_causal_subgraph", "(", "graph", ",", "nbunch", ")", "expand_downstream_causal...
Get the union of all of the 2-level deep causal downstream subgraphs from the nbunch. :param pybel.BELGraph graph: A BEL graph :param nbunch: A BEL node or list of BEL nodes :return: A subgraph of the original BEL graph :rtype: pybel.BELGraph
[ "Get", "the", "union", "of", "all", "of", "the", "2", "-", "level", "deep", "causal", "downstream", "subgraphs", "from", "the", "nbunch", "." ]
python
train
nerdvegas/rez
src/rez/vendor/amqp/serialization.py
https://github.com/nerdvegas/rez/blob/1d3b846d53b5b5404edfe8ddb9083f9ceec8c5e7/src/rez/vendor/amqp/serialization.py#L451-L479
def _load_properties(self, raw_bytes): """Given the raw bytes containing the property-flags and property-list from a content-frame-header, parse and insert into a dictionary stored in this object as an attribute named 'properties'.""" r = AMQPReader(raw_bytes) # # Read 1...
[ "def", "_load_properties", "(", "self", ",", "raw_bytes", ")", ":", "r", "=", "AMQPReader", "(", "raw_bytes", ")", "#", "# Read 16-bit shorts until we get one with a low bit set to zero", "#", "flags", "=", "[", "]", "while", "1", ":", "flag_bits", "=", "r", "."...
Given the raw bytes containing the property-flags and property-list from a content-frame-header, parse and insert into a dictionary stored in this object as an attribute named 'properties'.
[ "Given", "the", "raw", "bytes", "containing", "the", "property", "-", "flags", "and", "property", "-", "list", "from", "a", "content", "-", "frame", "-", "header", "parse", "and", "insert", "into", "a", "dictionary", "stored", "in", "this", "object", "as",...
python
train
omza/azurestoragewrap
azurestoragewrap/queue.py
https://github.com/omza/azurestoragewrap/blob/976878e95d82ff0f7d8a00a5e4a7a3fb6268ab08/azurestoragewrap/queue.py#L314-L336
def get(self, storagemodel:object, modeldefinition = None, hide = 0) -> StorageQueueModel: """ get the next message in queue """ try: if hide > 0: messages = modeldefinition['queueservice'].get_messages(storagemodel._queuename, num_messages=1, visibility_timeout = hide) ...
[ "def", "get", "(", "self", ",", "storagemodel", ":", "object", ",", "modeldefinition", "=", "None", ",", "hide", "=", "0", ")", "->", "StorageQueueModel", ":", "try", ":", "if", "hide", ">", "0", ":", "messages", "=", "modeldefinition", "[", "'queueservi...
get the next message in queue
[ "get", "the", "next", "message", "in", "queue" ]
python
train
PSU-OIT-ARC/elasticmodels
elasticmodels/analysis.py
https://github.com/PSU-OIT-ARC/elasticmodels/blob/67870508096f66123ef10b89789bbac06571cc80/elasticmodels/analysis.py#L42-L49
def diff_analysis(using): """ Returns a diff string comparing the analysis defined in ES, with the analysis defined in Python land for the connection `using` """ python_analysis = collect_analysis(using) es_analysis = existing_analysis(using) return compare_dicts(es_analysis, python_analysis...
[ "def", "diff_analysis", "(", "using", ")", ":", "python_analysis", "=", "collect_analysis", "(", "using", ")", "es_analysis", "=", "existing_analysis", "(", "using", ")", "return", "compare_dicts", "(", "es_analysis", ",", "python_analysis", ")" ]
Returns a diff string comparing the analysis defined in ES, with the analysis defined in Python land for the connection `using`
[ "Returns", "a", "diff", "string", "comparing", "the", "analysis", "defined", "in", "ES", "with", "the", "analysis", "defined", "in", "Python", "land", "for", "the", "connection", "using" ]
python
train
rameshg87/pyremotevbox
pyremotevbox/ZSI/TC.py
https://github.com/rameshg87/pyremotevbox/blob/123dffff27da57c8faa3ac1dd4c68b1cf4558b1a/pyremotevbox/ZSI/TC.py#L144-L157
def SimpleHREF(self, elt, ps, tag): '''Simple HREF for non-string and non-struct and non-array. Parameters: elt -- the DOM element being parsed ps -- the ParsedSoap object. tag -- ''' if len(_children(elt)): return elt href = _find_href(elt) ...
[ "def", "SimpleHREF", "(", "self", ",", "elt", ",", "ps", ",", "tag", ")", ":", "if", "len", "(", "_children", "(", "elt", ")", ")", ":", "return", "elt", "href", "=", "_find_href", "(", "elt", ")", "if", "not", "href", ":", "if", "self", ".", "...
Simple HREF for non-string and non-struct and non-array. Parameters: elt -- the DOM element being parsed ps -- the ParsedSoap object. tag --
[ "Simple", "HREF", "for", "non", "-", "string", "and", "non", "-", "struct", "and", "non", "-", "array", ".", "Parameters", ":", "elt", "--", "the", "DOM", "element", "being", "parsed", "ps", "--", "the", "ParsedSoap", "object", ".", "tag", "--" ]
python
train
CalebBell/thermo
thermo/chemical.py
https://github.com/CalebBell/thermo/blob/3857ed023a3e64fd3039a32d53576c24990ef1c3/thermo/chemical.py#L1297-L1311
def mass_fractions(self): r'''Dictionary of atom:mass-weighted fractional occurence of elements. Useful when performing mass balances. For atom-fraction occurences, see :obj:`atom_fractions`. Examples -------- >>> Chemical('water').mass_fractions {'H': 0.11189834...
[ "def", "mass_fractions", "(", "self", ")", ":", "if", "self", ".", "__mass_fractions", ":", "return", "self", ".", "__mass_fractions", "else", ":", "self", ".", "__mass_fractions", "=", "mass_fractions", "(", "self", ".", "atoms", ",", "self", ".", "MW", "...
r'''Dictionary of atom:mass-weighted fractional occurence of elements. Useful when performing mass balances. For atom-fraction occurences, see :obj:`atom_fractions`. Examples -------- >>> Chemical('water').mass_fractions {'H': 0.11189834407236524, 'O': 0.8881016559276347...
[ "r", "Dictionary", "of", "atom", ":", "mass", "-", "weighted", "fractional", "occurence", "of", "elements", ".", "Useful", "when", "performing", "mass", "balances", ".", "For", "atom", "-", "fraction", "occurences", "see", ":", "obj", ":", "atom_fractions", ...
python
valid
philippelt/netatmo-api-python
lnetatmo.py
https://github.com/philippelt/netatmo-api-python/blob/d749fca3637c07c2943aba7992f683fff1812f77/lnetatmo.py#L684-L695
def motionDetected(self, home=None, camera=None): """ Return True if movement has been detected """ try: cam_id = self.cameraByName(camera=camera, home=home)['id'] except TypeError: logger.warning("personSeenByCamera: Camera name or home is unknown") ...
[ "def", "motionDetected", "(", "self", ",", "home", "=", "None", ",", "camera", "=", "None", ")", ":", "try", ":", "cam_id", "=", "self", ".", "cameraByName", "(", "camera", "=", "camera", ",", "home", "=", "home", ")", "[", "'id'", "]", "except", "...
Return True if movement has been detected
[ "Return", "True", "if", "movement", "has", "been", "detected" ]
python
train
relekang/python-semantic-release
semantic_release/ci_checks.py
https://github.com/relekang/python-semantic-release/blob/76123f410180599a19e7c48da413880185bbea20/semantic_release/ci_checks.py#L31-L39
def travis(branch: str): """ Performs necessary checks to ensure that the travis build is one that should create releases. :param branch: The branch the environment should be running against. """ assert os.environ.get('TRAVIS_BRANCH') == branch assert os.environ.get('TRAVIS_PULL_REQUEST') =...
[ "def", "travis", "(", "branch", ":", "str", ")", ":", "assert", "os", ".", "environ", ".", "get", "(", "'TRAVIS_BRANCH'", ")", "==", "branch", "assert", "os", ".", "environ", ".", "get", "(", "'TRAVIS_PULL_REQUEST'", ")", "==", "'false'" ]
Performs necessary checks to ensure that the travis build is one that should create releases. :param branch: The branch the environment should be running against.
[ "Performs", "necessary", "checks", "to", "ensure", "that", "the", "travis", "build", "is", "one", "that", "should", "create", "releases", "." ]
python
train
lablup/backend.ai-client-py
src/ai/backend/client/kernel.py
https://github.com/lablup/backend.ai-client-py/blob/a063d774fea6f4350b89498c40d3c837ec3029a7/src/ai/backend/client/kernel.py#L364-L418
async def download(self, files: Sequence[Union[str, Path]], dest: Union[str, Path] = '.', show_progress: bool = False): ''' Downloads the given list of files from the compute session. :param files: The list of file paths in the compute session. ...
[ "async", "def", "download", "(", "self", ",", "files", ":", "Sequence", "[", "Union", "[", "str", ",", "Path", "]", "]", ",", "dest", ":", "Union", "[", "str", ",", "Path", "]", "=", "'.'", ",", "show_progress", ":", "bool", "=", "False", ")", ":...
Downloads the given list of files from the compute session. :param files: The list of file paths in the compute session. If they are relative paths, the path is calculated from ``/home/work`` in the compute session container. :param dest: The destination directory in the client-...
[ "Downloads", "the", "given", "list", "of", "files", "from", "the", "compute", "session", "." ]
python
train
datastore/datastore
datastore/core/basic.py
https://github.com/datastore/datastore/blob/7ccf0cd4748001d3dbf5e6dda369b0f63e0269d3/datastore/core/basic.py#L451-L456
def query(self, query): '''Returns an iterable of objects matching criteria expressed in `query`. LoggingDatastore logs the access. ''' self.logger.info('%s: query %s' % (self, query)) return super(LoggingDatastore, self).query(query)
[ "def", "query", "(", "self", ",", "query", ")", ":", "self", ".", "logger", ".", "info", "(", "'%s: query %s'", "%", "(", "self", ",", "query", ")", ")", "return", "super", "(", "LoggingDatastore", ",", "self", ")", ".", "query", "(", "query", ")" ]
Returns an iterable of objects matching criteria expressed in `query`. LoggingDatastore logs the access.
[ "Returns", "an", "iterable", "of", "objects", "matching", "criteria", "expressed", "in", "query", ".", "LoggingDatastore", "logs", "the", "access", "." ]
python
train
pmacosta/pmisc
pmisc/compat3.py
https://github.com/pmacosta/pmisc/blob/dd2bb32e59eee872f1ef2db2d9921a396ab9f50b/pmisc/compat3.py#L30-L38
def _readlines(fname, fpointer1=open, fpointer2=open): # pragma: no cover """Read all lines from file.""" # fpointer1, fpointer2 arguments to ease testing try: with fpointer1(fname, "r") as fobj: return fobj.readlines() except UnicodeDecodeError: # pragma: no cover with fpo...
[ "def", "_readlines", "(", "fname", ",", "fpointer1", "=", "open", ",", "fpointer2", "=", "open", ")", ":", "# pragma: no cover", "# fpointer1, fpointer2 arguments to ease testing", "try", ":", "with", "fpointer1", "(", "fname", ",", "\"r\"", ")", "as", "fobj", "...
Read all lines from file.
[ "Read", "all", "lines", "from", "file", "." ]
python
train
scot-dev/scot
scot/datatools.py
https://github.com/scot-dev/scot/blob/48598b79d4400dad893b134cd2194715511facda/scot/datatools.py#L28-L68
def cut_segments(x2d, tr, start, stop): """Cut continuous signal into segments. Parameters ---------- x2d : array, shape (m, n) Input data with m signals and n samples. tr : list of int Trigger positions. start : int Window start (offset relative to trigger). stop : ...
[ "def", "cut_segments", "(", "x2d", ",", "tr", ",", "start", ",", "stop", ")", ":", "if", "start", "!=", "int", "(", "start", ")", ":", "raise", "ValueError", "(", "\"start index must be an integer\"", ")", "if", "stop", "!=", "int", "(", "stop", ")", "...
Cut continuous signal into segments. Parameters ---------- x2d : array, shape (m, n) Input data with m signals and n samples. tr : list of int Trigger positions. start : int Window start (offset relative to trigger). stop : int Window end (offset relative to trig...
[ "Cut", "continuous", "signal", "into", "segments", "." ]
python
train
mcs07/PubChemPy
pubchempy.py
https://github.com/mcs07/PubChemPy/blob/e3c4f4a9b6120433e5cc3383464c7a79e9b2b86e/pubchempy.py#L795-L799
def sids(self): """Requires an extra request. Result is cached.""" if self.cid: results = get_json(self.cid, operation='sids') return results['InformationList']['Information'][0]['SID'] if results else []
[ "def", "sids", "(", "self", ")", ":", "if", "self", ".", "cid", ":", "results", "=", "get_json", "(", "self", ".", "cid", ",", "operation", "=", "'sids'", ")", "return", "results", "[", "'InformationList'", "]", "[", "'Information'", "]", "[", "0", "...
Requires an extra request. Result is cached.
[ "Requires", "an", "extra", "request", ".", "Result", "is", "cached", "." ]
python
train
cokelaer/spectrum
src/spectrum/window.py
https://github.com/cokelaer/spectrum/blob/bad6c32e3f10e185098748f67bb421b378b06afe/src/spectrum/window.py#L1206-L1227
def window_riesz(N): r"""Riesz tapering window :param N: window length .. math:: w(n) = 1 - \left| \frac{n}{N/2} \right|^2 with :math:`-N/2 \leq n \leq N/2`. .. plot:: :width: 80% :include-source: from spectrum import window_visu window_visu(64, 'riesz') .....
[ "def", "window_riesz", "(", "N", ")", ":", "n", "=", "linspace", "(", "-", "N", "/", "2.", ",", "(", "N", ")", "/", "2.", ",", "N", ")", "w", "=", "1", "-", "abs", "(", "n", "/", "(", "N", "/", "2.", ")", ")", "**", "2.", "return", "w" ...
r"""Riesz tapering window :param N: window length .. math:: w(n) = 1 - \left| \frac{n}{N/2} \right|^2 with :math:`-N/2 \leq n \leq N/2`. .. plot:: :width: 80% :include-source: from spectrum import window_visu window_visu(64, 'riesz') .. seealso:: :func:`create_...
[ "r", "Riesz", "tapering", "window" ]
python
valid
tensorflow/hub
tensorflow_hub/estimator.py
https://github.com/tensorflow/hub/blob/09f45963f6787322967b6fec61459f3ac56fbb27/tensorflow_hub/estimator.py#L171-L214
def _make_estimator_serving_session(estimator, serving_input_fn, checkpoint_path): """Returns a session constructed using `estimator` and `serving_input_fn`. The Estimator API does not provide an API to construct a graph and session, making it necessary for this function to re...
[ "def", "_make_estimator_serving_session", "(", "estimator", ",", "serving_input_fn", ",", "checkpoint_path", ")", ":", "with", "tf", ".", "Graph", "(", ")", ".", "as_default", "(", ")", "as", "g", ":", "mode", "=", "tf_v1", ".", "estimator", ".", "ModeKeys",...
Returns a session constructed using `estimator` and `serving_input_fn`. The Estimator API does not provide an API to construct a graph and session, making it necessary for this function to replicate how an estimator builds a graph. This code is based on `Estimator.export_savedmodel` (another function that h...
[ "Returns", "a", "session", "constructed", "using", "estimator", "and", "serving_input_fn", "." ]
python
train
google/grumpy
third_party/stdlib/fpformat.py
https://github.com/google/grumpy/blob/3ec87959189cfcdeae82eb68a47648ac25ceb10b/third_party/stdlib/fpformat.py#L90-L104
def fix(x, digs): """Format x as [-]ddd.ddd with 'digs' digits after the point and at least one digit before. If digs <= 0, the point is suppressed.""" if type(x) != type(''): x = repr(x) try: sign, intpart, fraction, expo = extract(x) except NotANumber: return x intpart, fra...
[ "def", "fix", "(", "x", ",", "digs", ")", ":", "if", "type", "(", "x", ")", "!=", "type", "(", "''", ")", ":", "x", "=", "repr", "(", "x", ")", "try", ":", "sign", ",", "intpart", ",", "fraction", ",", "expo", "=", "extract", "(", "x", ")",...
Format x as [-]ddd.ddd with 'digs' digits after the point and at least one digit before. If digs <= 0, the point is suppressed.
[ "Format", "x", "as", "[", "-", "]", "ddd", ".", "ddd", "with", "digs", "digits", "after", "the", "point", "and", "at", "least", "one", "digit", "before", ".", "If", "digs", "<", "=", "0", "the", "point", "is", "suppressed", "." ]
python
valid
pantsbuild/pants
src/python/pants/build_graph/target.py
https://github.com/pantsbuild/pants/blob/b72e650da0df685824ffdcc71988b8c282d0962d/src/python/pants/build_graph/target.py#L377-L382
def assert_list(self, putative_list, expected_type=string_types, key_arg=None): """ :API: public """ return assert_list(putative_list, expected_type, key_arg=key_arg, raise_type=lambda msg: TargetDefinitionException(self, msg))
[ "def", "assert_list", "(", "self", ",", "putative_list", ",", "expected_type", "=", "string_types", ",", "key_arg", "=", "None", ")", ":", "return", "assert_list", "(", "putative_list", ",", "expected_type", ",", "key_arg", "=", "key_arg", ",", "raise_type", "...
:API: public
[ ":", "API", ":", "public" ]
python
train
bitshares/uptick
uptick/vesting.py
https://github.com/bitshares/uptick/blob/66c102200fdbf96cef4fd55cc69d00e690f62001/uptick/vesting.py#L56-L63
def reserve(ctx, amount, symbol, account): """ Reserve/Burn tokens """ print_tx( ctx.bitshares.reserve( Amount(amount, symbol, bitshares_instance=ctx.bitshares), account=account ) )
[ "def", "reserve", "(", "ctx", ",", "amount", ",", "symbol", ",", "account", ")", ":", "print_tx", "(", "ctx", ".", "bitshares", ".", "reserve", "(", "Amount", "(", "amount", ",", "symbol", ",", "bitshares_instance", "=", "ctx", ".", "bitshares", ")", "...
Reserve/Burn tokens
[ "Reserve", "/", "Burn", "tokens" ]
python
train
not-na/peng3d
peng3d/gui/menus.py
https://github.com/not-na/peng3d/blob/1151be665b26cc8a479f6307086ba919e4d32d85/peng3d/gui/menus.py#L473-L487
def update_progressbar(self): """ Updates the progressbar by re-calculating the label. It is not required to manually call this method since setting any of the properties of this class will automatically trigger a re-calculation. """ n,nmin,nmax = self.wprogressb...
[ "def", "update_progressbar", "(", "self", ")", ":", "n", ",", "nmin", ",", "nmax", "=", "self", ".", "wprogressbar", ".", "n", ",", "self", ".", "wprogressbar", ".", "nmin", ",", "self", ".", "wprogressbar", ".", "nmax", "if", "(", "nmax", "-", "nmin...
Updates the progressbar by re-calculating the label. It is not required to manually call this method since setting any of the properties of this class will automatically trigger a re-calculation.
[ "Updates", "the", "progressbar", "by", "re", "-", "calculating", "the", "label", ".", "It", "is", "not", "required", "to", "manually", "call", "this", "method", "since", "setting", "any", "of", "the", "properties", "of", "this", "class", "will", "automatical...
python
test
googledatalab/pydatalab
solutionbox/structured_data/mltoolbox/_structured_data/trainer/task.py
https://github.com/googledatalab/pydatalab/blob/d9031901d5bca22fe0d5925d204e6698df9852e1/solutionbox/structured_data/mltoolbox/_structured_data/trainer/task.py#L30-L57
def get_reader_input_fn(train_config, preprocess_output_dir, model_type, data_paths, batch_size, shuffle, num_epochs=None): """Builds input layer for training.""" def get_input_features(): """Read the input features from the given data paths.""" _, examples = util.read_examples( ...
[ "def", "get_reader_input_fn", "(", "train_config", ",", "preprocess_output_dir", ",", "model_type", ",", "data_paths", ",", "batch_size", ",", "shuffle", ",", "num_epochs", "=", "None", ")", ":", "def", "get_input_features", "(", ")", ":", "\"\"\"Read the input feat...
Builds input layer for training.
[ "Builds", "input", "layer", "for", "training", "." ]
python
train
annoviko/pyclustering
pyclustering/gcolor/sync.py
https://github.com/annoviko/pyclustering/blob/98aa0dd89fd36f701668fb1eb29c8fb5662bf7d0/pyclustering/gcolor/sync.py#L116-L127
def _create_connections(self, graph_matrix): """! @brief Creates connection in the network in line with graph. @param[in] graph_matrix (list): Matrix representation of the graph. """ for row in range(0, len(graph_matrix)): for column in rang...
[ "def", "_create_connections", "(", "self", ",", "graph_matrix", ")", ":", "for", "row", "in", "range", "(", "0", ",", "len", "(", "graph_matrix", ")", ")", ":", "for", "column", "in", "range", "(", "0", ",", "len", "(", "graph_matrix", "[", "row", "]...
! @brief Creates connection in the network in line with graph. @param[in] graph_matrix (list): Matrix representation of the graph.
[ "!" ]
python
valid
fboender/ansible-cmdb
src/ansible-cmdb.py
https://github.com/fboender/ansible-cmdb/blob/ebd960ac10684e8c9ec2b12751bba2c4c9504ab7/src/ansible-cmdb.py#L48-L65
def get_data_dir(): """ Find out our installation prefix and data directory. These can be in different places depending on how ansible-cmdb was installed. """ data_dir_paths = [ os.path.join(os.path.dirname(ansiblecmdb.__file__), 'data'), os.path.join(os.path.dirname(sys.argv[0]), '....
[ "def", "get_data_dir", "(", ")", ":", "data_dir_paths", "=", "[", "os", ".", "path", ".", "join", "(", "os", ".", "path", ".", "dirname", "(", "ansiblecmdb", ".", "__file__", ")", ",", "'data'", ")", ",", "os", ".", "path", ".", "join", "(", "os", ...
Find out our installation prefix and data directory. These can be in different places depending on how ansible-cmdb was installed.
[ "Find", "out", "our", "installation", "prefix", "and", "data", "directory", ".", "These", "can", "be", "in", "different", "places", "depending", "on", "how", "ansible", "-", "cmdb", "was", "installed", "." ]
python
train