nwo stringlengths 5 106 | sha stringlengths 40 40 | path stringlengths 4 174 | language stringclasses 1
value | identifier stringlengths 1 140 | parameters stringlengths 0 87.7k | argument_list stringclasses 1
value | return_statement stringlengths 0 426k | docstring stringlengths 0 64.3k | docstring_summary stringlengths 0 26.3k | docstring_tokens list | function stringlengths 18 4.83M | function_tokens list | url stringlengths 83 304 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
SebKuzminsky/pycam | 55e3129f518e470040e79bb00515b4bfcf36c172 | pycam/Plugins/Tasks.py | python | Tasks._task_new | (self, widget=None, task_type="milling") | [] | def _task_new(self, widget=None, task_type="milling"):
with merge_history_and_block_events(self.core):
params = {"type": task_type}
params.update(self.core.get("get_default_parameter_values")("task",
set_name=task_ty... | [
"def",
"_task_new",
"(",
"self",
",",
"widget",
"=",
"None",
",",
"task_type",
"=",
"\"milling\"",
")",
":",
"with",
"merge_history_and_block_events",
"(",
"self",
".",
"core",
")",
":",
"params",
"=",
"{",
"\"type\"",
":",
"task_type",
"}",
"params",
".",... | https://github.com/SebKuzminsky/pycam/blob/55e3129f518e470040e79bb00515b4bfcf36c172/pycam/Plugins/Tasks.py#L220-L227 | ||||
KalleHallden/AutoTimer | 2d954216700c4930baa154e28dbddc34609af7ce | env/lib/python2.7/site-packages/pip/_vendor/requests/cookies.py | python | MockResponse.info | (self) | return self._headers | [] | def info(self):
return self._headers | [
"def",
"info",
"(",
"self",
")",
":",
"return",
"self",
".",
"_headers"
] | https://github.com/KalleHallden/AutoTimer/blob/2d954216700c4930baa154e28dbddc34609af7ce/env/lib/python2.7/site-packages/pip/_vendor/requests/cookies.py#L111-L112 | |||
wikimedia/pywikibot | 81a01ffaec7271bf5b4b170f85a80388420a4e78 | pywikibot/pagegenerators.py | python | GeneratorFactory._handle_interwiki | (self, value: str) | return LanguageLinksPageGenerator(page) | Handle `-interwiki` argument. | Handle `-interwiki` argument. | [
"Handle",
"-",
"interwiki",
"argument",
"."
] | def _handle_interwiki(self, value: str) -> HANDLER_RETURN_TYPE:
"""Handle `-interwiki` argument."""
if not value:
value = i18n.input('pywikibot-enter-page-processing')
page = pywikibot.Page(pywikibot.Link(value, self.site))
return LanguageLinksPageGenerator(page) | [
"def",
"_handle_interwiki",
"(",
"self",
",",
"value",
":",
"str",
")",
"->",
"HANDLER_RETURN_TYPE",
":",
"if",
"not",
"value",
":",
"value",
"=",
"i18n",
".",
"input",
"(",
"'pywikibot-enter-page-processing'",
")",
"page",
"=",
"pywikibot",
".",
"Page",
"("... | https://github.com/wikimedia/pywikibot/blob/81a01ffaec7271bf5b4b170f85a80388420a4e78/pywikibot/pagegenerators.py#L917-L922 | |
latenighttales/alcali | 8939fc7901b4d8273c4106a242afdf94c0b05ea8 | docker/saltconfig/salt/_returners/alcali.py | python | get_jids_filter | (count, filter_find_job=True) | Return a list of all job ids
:param int count: show not more than the count of most recent jobs
:param bool filter_find_jobs: filter out 'saltutil.find_job' jobs | Return a list of all job ids
:param int count: show not more than the count of most recent jobs
:param bool filter_find_jobs: filter out 'saltutil.find_job' jobs | [
"Return",
"a",
"list",
"of",
"all",
"job",
"ids",
":",
"param",
"int",
"count",
":",
"show",
"not",
"more",
"than",
"the",
"count",
"of",
"most",
"recent",
"jobs",
":",
"param",
"bool",
"filter_find_jobs",
":",
"filter",
"out",
"saltutil",
".",
"find_job... | def get_jids_filter(count, filter_find_job=True):
"""
Return a list of all job ids
:param int count: show not more than the count of most recent jobs
:param bool filter_find_jobs: filter out 'saltutil.find_job' jobs
"""
with _get_serv(ret=None, commit=True) as cur:
sql = """SELECT * FRO... | [
"def",
"get_jids_filter",
"(",
"count",
",",
"filter_find_job",
"=",
"True",
")",
":",
"with",
"_get_serv",
"(",
"ret",
"=",
"None",
",",
"commit",
"=",
"True",
")",
"as",
"cur",
":",
"sql",
"=",
"\"\"\"SELECT * FROM (\n SELECT DISTINCT `jid` ,... | https://github.com/latenighttales/alcali/blob/8939fc7901b4d8273c4106a242afdf94c0b05ea8/docker/saltconfig/salt/_returners/alcali.py#L469-L494 | ||
deepchem/deepchem | 054eb4b2b082e3df8e1a8e77f36a52137ae6e375 | contrib/atomicconv/splits/pdbbind_temporal_split.py | python | load_pdbbind_labels | (labels_file) | return contents_df | Loads pdbbind labels as dataframe
Parameters
----------
labels_file: str
Location of PDBbind datafile.
Returns
-------
contents_df: pd.DataFrame
Dataframe containing contents of PDBbind datafile. | Loads pdbbind labels as dataframe | [
"Loads",
"pdbbind",
"labels",
"as",
"dataframe"
] | def load_pdbbind_labels(labels_file):
"""Loads pdbbind labels as dataframe
Parameters
----------
labels_file: str
Location of PDBbind datafile.
Returns
-------
contents_df: pd.DataFrame
Dataframe containing contents of PDBbind datafile.
"""
contents = []
with open(labels_file) as f:
... | [
"def",
"load_pdbbind_labels",
"(",
"labels_file",
")",
":",
"contents",
"=",
"[",
"]",
"with",
"open",
"(",
"labels_file",
")",
"as",
"f",
":",
"for",
"line",
"in",
"f",
":",
"if",
"line",
".",
"startswith",
"(",
"\"#\"",
")",
":",
"continue",
"else",
... | https://github.com/deepchem/deepchem/blob/054eb4b2b082e3df8e1a8e77f36a52137ae6e375/contrib/atomicconv/splits/pdbbind_temporal_split.py#L8-L40 | |
securesystemslab/zippy | ff0e84ac99442c2c55fe1d285332cfd4e185e089 | zippy/lib-python/3/xml/sax/saxutils.py | python | XMLFilterBase.startElementNS | (self, name, qname, attrs) | [] | def startElementNS(self, name, qname, attrs):
self._cont_handler.startElementNS(name, qname, attrs) | [
"def",
"startElementNS",
"(",
"self",
",",
"name",
",",
"qname",
",",
"attrs",
")",
":",
"self",
".",
"_cont_handler",
".",
"startElementNS",
"(",
"name",
",",
"qname",
",",
"attrs",
")"
] | https://github.com/securesystemslab/zippy/blob/ff0e84ac99442c2c55fe1d285332cfd4e185e089/zippy/lib-python/3/xml/sax/saxutils.py#L240-L241 | ||||
zhl2008/awd-platform | 0416b31abea29743387b10b3914581fbe8e7da5e | web_flaskbb/Python-2.7.9/Lib/bsddb/dbshelve.py | python | DBShelfCursor.first | (self, flags=0) | return self.get_1(flags|db.DB_FIRST) | [] | def first(self, flags=0): return self.get_1(flags|db.DB_FIRST) | [
"def",
"first",
"(",
"self",
",",
"flags",
"=",
"0",
")",
":",
"return",
"self",
".",
"get_1",
"(",
"flags",
"|",
"db",
".",
"DB_FIRST",
")"
] | https://github.com/zhl2008/awd-platform/blob/0416b31abea29743387b10b3914581fbe8e7da5e/web_flaskbb/Python-2.7.9/Lib/bsddb/dbshelve.py#L334-L334 | |||
blinktrade/bitex | a4896e7faef9c4aa0ca5325f18b77db67003764e | apps/trade/models.py | python | Position.get_positions_by_account_broker | (session, account_id, broker_id) | return session.query(Position).filter_by(account_id = account_id).filter_by(broker_id = broker_id ) | [] | def get_positions_by_account_broker(session, account_id, broker_id):
return session.query(Position).filter_by(account_id = account_id).filter_by(broker_id = broker_id ) | [
"def",
"get_positions_by_account_broker",
"(",
"session",
",",
"account_id",
",",
"broker_id",
")",
":",
"return",
"session",
".",
"query",
"(",
"Position",
")",
".",
"filter_by",
"(",
"account_id",
"=",
"account_id",
")",
".",
"filter_by",
"(",
"broker_id",
"... | https://github.com/blinktrade/bitex/blob/a4896e7faef9c4aa0ca5325f18b77db67003764e/apps/trade/models.py#L736-L737 | |||
zhl2008/awd-platform | 0416b31abea29743387b10b3914581fbe8e7da5e | web_hxb2/lib/python3.5/site-packages/pip/_vendor/requests/utils.py | python | unquote_unreserved | (uri) | return ''.join(parts) | Un-escape any percent-escape sequences in a URI that are unreserved
characters. This leaves all reserved, illegal and non-ASCII bytes encoded.
:rtype: str | Un-escape any percent-escape sequences in a URI that are unreserved
characters. This leaves all reserved, illegal and non-ASCII bytes encoded. | [
"Un",
"-",
"escape",
"any",
"percent",
"-",
"escape",
"sequences",
"in",
"a",
"URI",
"that",
"are",
"unreserved",
"characters",
".",
"This",
"leaves",
"all",
"reserved",
"illegal",
"and",
"non",
"-",
"ASCII",
"bytes",
"encoded",
"."
] | def unquote_unreserved(uri):
"""Un-escape any percent-escape sequences in a URI that are unreserved
characters. This leaves all reserved, illegal and non-ASCII bytes encoded.
:rtype: str
"""
parts = uri.split('%')
for i in range(1, len(parts)):
h = parts[i][0:2]
if len(h) == 2 a... | [
"def",
"unquote_unreserved",
"(",
"uri",
")",
":",
"parts",
"=",
"uri",
".",
"split",
"(",
"'%'",
")",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"len",
"(",
"parts",
")",
")",
":",
"h",
"=",
"parts",
"[",
"i",
"]",
"[",
"0",
":",
"2",
"]",
... | https://github.com/zhl2008/awd-platform/blob/0416b31abea29743387b10b3914581fbe8e7da5e/web_hxb2/lib/python3.5/site-packages/pip/_vendor/requests/utils.py#L438-L459 | |
dj-stripe/dj-stripe | cf15b07c754525077098e2b906108425a1f657e0 | djstripe/models/billing.py | python | UsageRecordSummary.__str__ | (self) | return f"Usage Summary for {self.subscription_item} ({self.invoice}) is {self.total_usage}" | [] | def __str__(self):
return f"Usage Summary for {self.subscription_item} ({self.invoice}) is {self.total_usage}" | [
"def",
"__str__",
"(",
"self",
")",
":",
"return",
"f\"Usage Summary for {self.subscription_item} ({self.invoice}) is {self.total_usage}\""
] | https://github.com/dj-stripe/dj-stripe/blob/cf15b07c754525077098e2b906108425a1f657e0/djstripe/models/billing.py#L2077-L2078 | |||
linxid/Machine_Learning_Study_Path | 558e82d13237114bbb8152483977806fc0c222af | Machine Learning In Action/Chapter5-LogisticRegression/venv/Lib/site-packages/pip/_vendor/lockfile/linklockfile.py | python | LinkLockFile.i_am_locking | (self) | return (self.is_locked() and
os.path.exists(self.unique_name) and
os.stat(self.unique_name).st_nlink == 2) | [] | def i_am_locking(self):
return (self.is_locked() and
os.path.exists(self.unique_name) and
os.stat(self.unique_name).st_nlink == 2) | [
"def",
"i_am_locking",
"(",
"self",
")",
":",
"return",
"(",
"self",
".",
"is_locked",
"(",
")",
"and",
"os",
".",
"path",
".",
"exists",
"(",
"self",
".",
"unique_name",
")",
"and",
"os",
".",
"stat",
"(",
"self",
".",
"unique_name",
")",
".",
"st... | https://github.com/linxid/Machine_Learning_Study_Path/blob/558e82d13237114bbb8152483977806fc0c222af/Machine Learning In Action/Chapter5-LogisticRegression/venv/Lib/site-packages/pip/_vendor/lockfile/linklockfile.py#L66-L69 | |||
edgedb/edgedb | 872bf5abbb10f7c72df21f57635238ed27b9f280 | edb/edgeql/parser/grammar/ddl.py | python | CreateObjectTypeStmt.reduce_CreateAbstractObjectTypeStmt | (self, *kids) | r"""%reduce \
CREATE ABSTRACT TYPE NodeName \
OptExtendingSimple OptCreateObjectTypeCommandsBlock \ | r"""%reduce \
CREATE ABSTRACT TYPE NodeName \
OptExtendingSimple OptCreateObjectTypeCommandsBlock \ | [
"r",
"%reduce",
"\\",
"CREATE",
"ABSTRACT",
"TYPE",
"NodeName",
"\\",
"OptExtendingSimple",
"OptCreateObjectTypeCommandsBlock",
"\\"
] | def reduce_CreateAbstractObjectTypeStmt(self, *kids):
r"""%reduce \
CREATE ABSTRACT TYPE NodeName \
OptExtendingSimple OptCreateObjectTypeCommandsBlock \
"""
self.val = qlast.CreateObjectType(
name=kids[3].val,
bases=kids[4].val,
abstra... | [
"def",
"reduce_CreateAbstractObjectTypeStmt",
"(",
"self",
",",
"*",
"kids",
")",
":",
"self",
".",
"val",
"=",
"qlast",
".",
"CreateObjectType",
"(",
"name",
"=",
"kids",
"[",
"3",
"]",
".",
"val",
",",
"bases",
"=",
"kids",
"[",
"4",
"]",
".",
"val... | https://github.com/edgedb/edgedb/blob/872bf5abbb10f7c72df21f57635238ed27b9f280/edb/edgeql/parser/grammar/ddl.py#L1751-L1761 | ||
theotherp/nzbhydra | 4b03d7f769384b97dfc60dade4806c0fc987514e | libs/mailbox.py | python | Mailbox.__getitem__ | (self, key) | Return the keyed message; raise KeyError if it doesn't exist. | Return the keyed message; raise KeyError if it doesn't exist. | [
"Return",
"the",
"keyed",
"message",
";",
"raise",
"KeyError",
"if",
"it",
"doesn",
"t",
"exist",
"."
] | def __getitem__(self, key):
"""Return the keyed message; raise KeyError if it doesn't exist."""
if not self._factory:
return self.get_message(key)
else:
return self._factory(self.get_file(key)) | [
"def",
"__getitem__",
"(",
"self",
",",
"key",
")",
":",
"if",
"not",
"self",
".",
"_factory",
":",
"return",
"self",
".",
"get_message",
"(",
"key",
")",
"else",
":",
"return",
"self",
".",
"_factory",
"(",
"self",
".",
"get_file",
"(",
"key",
")",
... | https://github.com/theotherp/nzbhydra/blob/4b03d7f769384b97dfc60dade4806c0fc987514e/libs/mailbox.py#L77-L82 | ||
osmr/imgclsmob | f2993d3ce73a2f7ddba05da3891defb08547d504 | pytorch/pytorchcv/models/pyramidnet_cifar.py | python | pyramidnet110_a84_cifar100 | (num_classes=100, **kwargs) | return get_pyramidnet_cifar(
num_classes=num_classes,
blocks=110,
alpha=84,
bottleneck=False,
model_name="pyramidnet110_a84_cifar100",
**kwargs) | PyramidNet-110 (a=84) model for CIFAR-100 from 'Deep Pyramidal Residual Networks,' https://arxiv.org/abs/1610.02915.
Parameters:
----------
num_classes : int, default 100
Number of classification classes.
pretrained : bool, default False
Whether to load the pretrained weights for model.... | PyramidNet-110 (a=84) model for CIFAR-100 from 'Deep Pyramidal Residual Networks,' https://arxiv.org/abs/1610.02915. | [
"PyramidNet",
"-",
"110",
"(",
"a",
"=",
"84",
")",
"model",
"for",
"CIFAR",
"-",
"100",
"from",
"Deep",
"Pyramidal",
"Residual",
"Networks",
"https",
":",
"//",
"arxiv",
".",
"org",
"/",
"abs",
"/",
"1610",
".",
"02915",
"."
] | def pyramidnet110_a84_cifar100(num_classes=100, **kwargs):
"""
PyramidNet-110 (a=84) model for CIFAR-100 from 'Deep Pyramidal Residual Networks,' https://arxiv.org/abs/1610.02915.
Parameters:
----------
num_classes : int, default 100
Number of classification classes.
pretrained : bool, ... | [
"def",
"pyramidnet110_a84_cifar100",
"(",
"num_classes",
"=",
"100",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"get_pyramidnet_cifar",
"(",
"num_classes",
"=",
"num_classes",
",",
"blocks",
"=",
"110",
",",
"alpha",
"=",
"84",
",",
"bottleneck",
"=",
"Fal... | https://github.com/osmr/imgclsmob/blob/f2993d3ce73a2f7ddba05da3891defb08547d504/pytorch/pytorchcv/models/pyramidnet_cifar.py#L250-L269 | |
amundsen-io/amundsendatabuilder | a0af611350fde12438450d4bfd83b226ef220c3f | databuilder/extractor/csv_extractor.py | python | CsvTableColumnExtractor.extract | (self) | Yield the csv result one at a time.
convert the result to model if a model_class is provided | Yield the csv result one at a time.
convert the result to model if a model_class is provided | [
"Yield",
"the",
"csv",
"result",
"one",
"at",
"a",
"time",
".",
"convert",
"the",
"result",
"to",
"model",
"if",
"a",
"model_class",
"is",
"provided"
] | def extract(self) -> Any:
"""
Yield the csv result one at a time.
convert the result to model if a model_class is provided
"""
try:
return next(self._iter)
except StopIteration:
return None
except Exception as e:
raise e | [
"def",
"extract",
"(",
"self",
")",
"->",
"Any",
":",
"try",
":",
"return",
"next",
"(",
"self",
".",
"_iter",
")",
"except",
"StopIteration",
":",
"return",
"None",
"except",
"Exception",
"as",
"e",
":",
"raise",
"e"
] | https://github.com/amundsen-io/amundsendatabuilder/blob/a0af611350fde12438450d4bfd83b226ef220c3f/databuilder/extractor/csv_extractor.py#L245-L255 | ||
krintoxi/NoobSec-Toolkit | 38738541cbc03cedb9a3b3ed13b629f781ad64f6 | NoobSecToolkit /tools/inject/tamper/uppercase.py | python | tamper | (payload, **kwargs) | return retVal | Replaces each keyword character with upper case value
Tested against:
* Microsoft SQL Server 2005
* MySQL 4, 5.0 and 5.5
* Oracle 10g
* PostgreSQL 8.3, 8.4, 9.0
Notes:
* Useful to bypass very weak and bespoke web application firewalls
that has poorly written p... | Replaces each keyword character with upper case value | [
"Replaces",
"each",
"keyword",
"character",
"with",
"upper",
"case",
"value"
] | def tamper(payload, **kwargs):
"""
Replaces each keyword character with upper case value
Tested against:
* Microsoft SQL Server 2005
* MySQL 4, 5.0 and 5.5
* Oracle 10g
* PostgreSQL 8.3, 8.4, 9.0
Notes:
* Useful to bypass very weak and bespoke web application fi... | [
"def",
"tamper",
"(",
"payload",
",",
"*",
"*",
"kwargs",
")",
":",
"retVal",
"=",
"payload",
"if",
"payload",
":",
"for",
"match",
"in",
"re",
".",
"finditer",
"(",
"r\"[A-Za-z_]+\"",
",",
"retVal",
")",
":",
"word",
"=",
"match",
".",
"group",
"(",... | https://github.com/krintoxi/NoobSec-Toolkit/blob/38738541cbc03cedb9a3b3ed13b629f781ad64f6/NoobSecToolkit /tools/inject/tamper/uppercase.py#L18-L46 | |
HeinleinSupport/check_mk_extensions | aa7d7389b812ed00f91dad61d66fb676284897d8 | postfix_mailq_details/web/plugins/wato/postfix_mailq_details.py | python | _parameter_valuespec_postfix_mailq_details | () | return Transform(
Dictionary(
elements = [
('level',
Tuple(
help = _("These levels are applied to the number of Email that are "
"currently in the specified mail queue."),
elements = [
... | [] | def _parameter_valuespec_postfix_mailq_details():
return Transform(
Dictionary(
elements = [
('level',
Tuple(
help = _("These levels are applied to the number of Email that are "
"currently in the specified mail ... | [
"def",
"_parameter_valuespec_postfix_mailq_details",
"(",
")",
":",
"return",
"Transform",
"(",
"Dictionary",
"(",
"elements",
"=",
"[",
"(",
"'level'",
",",
"Tuple",
"(",
"help",
"=",
"_",
"(",
"\"These levels are applied to the number of Email that are \"",
"\"current... | https://github.com/HeinleinSupport/check_mk_extensions/blob/aa7d7389b812ed00f91dad61d66fb676284897d8/postfix_mailq_details/web/plugins/wato/postfix_mailq_details.py#L39-L58 | |||
quantumlib/Cirq | 89f88b01d69222d3f1ec14d649b7b3a85ed9211f | cirq-core/cirq/contrib/qasm_import/_parser.py | python | QasmParser.p_term | (self, p) | term : NUMBER
| NATURAL_NUMBER
| PI | term : NUMBER
| NATURAL_NUMBER
| PI | [
"term",
":",
"NUMBER",
"|",
"NATURAL_NUMBER",
"|",
"PI"
] | def p_term(self, p):
"""term : NUMBER
| NATURAL_NUMBER
| PI"""
p[0] = p[1] | [
"def",
"p_term",
"(",
"self",
",",
"p",
")",
":",
"p",
"[",
"0",
"]",
"=",
"p",
"[",
"1",
"]"
] | https://github.com/quantumlib/Cirq/blob/89f88b01d69222d3f1ec14d649b7b3a85ed9211f/cirq-core/cirq/contrib/qasm_import/_parser.py#L401-L405 | ||
errbotio/errbot | 66e1de8e1d7daa62be7f9ed1b2ac8832f09fa25d | errbot/flow.py | python | FlowExecutor._create_new_flow | (
flow_root, requestor: Identifier, initial_command
) | return None, None | Helper method to create a new FLow. | Helper method to create a new FLow. | [
"Helper",
"method",
"to",
"create",
"a",
"new",
"FLow",
"."
] | def _create_new_flow(
flow_root, requestor: Identifier, initial_command
) -> Tuple[Optional[Flow], Optional[FlowNode]]:
"""
Helper method to create a new FLow.
"""
empty_context = {}
flow = Flow(flow_root, requestor, empty_context)
for possible_next_step in fl... | [
"def",
"_create_new_flow",
"(",
"flow_root",
",",
"requestor",
":",
"Identifier",
",",
"initial_command",
")",
"->",
"Tuple",
"[",
"Optional",
"[",
"Flow",
"]",
",",
"Optional",
"[",
"FlowNode",
"]",
"]",
":",
"empty_context",
"=",
"{",
"}",
"flow",
"=",
... | https://github.com/errbotio/errbot/blob/66e1de8e1d7daa62be7f9ed1b2ac8832f09fa25d/errbot/flow.py#L372-L384 | |
giantbranch/python-hacker-code | addbc8c73e7e6fb9e4fcadcec022fa1d3da4b96d | 我手敲的代码(中文注释)/chapter11/volatility/cache.py | python | CacheDecorator._cachewrapper | (self, f, s, *args, **kwargs) | return result | Wrapper for caching function calls | Wrapper for caching function calls | [
"Wrapper",
"for",
"caching",
"function",
"calls"
] | def _cachewrapper(self, f, s, *args, **kwargs):
"""Wrapper for caching function calls"""
## See if the path is callable:
if callable(self.path):
path = self.path(s, *args, **kwargs)
else:
path = self.path
## Check if the result can be retrieved
se... | [
"def",
"_cachewrapper",
"(",
"self",
",",
"f",
",",
"s",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"## See if the path is callable:",
"if",
"callable",
"(",
"self",
".",
"path",
")",
":",
"path",
"=",
"self",
".",
"path",
"(",
"s",
",",
... | https://github.com/giantbranch/python-hacker-code/blob/addbc8c73e7e6fb9e4fcadcec022fa1d3da4b96d/我手敲的代码(中文注释)/chapter11/volatility/cache.py#L561-L586 | |
shiyanlou/louplus-python | 4c61697259e286e3d9116c3299f170d019ba3767 | taobei/challenge-06/tbbuy/handlers/cart_product.py | python | cart_product_info | (id) | return json_response(cart_product=CartProductSchema().dump(cart_product)) | 查询购物车商品 | 查询购物车商品 | [
"查询购物车商品"
] | def cart_product_info(id):
"""查询购物车商品
"""
cart_product = CartProduct.query.filter(CartProduct.id == id).first()
if cart_product is None:
return json_response(ResponseCode.NOT_FOUND)
return json_response(cart_product=CartProductSchema().dump(cart_product)) | [
"def",
"cart_product_info",
"(",
"id",
")",
":",
"cart_product",
"=",
"CartProduct",
".",
"query",
".",
"filter",
"(",
"CartProduct",
".",
"id",
"==",
"id",
")",
".",
"first",
"(",
")",
"if",
"cart_product",
"is",
"None",
":",
"return",
"json_response",
... | https://github.com/shiyanlou/louplus-python/blob/4c61697259e286e3d9116c3299f170d019ba3767/taobei/challenge-06/tbbuy/handlers/cart_product.py#L97-L105 | |
gramps-project/gramps | 04d4651a43eb210192f40a9f8c2bad8ee8fa3753 | gramps/gen/lib/ldsord.py | python | LdsOrd.type2xml | (self) | return "" | Return type-representing string suitable for XML. | Return type-representing string suitable for XML. | [
"Return",
"type",
"-",
"representing",
"string",
"suitable",
"for",
"XML",
"."
] | def type2xml(self):
"""
Return type-representing string suitable for XML.
"""
for item in LdsOrd._TYPE_MAP:
if item[0] == self.type:
return item[2]
return "" | [
"def",
"type2xml",
"(",
"self",
")",
":",
"for",
"item",
"in",
"LdsOrd",
".",
"_TYPE_MAP",
":",
"if",
"item",
"[",
"0",
"]",
"==",
"self",
".",
"type",
":",
"return",
"item",
"[",
"2",
"]",
"return",
"\"\""
] | https://github.com/gramps-project/gramps/blob/04d4651a43eb210192f40a9f8c2bad8ee8fa3753/gramps/gen/lib/ldsord.py#L350-L357 | |
irmen/synthesizer | 9ee0969cb9756d63dffb3b4be2ff35e3ccfadbf9 | synthplayer/sample.py | python | Sample.envelope | (self, attack: float, decay: float, sustainlevel: float, release: float) | return self | Apply an ADSR volume envelope. A,D,R are in seconds, Sustainlevel is a factor. | Apply an ADSR volume envelope. A,D,R are in seconds, Sustainlevel is a factor. | [
"Apply",
"an",
"ADSR",
"volume",
"envelope",
".",
"A",
"D",
"R",
"are",
"in",
"seconds",
"Sustainlevel",
"is",
"a",
"factor",
"."
] | def envelope(self, attack: float, decay: float, sustainlevel: float, release: float) -> 'Sample':
"""Apply an ADSR volume envelope. A,D,R are in seconds, Sustainlevel is a factor."""
if self.__locked:
raise RuntimeError("cannot modify a locked sample")
assert attack >= 0 and decay >=... | [
"def",
"envelope",
"(",
"self",
",",
"attack",
":",
"float",
",",
"decay",
":",
"float",
",",
"sustainlevel",
":",
"float",
",",
"release",
":",
"float",
")",
"->",
"'Sample'",
":",
"if",
"self",
".",
"__locked",
":",
"raise",
"RuntimeError",
"(",
"\"c... | https://github.com/irmen/synthesizer/blob/9ee0969cb9756d63dffb3b4be2ff35e3ccfadbf9/synthplayer/sample.py#L779-L797 | |
aiortc/aioquic | 9e360a57d6b8cf78844fba814d9968cd5f602648 | src/aioquic/quic/connection.py | python | QuicConnection._handle_max_data_frame | (
self, context: QuicReceiveContext, frame_type: int, buf: Buffer
) | Handle a MAX_DATA frame.
This adjusts the total amount of we can send to the peer. | Handle a MAX_DATA frame. | [
"Handle",
"a",
"MAX_DATA",
"frame",
"."
] | def _handle_max_data_frame(
self, context: QuicReceiveContext, frame_type: int, buf: Buffer
) -> None:
"""
Handle a MAX_DATA frame.
This adjusts the total amount of we can send to the peer.
"""
max_data = buf.pull_uint_var()
# log frame
if self._quic... | [
"def",
"_handle_max_data_frame",
"(",
"self",
",",
"context",
":",
"QuicReceiveContext",
",",
"frame_type",
":",
"int",
",",
"buf",
":",
"Buffer",
")",
"->",
"None",
":",
"max_data",
"=",
"buf",
".",
"pull_uint_var",
"(",
")",
"# log frame",
"if",
"self",
... | https://github.com/aiortc/aioquic/blob/9e360a57d6b8cf78844fba814d9968cd5f602648/src/aioquic/quic/connection.py#L1623-L1643 | ||
masa-su/pixyz | a9baf067730035d03351476f5c2e1e43016808ce | pixyz/losses/losses.py | python | Loss.__init__ | (self, input_var=None) | Parameters
----------
input_var : :obj:`list` of :obj:`str`, defaults to None
Input variables of this loss function.
In general, users do not need to set them explicitly
because these depend on the given distributions and each loss function. | Parameters
----------
input_var : :obj:`list` of :obj:`str`, defaults to None
Input variables of this loss function.
In general, users do not need to set them explicitly
because these depend on the given distributions and each loss function. | [
"Parameters",
"----------",
"input_var",
":",
":",
"obj",
":",
"list",
"of",
":",
"obj",
":",
"str",
"defaults",
"to",
"None",
"Input",
"variables",
"of",
"this",
"loss",
"function",
".",
"In",
"general",
"users",
"do",
"not",
"need",
"to",
"set",
"them"... | def __init__(self, input_var=None):
"""
Parameters
----------
input_var : :obj:`list` of :obj:`str`, defaults to None
Input variables of this loss function.
In general, users do not need to set them explicitly
because these depend on the given distribu... | [
"def",
"__init__",
"(",
"self",
",",
"input_var",
"=",
"None",
")",
":",
"super",
"(",
")",
".",
"__init__",
"(",
")",
"self",
".",
"_input_var",
"=",
"deepcopy",
"(",
"input_var",
")"
] | https://github.com/masa-su/pixyz/blob/a9baf067730035d03351476f5c2e1e43016808ce/pixyz/losses/losses.py#L58-L69 | ||
1012598167/flask_mongodb_game | 60c7e0351586656ec38f851592886338e50b4110 | python_flask/venv/Lib/site-packages/pip-19.0.3-py3.6.egg/pip/_vendor/distlib/markers.py | python | Evaluator.evaluate | (self, expr, context) | return result | Evaluate a marker expression returned by the :func:`parse_requirement`
function in the specified context. | Evaluate a marker expression returned by the :func:`parse_requirement`
function in the specified context. | [
"Evaluate",
"a",
"marker",
"expression",
"returned",
"by",
"the",
":",
"func",
":",
"parse_requirement",
"function",
"in",
"the",
"specified",
"context",
"."
] | def evaluate(self, expr, context):
"""
Evaluate a marker expression returned by the :func:`parse_requirement`
function in the specified context.
"""
if isinstance(expr, string_types):
if expr[0] in '\'"':
result = expr[1:-1]
else:
... | [
"def",
"evaluate",
"(",
"self",
",",
"expr",
",",
"context",
")",
":",
"if",
"isinstance",
"(",
"expr",
",",
"string_types",
")",
":",
"if",
"expr",
"[",
"0",
"]",
"in",
"'\\'\"'",
":",
"result",
"=",
"expr",
"[",
"1",
":",
"-",
"1",
"]",
"else",... | https://github.com/1012598167/flask_mongodb_game/blob/60c7e0351586656ec38f851592886338e50b4110/python_flask/venv/Lib/site-packages/pip-19.0.3-py3.6.egg/pip/_vendor/distlib/markers.py#L50-L75 | |
AIChallenger/AI_Challenger_2018 | f0e4376152c8fe5a098ed92a973cec96b13e1a24 | Baselines/autonomous_driving_perception208_baseline/detection/utils/np_box_list_ops.py | python | gather | (boxlist, indices, fields=None) | return subboxlist | Gather boxes from BoxList according to indices and return new BoxList.
By default, gather returns boxes corresponding to the input index list, as
well as all additional fields stored in the boxlist (indexing into the
first dimension). However one can optionally only gather from a
subset of fields.
Args:
... | Gather boxes from BoxList according to indices and return new BoxList. | [
"Gather",
"boxes",
"from",
"BoxList",
"according",
"to",
"indices",
"and",
"return",
"new",
"BoxList",
"."
] | def gather(boxlist, indices, fields=None):
"""Gather boxes from BoxList according to indices and return new BoxList.
By default, gather returns boxes corresponding to the input index list, as
well as all additional fields stored in the boxlist (indexing into the
first dimension). However one can optionally on... | [
"def",
"gather",
"(",
"boxlist",
",",
"indices",
",",
"fields",
"=",
"None",
")",
":",
"if",
"indices",
".",
"size",
":",
"if",
"np",
".",
"amax",
"(",
"indices",
")",
">=",
"boxlist",
".",
"num_boxes",
"(",
")",
"or",
"np",
".",
"amin",
"(",
"in... | https://github.com/AIChallenger/AI_Challenger_2018/blob/f0e4376152c8fe5a098ed92a973cec96b13e1a24/Baselines/autonomous_driving_perception208_baseline/detection/utils/np_box_list_ops.py#L95-L127 | |
deepset-ai/haystack | 79fdda8a7cf393d774803608a4874f2a6e63cf6f | haystack/nodes/retriever/dense.py | python | TableTextRetriever.load | (cls,
load_dir: Union[Path, str],
document_store: BaseDocumentStore,
max_seq_len_query: int = 64,
max_seq_len_passage: int = 256,
max_seq_len_table: int = 256,
use_gpu: bool = True,
batch_size: int = 16,
embed_meta_f... | return mm_retriever | Load TableTextRetriever from the specified directory. | Load TableTextRetriever from the specified directory. | [
"Load",
"TableTextRetriever",
"from",
"the",
"specified",
"directory",
"."
] | def load(cls,
load_dir: Union[Path, str],
document_store: BaseDocumentStore,
max_seq_len_query: int = 64,
max_seq_len_passage: int = 256,
max_seq_len_table: int = 256,
use_gpu: bool = True,
batch_size: int = 16,
embe... | [
"def",
"load",
"(",
"cls",
",",
"load_dir",
":",
"Union",
"[",
"Path",
",",
"str",
"]",
",",
"document_store",
":",
"BaseDocumentStore",
",",
"max_seq_len_query",
":",
"int",
"=",
"64",
",",
"max_seq_len_passage",
":",
"int",
"=",
"256",
",",
"max_seq_len_... | https://github.com/deepset-ai/haystack/blob/79fdda8a7cf393d774803608a4874f2a6e63cf6f/haystack/nodes/retriever/dense.py#L921-L959 | |
spectacles/CodeComplice | 8ca8ee4236f72b58caa4209d2fbd5fa56bd31d62 | libs/codeintel2/langintel.py | python | LangIntel.cb_function_detail_from_elem | (self, elem) | [] | def cb_function_detail_from_elem(self, elem):
# by default (some languages may choose to override)
sig = elem.get("signature")
if sig:
return sig
else:
return elem.get("name")+"(...)" | [
"def",
"cb_function_detail_from_elem",
"(",
"self",
",",
"elem",
")",
":",
"# by default (some languages may choose to override)",
"sig",
"=",
"elem",
".",
"get",
"(",
"\"signature\"",
")",
"if",
"sig",
":",
"return",
"sig",
"else",
":",
"return",
"elem",
".",
"... | https://github.com/spectacles/CodeComplice/blob/8ca8ee4236f72b58caa4209d2fbd5fa56bd31d62/libs/codeintel2/langintel.py#L153-L159 | ||||
openhatch/oh-mainline | ce29352a034e1223141dcc2f317030bbc3359a51 | vendor/packages/celery/celery/concurrency/processes/pool.py | python | Pool.apply_async | (self, func, args=(), kwds={},
callback=None, accept_callback=None, timeout_callback=None,
waitforslot=False, error_callback=None,
soft_timeout=None, timeout=None) | return result | Asynchronous equivalent of `apply()` builtin.
Callback is called when the functions return value is ready.
The accept callback is called when the job is accepted to be executed.
Simplified the flow is like this:
>>> if accept_callback:
... accept_callback()
... | Asynchronous equivalent of `apply()` builtin. | [
"Asynchronous",
"equivalent",
"of",
"apply",
"()",
"builtin",
"."
] | def apply_async(self, func, args=(), kwds={},
callback=None, accept_callback=None, timeout_callback=None,
waitforslot=False, error_callback=None,
soft_timeout=None, timeout=None):
'''
Asynchronous equivalent of `apply()` builtin.
Callback is called when the f... | [
"def",
"apply_async",
"(",
"self",
",",
"func",
",",
"args",
"=",
"(",
")",
",",
"kwds",
"=",
"{",
"}",
",",
"callback",
"=",
"None",
",",
"accept_callback",
"=",
"None",
",",
"timeout_callback",
"=",
"None",
",",
"waitforslot",
"=",
"False",
",",
"e... | https://github.com/openhatch/oh-mainline/blob/ce29352a034e1223141dcc2f317030bbc3359a51/vendor/packages/celery/celery/concurrency/processes/pool.py#L770-L806 | |
home-assistant/core | 265ebd17a3f17ed8dc1e9bdede03ac8e323f1ab1 | homeassistant/components/mpchc/media_player.py | python | MpcHcDevice.media_play | (self) | Send play command. | Send play command. | [
"Send",
"play",
"command",
"."
] | def media_play(self):
"""Send play command."""
self._send_command(887) | [
"def",
"media_play",
"(",
"self",
")",
":",
"self",
".",
"_send_command",
"(",
"887",
")"
] | https://github.com/home-assistant/core/blob/265ebd17a3f17ed8dc1e9bdede03ac8e323f1ab1/homeassistant/components/mpchc/media_player.py#L181-L183 | ||
apache/tvm | 6eb4ed813ebcdcd9558f0906a1870db8302ff1e0 | python/tvm/relay/op/contrib/cutlass.py | python | partition_for_cutlass | (mod, params=None) | return seq(mod) | Partition the input module into CUTLASS-supported subgraphs. | Partition the input module into CUTLASS-supported subgraphs. | [
"Partition",
"the",
"input",
"module",
"into",
"CUTLASS",
"-",
"supported",
"subgraphs",
"."
] | def partition_for_cutlass(mod, params=None):
"""Partition the input module into CUTLASS-supported subgraphs."""
dense_pat = ("cutlass.dense", make_gemm_pattern(False, None), check_gemm)
dense_bias_pat = ("cutlass.dense_bias", make_gemm_pattern(True, None), check_gemm)
dense_bias_relu_pat = ("cutlass.den... | [
"def",
"partition_for_cutlass",
"(",
"mod",
",",
"params",
"=",
"None",
")",
":",
"dense_pat",
"=",
"(",
"\"cutlass.dense\"",
",",
"make_gemm_pattern",
"(",
"False",
",",
"None",
")",
",",
"check_gemm",
")",
"dense_bias_pat",
"=",
"(",
"\"cutlass.dense_bias\"",
... | https://github.com/apache/tvm/blob/6eb4ed813ebcdcd9558f0906a1870db8302ff1e0/python/tvm/relay/op/contrib/cutlass.py#L178-L265 | |
blawar/nut | 2cf351400418399a70164987e28670309f6c9cb5 | Fs/driver/curl.py | python | FileContext.read | (self, sz=None) | return output.getvalue() | [] | def read(self, sz=None):
curl = pycurl.Curl()
curl.setopt(pycurl.URL, self.url)
curl.setopt(pycurl.CAINFO, certifi.where())
output = io.BytesIO()
curl.setopt(pycurl.WRITEFUNCTION, output.write)
self.setup(curl, None, sz)
curl.perform()
return output.getvalue() | [
"def",
"read",
"(",
"self",
",",
"sz",
"=",
"None",
")",
":",
"curl",
"=",
"pycurl",
".",
"Curl",
"(",
")",
"curl",
".",
"setopt",
"(",
"pycurl",
".",
"URL",
",",
"self",
".",
"url",
")",
"curl",
".",
"setopt",
"(",
"pycurl",
".",
"CAINFO",
","... | https://github.com/blawar/nut/blob/2cf351400418399a70164987e28670309f6c9cb5/Fs/driver/curl.py#L31-L40 | |||
sonyisme/keras-recommendation | e84fab6706faf8405f4bfc75cf50ecff6433eb77 | keras-master/keras/initializations.py | python | orthogonal | (shape, scale=1.1) | return sharedX(scale * q[:shape[0], :shape[1]]) | From Lasagne | From Lasagne | [
"From",
"Lasagne"
] | def orthogonal(shape, scale=1.1):
''' From Lasagne
'''
flat_shape = (shape[0], np.prod(shape[1:]))
a = np.random.normal(0.0, 1.0, flat_shape)
u, _, v = np.linalg.svd(a, full_matrices=False)
q = u if u.shape == flat_shape else v # pick the one with the correct shape
q = q.reshape(shape)
r... | [
"def",
"orthogonal",
"(",
"shape",
",",
"scale",
"=",
"1.1",
")",
":",
"flat_shape",
"=",
"(",
"shape",
"[",
"0",
"]",
",",
"np",
".",
"prod",
"(",
"shape",
"[",
"1",
":",
"]",
")",
")",
"a",
"=",
"np",
".",
"random",
".",
"normal",
"(",
"0.0... | https://github.com/sonyisme/keras-recommendation/blob/e84fab6706faf8405f4bfc75cf50ecff6433eb77/keras-master/keras/initializations.py#L52-L60 | |
mdiazcl/fuzzbunch-debian | 2b76c2249ade83a389ae3badb12a1bd09901fd2c | windows/Resources/Python/Core/Lib/tarfile.py | python | TarIter.next | (self) | return tarinfo | Return the next item using TarFile's next() method.
When all members have been read, set TarFile as _loaded. | Return the next item using TarFile's next() method.
When all members have been read, set TarFile as _loaded. | [
"Return",
"the",
"next",
"item",
"using",
"TarFile",
"s",
"next",
"()",
"method",
".",
"When",
"all",
"members",
"have",
"been",
"read",
"set",
"TarFile",
"as",
"_loaded",
"."
] | def next(self):
"""Return the next item using TarFile's next() method.
When all members have been read, set TarFile as _loaded.
"""
if not self.tarfile._loaded:
tarinfo = self.tarfile.next()
if not tarinfo:
self.tarfile._loaded = True
... | [
"def",
"next",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"tarfile",
".",
"_loaded",
":",
"tarinfo",
"=",
"self",
".",
"tarfile",
".",
"next",
"(",
")",
"if",
"not",
"tarinfo",
":",
"self",
".",
"tarfile",
".",
"_loaded",
"=",
"True",
"raise"... | https://github.com/mdiazcl/fuzzbunch-debian/blob/2b76c2249ade83a389ae3badb12a1bd09901fd2c/windows/Resources/Python/Core/Lib/tarfile.py#L2171-L2187 | |
hitchtest/hitch | 7ccfb878566d51259f448fab7d7ae5bdb92025bf | hitch/commandline.py | python | upgrade | () | Upgrade all installed hitch packages. | Upgrade all installed hitch packages. | [
"Upgrade",
"all",
"installed",
"hitch",
"packages",
"."
] | def upgrade():
"""Upgrade all installed hitch packages."""
hitchdir.check_hitch_directory_integrity()
update_requirements()
pip = get_pip()
package_list = [
p for p in check_output([pip, "freeze"]).decode('utf8').split('\n')
if p != "" and "==" in p
]
version_fixed_pack... | [
"def",
"upgrade",
"(",
")",
":",
"hitchdir",
".",
"check_hitch_directory_integrity",
"(",
")",
"update_requirements",
"(",
")",
"pip",
"=",
"get_pip",
"(",
")",
"package_list",
"=",
"[",
"p",
"for",
"p",
"in",
"check_output",
"(",
"[",
"pip",
",",
"\"freez... | https://github.com/hitchtest/hitch/blob/7ccfb878566d51259f448fab7d7ae5bdb92025bf/hitch/commandline.py#L222-L243 | ||
Nicotine-Plus/nicotine-plus | 6583532193e132206bb2096c77c6ad1ce96c21fa | pynicotine/pynicotine.py | python | NicotineCore.room_ticker_remove | (self, msg) | Server code: 115 | Server code: 115 | [
"Server",
"code",
":",
"115"
] | def room_ticker_remove(self, msg):
""" Server code: 115 """
log.add_msg_contents(msg)
self.chatrooms.ticker_remove(msg) | [
"def",
"room_ticker_remove",
"(",
"self",
",",
"msg",
")",
":",
"log",
".",
"add_msg_contents",
"(",
"msg",
")",
"self",
".",
"chatrooms",
".",
"ticker_remove",
"(",
"msg",
")"
] | https://github.com/Nicotine-Plus/nicotine-plus/blob/6583532193e132206bb2096c77c6ad1ce96c21fa/pynicotine/pynicotine.py#L836-L840 | ||
dagster-io/dagster | b27d569d5fcf1072543533a0c763815d96f90b8f | python_modules/dagster/dagster/core/storage/event_log/sql_event_log.py | python | SqlEventLogStorage.update_event_log_record | (self, record_id, event) | Utility method for migration scripts to update SQL representation of event records. | Utility method for migration scripts to update SQL representation of event records. | [
"Utility",
"method",
"for",
"migration",
"scripts",
"to",
"update",
"SQL",
"representation",
"of",
"event",
"records",
"."
] | def update_event_log_record(self, record_id, event):
"""Utility method for migration scripts to update SQL representation of event records."""
check.int_param(record_id, "record_id")
check.inst_param(event, "event", EventLogEntry)
dagster_event_type = None
asset_key_str = None
... | [
"def",
"update_event_log_record",
"(",
"self",
",",
"record_id",
",",
"event",
")",
":",
"check",
".",
"int_param",
"(",
"record_id",
",",
"\"record_id\"",
")",
"check",
".",
"inst_param",
"(",
"event",
",",
"\"event\"",
",",
"EventLogEntry",
")",
"dagster_eve... | https://github.com/dagster-io/dagster/blob/b27d569d5fcf1072543533a0c763815d96f90b8f/python_modules/dagster/dagster/core/storage/event_log/sql_event_log.py#L457-L480 | ||
beeware/ouroboros | a29123c6fab6a807caffbb7587cf548e0c370296 | ouroboros/turtle.py | python | __methodDict | (cls, _dict) | helper function for Scrolled Canvas | helper function for Scrolled Canvas | [
"helper",
"function",
"for",
"Scrolled",
"Canvas"
] | def __methodDict(cls, _dict):
"""helper function for Scrolled Canvas"""
baseList = list(cls.__bases__)
baseList.reverse()
for _super in baseList:
__methodDict(_super, _dict)
for key, value in cls.__dict__.items():
if type(value) == types.FunctionType:
_dict[key] = value | [
"def",
"__methodDict",
"(",
"cls",
",",
"_dict",
")",
":",
"baseList",
"=",
"list",
"(",
"cls",
".",
"__bases__",
")",
"baseList",
".",
"reverse",
"(",
")",
"for",
"_super",
"in",
"baseList",
":",
"__methodDict",
"(",
"_super",
",",
"_dict",
")",
"for"... | https://github.com/beeware/ouroboros/blob/a29123c6fab6a807caffbb7587cf548e0c370296/ouroboros/turtle.py#L288-L296 | ||
evennia/evennia | fa79110ba6b219932f22297838e8ac72ebc0be0e | evennia/server/portal/amp_server.py | python | AMPServerProtocol.data_to_server | (self, command, sessid, **kwargs) | Send data across the wire to the Server.
Args:
command (AMP Command): A protocol send command.
sessid (int): A unique Session id.
kwargs (any): Data to send. This will be pickled.
Returns:
deferred (deferred or None): A deferred with an errback.
... | Send data across the wire to the Server. | [
"Send",
"data",
"across",
"the",
"wire",
"to",
"the",
"Server",
"."
] | def data_to_server(self, command, sessid, **kwargs):
"""
Send data across the wire to the Server.
Args:
command (AMP Command): A protocol send command.
sessid (int): A unique Session id.
kwargs (any): Data to send. This will be pickled.
Returns:
... | [
"def",
"data_to_server",
"(",
"self",
",",
"command",
",",
"sessid",
",",
"*",
"*",
"kwargs",
")",
":",
"# print(\"portal data_to_server: {}, {}, {}\".format(command, sessid, kwargs))",
"if",
"self",
".",
"factory",
".",
"server_connection",
":",
"return",
"self",
"."... | https://github.com/evennia/evennia/blob/fa79110ba6b219932f22297838e8ac72ebc0be0e/evennia/server/portal/amp_server.py#L126-L150 | ||
jantman/awslimitchecker | 411ad9e734ddb16d87720ff5b994f19f47b8b098 | awslimitchecker/services/vpc.py | python | _VpcService._find_usage_network_interfaces | (self) | find usage of network interfaces | find usage of network interfaces | [
"find",
"usage",
"of",
"network",
"interfaces"
] | def _find_usage_network_interfaces(self):
"""find usage of network interfaces"""
enis = paginate_dict(
self.conn.describe_network_interfaces,
alc_marker_path=['NextToken'],
alc_data_path=['NetworkInterfaces'],
alc_marker_param='NextToken',
Filt... | [
"def",
"_find_usage_network_interfaces",
"(",
"self",
")",
":",
"enis",
"=",
"paginate_dict",
"(",
"self",
".",
"conn",
".",
"describe_network_interfaces",
",",
"alc_marker_path",
"=",
"[",
"'NextToken'",
"]",
",",
"alc_data_path",
"=",
"[",
"'NetworkInterfaces'",
... | https://github.com/jantman/awslimitchecker/blob/411ad9e734ddb16d87720ff5b994f19f47b8b098/awslimitchecker/services/vpc.py#L238-L251 | ||
rockstor/rockstor-core | 81a0d5f5e0a6dfe5a922199828f66eeab0253e65 | src/rockstor/system/ssh.py | python | update_sftp_config | (input_map) | Fetch sftp-related customization settings from database
and writes them to SSHD_CONFIG.
:param input_map: dictionary of user,directory pairs.
:return: | Fetch sftp-related customization settings from database
and writes them to SSHD_CONFIG.
:param input_map: dictionary of user,directory pairs.
:return: | [
"Fetch",
"sftp",
"-",
"related",
"customization",
"settings",
"from",
"database",
"and",
"writes",
"them",
"to",
"SSHD_CONFIG",
".",
":",
"param",
"input_map",
":",
"dictionary",
"of",
"user",
"directory",
"pairs",
".",
":",
"return",
":"
] | def update_sftp_config(input_map):
"""
Fetch sftp-related customization settings from database
and writes them to SSHD_CONFIG.
:param input_map: dictionary of user,directory pairs.
:return:
"""
fo, npath = mkstemp()
userstr = "AllowUsers root {}".format(" ".join(input_map.keys()))
wi... | [
"def",
"update_sftp_config",
"(",
"input_map",
")",
":",
"fo",
",",
"npath",
"=",
"mkstemp",
"(",
")",
"userstr",
"=",
"\"AllowUsers root {}\"",
".",
"format",
"(",
"\" \"",
".",
"join",
"(",
"input_map",
".",
"keys",
"(",
")",
")",
")",
"with",
"open",
... | https://github.com/rockstor/rockstor-core/blob/81a0d5f5e0a6dfe5a922199828f66eeab0253e65/src/rockstor/system/ssh.py#L36-L71 | ||
danecjensen/subscribely | 4d6ac60358b5fe26f0c01be68f1ba063df3b1ea0 | src/httplib2/__init__.py | python | Http.__init__ | (self, cache=None, timeout=None,
proxy_info=ProxyInfo.from_environment,
ca_certs=None, disable_ssl_certificate_validation=False) | If 'cache' is a string then it is used as a directory name for
a disk cache. Otherwise it must be an object that supports the
same interface as FileCache.
All timeouts are in seconds. If None is passed for timeout
then Python's default timeout for sockets will be used. See
for e... | If 'cache' is a string then it is used as a directory name for
a disk cache. Otherwise it must be an object that supports the
same interface as FileCache. | [
"If",
"cache",
"is",
"a",
"string",
"then",
"it",
"is",
"used",
"as",
"a",
"directory",
"name",
"for",
"a",
"disk",
"cache",
".",
"Otherwise",
"it",
"must",
"be",
"an",
"object",
"that",
"supports",
"the",
"same",
"interface",
"as",
"FileCache",
"."
] | def __init__(self, cache=None, timeout=None,
proxy_info=ProxyInfo.from_environment,
ca_certs=None, disable_ssl_certificate_validation=False):
"""
If 'cache' is a string then it is used as a directory name for
a disk cache. Otherwise it must be an object that sup... | [
"def",
"__init__",
"(",
"self",
",",
"cache",
"=",
"None",
",",
"timeout",
"=",
"None",
",",
"proxy_info",
"=",
"ProxyInfo",
".",
"from_environment",
",",
"ca_certs",
"=",
"None",
",",
"disable_ssl_certificate_validation",
"=",
"False",
")",
":",
"self",
"."... | https://github.com/danecjensen/subscribely/blob/4d6ac60358b5fe26f0c01be68f1ba063df3b1ea0/src/httplib2/__init__.py#L1133-L1198 | ||
neulab/xnmt | d93f8f3710f986f36eb54e9ff3976a6b683da2a4 | xnmt/transducers/base.py | python | SeqTransducer.transduce | (self, seq: 'expression_seqs.ExpressionSequence') | Parameters should be :class:`expression_seqs.ExpressionSequence` objects wherever appropriate
Args:
seq: An expression sequence representing the input to the transduction
Returns:
result of transduction, an expression sequence | Parameters should be :class:`expression_seqs.ExpressionSequence` objects wherever appropriate | [
"Parameters",
"should",
"be",
":",
"class",
":",
"expression_seqs",
".",
"ExpressionSequence",
"objects",
"wherever",
"appropriate"
] | def transduce(self, seq: 'expression_seqs.ExpressionSequence') -> 'expression_seqs.ExpressionSequence':
"""
Parameters should be :class:`expression_seqs.ExpressionSequence` objects wherever appropriate
Args:
seq: An expression sequence representing the input to the transduction
Returns:
re... | [
"def",
"transduce",
"(",
"self",
",",
"seq",
":",
"'expression_seqs.ExpressionSequence'",
")",
"->",
"'expression_seqs.ExpressionSequence'",
":",
"raise",
"NotImplementedError",
"(",
"\"SeqTransducer.transduce() must be implemented by SeqTransducer sub-classes\"",
")"
] | https://github.com/neulab/xnmt/blob/d93f8f3710f986f36eb54e9ff3976a6b683da2a4/xnmt/transducers/base.py#L46-L56 | ||
hyperledger/fabric-sdk-py | 8ee33a8981887e37950dc0f36a7ec63b3a5ba5c3 | hfc/fabric/channel/instantiation.py | python | Instantiation.handle | (self, tran_prop_req, scheduler=None) | return _instantiate_chaincode(self._chain, tran_prop_req, scheduler) | Execute chaincode instantiation transaction proposal request.
:param tran_prop_req: chaincode instantiation transaction proposal request
:param scheduler: see rx.Scheduler, defaults to None
:return: An rx.Observer wrapper of chaincode instantiation response | Execute chaincode instantiation transaction proposal request. | [
"Execute",
"chaincode",
"instantiation",
"transaction",
"proposal",
"request",
"."
] | def handle(self, tran_prop_req, scheduler=None):
"""Execute chaincode instantiation transaction proposal request.
:param tran_prop_req: chaincode instantiation transaction proposal request
:param scheduler: see rx.Scheduler, defaults to None
:return: An rx.Observer wrapper of chaincode ... | [
"def",
"handle",
"(",
"self",
",",
"tran_prop_req",
",",
"scheduler",
"=",
"None",
")",
":",
"return",
"_instantiate_chaincode",
"(",
"self",
".",
"_chain",
",",
"tran_prop_req",
",",
"scheduler",
")"
] | https://github.com/hyperledger/fabric-sdk-py/blob/8ee33a8981887e37950dc0f36a7ec63b3a5ba5c3/hfc/fabric/channel/instantiation.py#L36-L43 | |
hubblestack/hubble | 763142474edcecdec5fd25591dc29c3536e8f969 | hubblestack/audit/readfile.py | python | validate_params | (block_id, block_dict, extra_args=None) | Validate all mandatory params required for this module
:param block_id:
id of the block
:param block_dict:
parameter for this module
:param extra_args:
Extra argument dictionary, (If any)
Example: {'chaining_args': {'result': '/some/path', 'status': True},
... | Validate all mandatory params required for this module | [
"Validate",
"all",
"mandatory",
"params",
"required",
"for",
"this",
"module"
] | def validate_params(block_id, block_dict, extra_args=None):
"""
Validate all mandatory params required for this module
:param block_id:
id of the block
:param block_dict:
parameter for this module
:param extra_args:
Extra argument dictionary, (If any)
Example: {'chai... | [
"def",
"validate_params",
"(",
"block_id",
",",
"block_dict",
",",
"extra_args",
"=",
"None",
")",
":",
"log",
".",
"debug",
"(",
"'Module: readfile Start validating params for check-id: {0}'",
".",
"format",
"(",
"block_id",
")",
")",
"filepath",
"=",
"runner_utils... | https://github.com/hubblestack/hubble/blob/763142474edcecdec5fd25591dc29c3536e8f969/hubblestack/audit/readfile.py#L238-L268 | ||
sagemath/sage | f9b2db94f675ff16963ccdefba4f1a3393b3fe0d | src/sage/parallel/parallelism.py | python | Parallelism.set | (self, field=None, nproc=None) | r"""
Set the number of processes to be launched for parallel computations
regarding some specific field.
INPUT:
- ``field`` -- (default: ``None``) string specifying the computational
field for which the number of parallel processes is to be set; if
``None``, all fie... | r"""
Set the number of processes to be launched for parallel computations
regarding some specific field. | [
"r",
"Set",
"the",
"number",
"of",
"processes",
"to",
"be",
"launched",
"for",
"parallel",
"computations",
"regarding",
"some",
"specific",
"field",
"."
] | def set(self, field=None, nproc=None):
r"""
Set the number of processes to be launched for parallel computations
regarding some specific field.
INPUT:
- ``field`` -- (default: ``None``) string specifying the computational
field for which the number of parallel process... | [
"def",
"set",
"(",
"self",
",",
"field",
"=",
"None",
",",
"nproc",
"=",
"None",
")",
":",
"if",
"field",
"is",
"None",
":",
"for",
"fi",
"in",
"self",
".",
"_nproc",
":",
"self",
".",
"set",
"(",
"field",
"=",
"fi",
",",
"nproc",
"=",
"nproc",... | https://github.com/sagemath/sage/blob/f9b2db94f675ff16963ccdefba4f1a3393b3fe0d/src/sage/parallel/parallelism.py#L170-L247 | ||
HealthCatalyst/healthcareai-py | cb82b94990fb3046edccb3740ae5653adce70940 | healthcareai/common/cardinality_checks.py | python | calculate_cardinality | (dataframe) | return results | Find cardinality of columns in a dataframe.
This function counts the number of rows in the dataframe, counts the unique
values in each column and sorts by the ratio of unique values relative to
the number of rows.
This is useful for profiling training data.
Args:
dataframe (pandas.core.fr... | Find cardinality of columns in a dataframe. | [
"Find",
"cardinality",
"of",
"columns",
"in",
"a",
"dataframe",
"."
] | def calculate_cardinality(dataframe):
"""
Find cardinality of columns in a dataframe.
This function counts the number of rows in the dataframe, counts the unique
values in each column and sorts by the ratio of unique values relative to
the number of rows.
This is useful for profiling training ... | [
"def",
"calculate_cardinality",
"(",
"dataframe",
")",
":",
"record_count",
"=",
"len",
"(",
"dataframe",
")",
"result_list",
"=",
"[",
"]",
"for",
"column",
"in",
"dataframe",
":",
"count",
"=",
"len",
"(",
"dataframe",
"[",
"column",
"]",
".",
"unique",
... | https://github.com/HealthCatalyst/healthcareai-py/blob/cb82b94990fb3046edccb3740ae5653adce70940/healthcareai/common/cardinality_checks.py#L9-L40 | |
pysmt/pysmt | ade4dc2a825727615033a96d31c71e9f53ce4764 | pysmt/solvers/msat.py | python | MathSATOptions.__init__ | (self, **base_options) | [] | def __init__(self, **base_options):
SolverOptions.__init__(self, **base_options) | [
"def",
"__init__",
"(",
"self",
",",
"*",
"*",
"base_options",
")",
":",
"SolverOptions",
".",
"__init__",
"(",
"self",
",",
"*",
"*",
"base_options",
")"
] | https://github.com/pysmt/pysmt/blob/ade4dc2a825727615033a96d31c71e9f53ce4764/pysmt/solvers/msat.py#L139-L140 | ||||
rcorcs/NatI | fdf014f4292afdc95250add7b6658468043228e1 | en/wordnet/wntools.py | python | hypernyms | (source) | return closure(source, HYPERNYM) | Return source and its hypernyms. If source is a Word, return
the union of the hypernyms of its senses. | Return source and its hypernyms. If source is a Word, return
the union of the hypernyms of its senses. | [
"Return",
"source",
"and",
"its",
"hypernyms",
".",
"If",
"source",
"is",
"a",
"Word",
"return",
"the",
"union",
"of",
"the",
"hypernyms",
"of",
"its",
"senses",
"."
] | def hypernyms(source):
"""Return source and its hypernyms. If source is a Word, return
the union of the hypernyms of its senses."""
return closure(source, HYPERNYM) | [
"def",
"hypernyms",
"(",
"source",
")",
":",
"return",
"closure",
"(",
"source",
",",
"HYPERNYM",
")"
] | https://github.com/rcorcs/NatI/blob/fdf014f4292afdc95250add7b6658468043228e1/en/wordnet/wntools.py#L96-L100 | |
replit-archive/empythoned | 977ec10ced29a3541a4973dc2b59910805695752 | cpython/Tools/gdb/libpython.py | python | PyObjectPtr.is_optimized_out | (self) | return self._gdbval.is_optimized_out | Is the value of the underlying PyObject* visible to the debugger?
This can vary with the precise version of the compiler used to build
Python, and the precise version of gdb.
See e.g. https://bugzilla.redhat.com/show_bug.cgi?id=556975 with
PyEval_EvalFrameEx's "f" | Is the value of the underlying PyObject* visible to the debugger? | [
"Is",
"the",
"value",
"of",
"the",
"underlying",
"PyObject",
"*",
"visible",
"to",
"the",
"debugger?"
] | def is_optimized_out(self):
'''
Is the value of the underlying PyObject* visible to the debugger?
This can vary with the precise version of the compiler used to build
Python, and the precise version of gdb.
See e.g. https://bugzilla.redhat.com/show_bug.cgi?id=556975 with
... | [
"def",
"is_optimized_out",
"(",
"self",
")",
":",
"return",
"self",
".",
"_gdbval",
".",
"is_optimized_out"
] | https://github.com/replit-archive/empythoned/blob/977ec10ced29a3541a4973dc2b59910805695752/cpython/Tools/gdb/libpython.py#L198-L208 | |
CvvT/dumpDex | 92ab3b7e996194a06bf1dd5538a4954e8a5ee9c1 | python/idc.py | python | GetStringType | (ea) | Get string type
@param ea: linear address
@return: One of ASCSTR_... constants | Get string type | [
"Get",
"string",
"type"
] | def GetStringType(ea):
"""
Get string type
@param ea: linear address
@return: One of ASCSTR_... constants
"""
ti = idaapi.opinfo_t()
if idaapi.get_opinfo(ea, 0, GetFlags(ea), ti):
return ti.strtype
else:
return None | [
"def",
"GetStringType",
"(",
"ea",
")",
":",
"ti",
"=",
"idaapi",
".",
"opinfo_t",
"(",
")",
"if",
"idaapi",
".",
"get_opinfo",
"(",
"ea",
",",
"0",
",",
"GetFlags",
"(",
"ea",
")",
",",
"ti",
")",
":",
"return",
"ti",
".",
"strtype",
"else",
":"... | https://github.com/CvvT/dumpDex/blob/92ab3b7e996194a06bf1dd5538a4954e8a5ee9c1/python/idc.py#L2448-L2461 | ||
harvey1673/pyktrader | 9e4f9211cb716786d443f3800010d901bc261610 | agent.py | python | Agent.__init__ | (self, name, trader, cuser, instruments, strategies = [], tday=datetime.date.today(), config = {}) | trader为交易对象
tday为当前日,为0则为当日 | trader为交易对象
tday为当前日,为0则为当日 | [
"trader为交易对象",
"tday为当前日",
"为0则为当日"
] | def __init__(self, name, trader, cuser, instruments, strategies = [], tday=datetime.date.today(), config = {}):
'''
trader为交易对象
tday为当前日,为0则为当日
'''
self.tick_id = 0
self.timer_count = 0
self.request_id = 1
folder = 'C:\\dev\\src\\ktlib\\pythonctp\\... | [
"def",
"__init__",
"(",
"self",
",",
"name",
",",
"trader",
",",
"cuser",
",",
"instruments",
",",
"strategies",
"=",
"[",
"]",
",",
"tday",
"=",
"datetime",
".",
"date",
".",
"today",
"(",
")",
",",
"config",
"=",
"{",
"}",
")",
":",
"self",
"."... | https://github.com/harvey1673/pyktrader/blob/9e4f9211cb716786d443f3800010d901bc261610/agent.py#L651-L736 | ||
CLUEbenchmark/CLUE | 5bd39732734afecb490cf18a5212e692dbf2c007 | baselines/models/bert_wwm_ext/tokenization.py | python | _is_control | (char) | return False | Checks whether `chars` is a control character. | Checks whether `chars` is a control character. | [
"Checks",
"whether",
"chars",
"is",
"a",
"control",
"character",
"."
] | def _is_control(char):
"""Checks whether `chars` is a control character."""
# These are technically control characters but we count them as whitespace
# characters.
if char == "\t" or char == "\n" or char == "\r":
return False
cat = unicodedata.category(char)
if cat in ("Cc", "Cf"):
return True
re... | [
"def",
"_is_control",
"(",
"char",
")",
":",
"# These are technically control characters but we count them as whitespace",
"# characters.",
"if",
"char",
"==",
"\"\\t\"",
"or",
"char",
"==",
"\"\\n\"",
"or",
"char",
"==",
"\"\\r\"",
":",
"return",
"False",
"cat",
"=",... | https://github.com/CLUEbenchmark/CLUE/blob/5bd39732734afecb490cf18a5212e692dbf2c007/baselines/models/bert_wwm_ext/tokenization.py#L374-L383 | |
cuthbertLab/music21 | bd30d4663e52955ed922c10fdf541419d8c67671 | music21/romanText/translate.py | python | _copyMultipleMeasures | (t, p, kCurrent) | return measures, kCurrent | Given a RomanText token for a RTMeasure, a
Part used as the current container, and the current Key,
return a Measure range copied from the past of the Part.
This is used for cases such as:
m23-25 = m20-22 | Given a RomanText token for a RTMeasure, a
Part used as the current container, and the current Key,
return a Measure range copied from the past of the Part. | [
"Given",
"a",
"RomanText",
"token",
"for",
"a",
"RTMeasure",
"a",
"Part",
"used",
"as",
"the",
"current",
"container",
"and",
"the",
"current",
"Key",
"return",
"a",
"Measure",
"range",
"copied",
"from",
"the",
"past",
"of",
"the",
"Part",
"."
] | def _copyMultipleMeasures(t, p, kCurrent):
'''
Given a RomanText token for a RTMeasure, a
Part used as the current container, and the current Key,
return a Measure range copied from the past of the Part.
This is used for cases such as:
m23-25 = m20-22
'''
# the key provided needs to be ... | [
"def",
"_copyMultipleMeasures",
"(",
"t",
",",
"p",
",",
"kCurrent",
")",
":",
"# the key provided needs to be the current key",
"# environLocal.printDebug(['calling _copyMultipleMeasures()'])",
"targetNumbers",
",",
"unused_targetRepeat",
"=",
"t",
".",
"getCopyTarget",
"(",
... | https://github.com/cuthbertLab/music21/blob/bd30d4663e52955ed922c10fdf541419d8c67671/music21/romanText/translate.py#L224-L286 | |
DataBiosphere/toil | 2e148eee2114ece8dcc3ec8a83f36333266ece0d | src/toil/deferred.py | python | DeferredFunctionManager._runAllDeferredFunctions | (self, fileObj) | Read and run deferred functions until EOF from the given open file. | Read and run deferred functions until EOF from the given open file. | [
"Read",
"and",
"run",
"deferred",
"functions",
"until",
"EOF",
"from",
"the",
"given",
"open",
"file",
"."
] | def _runAllDeferredFunctions(self, fileObj):
"""
Read and run deferred functions until EOF from the given open file.
"""
try:
while True:
# Load each function
deferredFunction = dill.load(fileObj)
logger.debug("Loaded deferred ... | [
"def",
"_runAllDeferredFunctions",
"(",
"self",
",",
"fileObj",
")",
":",
"try",
":",
"while",
"True",
":",
"# Load each function",
"deferredFunction",
"=",
"dill",
".",
"load",
"(",
"fileObj",
")",
"logger",
".",
"debug",
"(",
"\"Loaded deferred function %s\"",
... | https://github.com/DataBiosphere/toil/blob/2e148eee2114ece8dcc3ec8a83f36333266ece0d/src/toil/deferred.py#L235-L249 | ||
kerlomz/captcha_platform | f7d719bd1239a987996e266bd7fe35c96003b378 | sdk/onnx/sdk.py | python | Pretreatment.preprocessing_by_func | (exec_map, key, src_arr) | return cv2.cvtColor(target_arr, cv2.COLOR_BGR2RGB) | [] | def preprocessing_by_func(exec_map, key, src_arr):
if not exec_map:
return src_arr
target_arr = cv2.cvtColor(src_arr, cv2.COLOR_RGB2BGR)
for sentence in exec_map.get(key):
if sentence.startswith("@@"):
target_arr = eval(sentence[2:])
elif sente... | [
"def",
"preprocessing_by_func",
"(",
"exec_map",
",",
"key",
",",
"src_arr",
")",
":",
"if",
"not",
"exec_map",
":",
"return",
"src_arr",
"target_arr",
"=",
"cv2",
".",
"cvtColor",
"(",
"src_arr",
",",
"cv2",
".",
"COLOR_RGB2BGR",
")",
"for",
"sentence",
"... | https://github.com/kerlomz/captcha_platform/blob/f7d719bd1239a987996e266bd7fe35c96003b378/sdk/onnx/sdk.py#L506-L515 | |||
triaquae/triaquae | bbabf736b3ba56a0c6498e7f04e16c13b8b8f2b9 | TriAquae/models/Centos_5.9/paramiko/channel.py | python | Channel._request_success | (self, m) | return | [] | def _request_success(self, m):
self._log(DEBUG, 'Sesch channel %d request ok' % self.chanid)
self.event_ready = True
self.event.set()
return | [
"def",
"_request_success",
"(",
"self",
",",
"m",
")",
":",
"self",
".",
"_log",
"(",
"DEBUG",
",",
"'Sesch channel %d request ok'",
"%",
"self",
".",
"chanid",
")",
"self",
".",
"event_ready",
"=",
"True",
"self",
".",
"event",
".",
"set",
"(",
")",
"... | https://github.com/triaquae/triaquae/blob/bbabf736b3ba56a0c6498e7f04e16c13b8b8f2b9/TriAquae/models/Centos_5.9/paramiko/channel.py#L955-L959 | |||
progrium/duplex | d076a972fc21e19dff1f0a619e29dc6c5b25cc1f | python/duplex/sync.py | python | Channel.close | (self) | [] | def close(self):
self.peer.close() | [
"def",
"close",
"(",
"self",
")",
":",
"self",
".",
"peer",
".",
"close",
"(",
")"
] | https://github.com/progrium/duplex/blob/d076a972fc21e19dff1f0a619e29dc6c5b25cc1f/python/duplex/sync.py#L26-L27 | ||||
microsoft/ptvsd | 99c8513921021d2cc7cd82e132b65c644c256768 | src/ptvsd/_vendored/pydevd/_pydev_bundle/_pydev_completer.py | python | extract_token_and_qualifier | (text, line=0, column=0) | return TokenAndQualifier(token, qualifier) | Extracts the token a qualifier from the text given the line/colum
(see test_extract_token_and_qualifier for examples).
:param unicode text:
:param int line: 0-based
:param int column: 0-based | Extracts the token a qualifier from the text given the line/colum
(see test_extract_token_and_qualifier for examples). | [
"Extracts",
"the",
"token",
"a",
"qualifier",
"from",
"the",
"text",
"given",
"the",
"line",
"/",
"colum",
"(",
"see",
"test_extract_token_and_qualifier",
"for",
"examples",
")",
"."
] | def extract_token_and_qualifier(text, line=0, column=0):
'''
Extracts the token a qualifier from the text given the line/colum
(see test_extract_token_and_qualifier for examples).
:param unicode text:
:param int line: 0-based
:param int column: 0-based
'''
# Note: not using the tokenize... | [
"def",
"extract_token_and_qualifier",
"(",
"text",
",",
"line",
"=",
"0",
",",
"column",
"=",
"0",
")",
":",
"# Note: not using the tokenize module because text should be unicode and",
"# line/column refer to the unicode text (otherwise we'd have to know",
"# those ranges after conve... | https://github.com/microsoft/ptvsd/blob/99c8513921021d2cc7cd82e132b65c644c256768/src/ptvsd/_vendored/pydevd/_pydev_bundle/_pydev_completer.py#L243-L291 | |
wistbean/learn_python3_spider | 73c873f4845f4385f097e5057407d03dd37a117b | stackoverflow/venv/lib/python3.6/site-packages/twisted/internet/interfaces.py | python | IResolver.query | (query, timeout=None) | Dispatch C{query} to the method which can handle its type.
@type query: L{twisted.names.dns.Query}
@param query: The DNS query being issued, to which a response is to be
generated.
@type timeout: Sequence of C{int}
@param timeout: Number of seconds after which to reissue th... | Dispatch C{query} to the method which can handle its type. | [
"Dispatch",
"C",
"{",
"query",
"}",
"to",
"the",
"method",
"which",
"can",
"handle",
"its",
"type",
"."
] | def query(query, timeout=None):
"""
Dispatch C{query} to the method which can handle its type.
@type query: L{twisted.names.dns.Query}
@param query: The DNS query being issued, to which a response is to be
generated.
@type timeout: Sequence of C{int}
@param ... | [
"def",
"query",
"(",
"query",
",",
"timeout",
"=",
"None",
")",
":"
] | https://github.com/wistbean/learn_python3_spider/blob/73c873f4845f4385f097e5057407d03dd37a117b/stackoverflow/venv/lib/python3.6/site-packages/twisted/internet/interfaces.py#L192-L212 | ||
cloudmatrix/esky | 6fde3201f0335064931a6c7f7847fc5ad39001b4 | esky/sudo/__init__.py | python | _get_sudo_argtypes | (obj,methname) | return None | Get the argtypes list for the given method.
This searches the base classes of obj if the given method is not declared
allowed_from_sudo, so that people don't have to constantly re-apply the
decorator. | Get the argtypes list for the given method. | [
"Get",
"the",
"argtypes",
"list",
"for",
"the",
"given",
"method",
"."
] | def _get_sudo_argtypes(obj,methname):
"""Get the argtypes list for the given method.
This searches the base classes of obj if the given method is not declared
allowed_from_sudo, so that people don't have to constantly re-apply the
decorator.
"""
for base in _get_mro(obj):
try:
... | [
"def",
"_get_sudo_argtypes",
"(",
"obj",
",",
"methname",
")",
":",
"for",
"base",
"in",
"_get_mro",
"(",
"obj",
")",
":",
"try",
":",
"argtypes",
"=",
"base",
".",
"__dict__",
"[",
"methname",
"]",
".",
"_esky_sudo_argtypes",
"except",
"(",
"KeyError",
... | https://github.com/cloudmatrix/esky/blob/6fde3201f0335064931a6c7f7847fc5ad39001b4/esky/sudo/__init__.py#L279-L293 | |
readthedocs/sphinx-autoapi | 71c6ceebe0b02c34027fcd3d56c8641e9b94c7af | autoapi/directives.py | python | NestedParse.run | (self) | return node.children | [] | def run(self):
node = nodes.container()
node.document = self.state.document
nested_parse_with_titles(self.state, self.content, node)
try:
title_node = node[0][0]
if isinstance(title_node, nodes.title):
del node[0][0]
except IndexError:
... | [
"def",
"run",
"(",
"self",
")",
":",
"node",
"=",
"nodes",
".",
"container",
"(",
")",
"node",
".",
"document",
"=",
"self",
".",
"state",
".",
"document",
"nested_parse_with_titles",
"(",
"self",
".",
"state",
",",
"self",
".",
"content",
",",
"node",... | https://github.com/readthedocs/sphinx-autoapi/blob/71c6ceebe0b02c34027fcd3d56c8641e9b94c7af/autoapi/directives.py#L53-L63 | |||
kwotsin/mimicry | 70ce919b0684b14af264881cc6acf4eccaff42b2 | torch_mimicry/metrics/fid/fid_utils.py | python | calculate_activation_statistics | (images, sess, batch_size=50, verbose=True) | return mu, sigma | Calculation of the statistics used by the FID.
Args:
images (ndarray): Numpy array of shape (N, H, W, 3) and values in
the range [0, 255].
sess (Session): TensorFlow session object.
batch_size (int): Batch size for inference.
verbose (bool): If True, prints out logging i... | Calculation of the statistics used by the FID. | [
"Calculation",
"of",
"the",
"statistics",
"used",
"by",
"the",
"FID",
"."
] | def calculate_activation_statistics(images, sess, batch_size=50, verbose=True):
"""
Calculation of the statistics used by the FID.
Args:
images (ndarray): Numpy array of shape (N, H, W, 3) and values in
the range [0, 255].
sess (Session): TensorFlow session object.
batch... | [
"def",
"calculate_activation_statistics",
"(",
"images",
",",
"sess",
",",
"batch_size",
"=",
"50",
",",
"verbose",
"=",
"True",
")",
":",
"act",
"=",
"inception_utils",
".",
"get_activations",
"(",
"images",
",",
"sess",
",",
"batch_size",
",",
"verbose",
"... | https://github.com/kwotsin/mimicry/blob/70ce919b0684b14af264881cc6acf4eccaff42b2/torch_mimicry/metrics/fid/fid_utils.py#L70-L89 | |
lucadelu/pyModis | de86ccf28fffcb759d18b4b5b5a601304ec4fd14 | pymodis/convertmodis.py | python | processModis.__init__ | (self, hdfname, confile, mrtpath) | Function to initialize the object | Function to initialize the object | [
"Function",
"to",
"initialize",
"the",
"object"
] | def __init__(self, hdfname, confile, mrtpath):
"""Function to initialize the object"""
# check if the hdf file exists
if os.path.exists(hdfname):
self.name = hdfname
else:
raise Exception('%s does not exist' % hdfname)
# check if confile exists
if ... | [
"def",
"__init__",
"(",
"self",
",",
"hdfname",
",",
"confile",
",",
"mrtpath",
")",
":",
"# check if the hdf file exists",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"hdfname",
")",
":",
"self",
".",
"name",
"=",
"hdfname",
"else",
":",
"raise",
"Exce... | https://github.com/lucadelu/pyModis/blob/de86ccf28fffcb759d18b4b5b5a601304ec4fd14/pymodis/convertmodis.py#L210-L224 | ||
pyparallel/pyparallel | 11e8c6072d48c8f13641925d17b147bf36ee0ba3 | Lib/site-packages/pip-7.1.2-py3.3.egg/pip/_vendor/distlib/_backport/shutil.py | python | _check_unpack_options | (extensions, function, extra_args) | Checks what gets registered as an unpacker. | Checks what gets registered as an unpacker. | [
"Checks",
"what",
"gets",
"registered",
"as",
"an",
"unpacker",
"."
] | def _check_unpack_options(extensions, function, extra_args):
"""Checks what gets registered as an unpacker."""
# first make sure no other unpacker is registered for this extension
existing_extensions = {}
for name, info in _UNPACK_FORMATS.items():
for ext in info[0]:
existing_extensi... | [
"def",
"_check_unpack_options",
"(",
"extensions",
",",
"function",
",",
"extra_args",
")",
":",
"# first make sure no other unpacker is registered for this extension",
"existing_extensions",
"=",
"{",
"}",
"for",
"name",
",",
"info",
"in",
"_UNPACK_FORMATS",
".",
"items"... | https://github.com/pyparallel/pyparallel/blob/11e8c6072d48c8f13641925d17b147bf36ee0ba3/Lib/site-packages/pip-7.1.2-py3.3.egg/pip/_vendor/distlib/_backport/shutil.py#L610-L625 | ||
zhl2008/awd-platform | 0416b31abea29743387b10b3914581fbe8e7da5e | web_hxb2/lib/python3.5/site-packages/wagtail/wagtailsearch/backends/elasticsearch.py | python | ElasticsearchIndexRebuilder.reset_index | (self) | [] | def reset_index(self):
self.index.reset() | [
"def",
"reset_index",
"(",
"self",
")",
":",
"self",
".",
"index",
".",
"reset",
"(",
")"
] | https://github.com/zhl2008/awd-platform/blob/0416b31abea29743387b10b3914581fbe8e7da5e/web_hxb2/lib/python3.5/site-packages/wagtail/wagtailsearch/backends/elasticsearch.py#L615-L616 | ||||
mesalock-linux/mesapy | ed546d59a21b36feb93e2309d5c6b75aa0ad95c9 | rpython/jit/metainterp/optimizeopt/virtualstate.py | python | AbstractVirtualStateInfo.generate_guards | (self, other, op, runtime_op, state) | generate guards (output in the list extra_guards) that make runtime
values of the shape other match the shape of self. if that's not
possible, VirtualStatesCantMatch is thrown and bad gets keys set which
parts of the state are the problem.
the function can peek into the information abou... | generate guards (output in the list extra_guards) that make runtime
values of the shape other match the shape of self. if that's not
possible, VirtualStatesCantMatch is thrown and bad gets keys set which
parts of the state are the problem. | [
"generate",
"guards",
"(",
"output",
"in",
"the",
"list",
"extra_guards",
")",
"that",
"make",
"runtime",
"values",
"of",
"the",
"shape",
"other",
"match",
"the",
"shape",
"of",
"self",
".",
"if",
"that",
"s",
"not",
"possible",
"VirtualStatesCantMatch",
"is... | def generate_guards(self, other, op, runtime_op, state):
""" generate guards (output in the list extra_guards) that make runtime
values of the shape other match the shape of self. if that's not
possible, VirtualStatesCantMatch is thrown and bad gets keys set which
parts of the state are ... | [
"def",
"generate_guards",
"(",
"self",
",",
"other",
",",
"op",
",",
"runtime_op",
",",
"state",
")",
":",
"assert",
"self",
".",
"position",
"!=",
"-",
"1",
"if",
"self",
".",
"position",
"in",
"state",
".",
"renum",
":",
"if",
"state",
".",
"renum"... | https://github.com/mesalock-linux/mesapy/blob/ed546d59a21b36feb93e2309d5c6b75aa0ad95c9/rpython/jit/metainterp/optimizeopt/virtualstate.py#L70-L99 | ||
stratosphereips/Manati | 20e55d49edf00f8503807c62397d02a0dad9ddff | manati/api_manager/core/modules_manager.py | python | ModulesManager.get_filtered_weblogs_json | (**kwargs) | return json.dumps(weblogs_json) | [] | def get_filtered_weblogs_json(**kwargs):
weblogs_qs = Weblog.objects.filter(Q(**kwargs))
weblogs_json = WeblogSerializer(weblogs_qs, many=True).data
return json.dumps(weblogs_json) | [
"def",
"get_filtered_weblogs_json",
"(",
"*",
"*",
"kwargs",
")",
":",
"weblogs_qs",
"=",
"Weblog",
".",
"objects",
".",
"filter",
"(",
"Q",
"(",
"*",
"*",
"kwargs",
")",
")",
"weblogs_json",
"=",
"WeblogSerializer",
"(",
"weblogs_qs",
",",
"many",
"=",
... | https://github.com/stratosphereips/Manati/blob/20e55d49edf00f8503807c62397d02a0dad9ddff/manati/api_manager/core/modules_manager.py#L150-L153 | |||
aws-samples/aws-kube-codesuite | ab4e5ce45416b83bffb947ab8d234df5437f4fca | src/click/utils.py | python | get_binary_stream | (name) | return opener() | Returns a system stream for byte processing. This essentially
returns the stream from the sys module with the given name but it
solves some compatibility issues between different Python versions.
Primarily this function is necessary for getting binary streams on
Python 3.
:param name: the name of ... | Returns a system stream for byte processing. This essentially
returns the stream from the sys module with the given name but it
solves some compatibility issues between different Python versions.
Primarily this function is necessary for getting binary streams on
Python 3. | [
"Returns",
"a",
"system",
"stream",
"for",
"byte",
"processing",
".",
"This",
"essentially",
"returns",
"the",
"stream",
"from",
"the",
"sys",
"module",
"with",
"the",
"given",
"name",
"but",
"it",
"solves",
"some",
"compatibility",
"issues",
"between",
"diffe... | def get_binary_stream(name):
"""Returns a system stream for byte processing. This essentially
returns the stream from the sys module with the given name but it
solves some compatibility issues between different Python versions.
Primarily this function is necessary for getting binary streams on
Pyth... | [
"def",
"get_binary_stream",
"(",
"name",
")",
":",
"opener",
"=",
"binary_streams",
".",
"get",
"(",
"name",
")",
"if",
"opener",
"is",
"None",
":",
"raise",
"TypeError",
"(",
"'Unknown standard stream %r'",
"%",
"name",
")",
"return",
"opener",
"(",
")"
] | https://github.com/aws-samples/aws-kube-codesuite/blob/ab4e5ce45416b83bffb947ab8d234df5437f4fca/src/click/utils.py#L263-L276 | |
enthought/traitsui | b7c38c7a47bf6ae7971f9ddab70c8a358647dd25 | traitsui/qt4/list_editor.py | python | NotebookEditor.update_page_name | (self, object, name, old, new) | Handles the trait defining a particular page's name being changed. | Handles the trait defining a particular page's name being changed. | [
"Handles",
"the",
"trait",
"defining",
"a",
"particular",
"page",
"s",
"name",
"being",
"changed",
"."
] | def update_page_name(self, object, name, old, new):
"""Handles the trait defining a particular page's name being changed."""
for i, value in enumerate(self._uis):
page, ui, _, _ = value
if object is ui.info.object:
name = None
handler = getattr(
... | [
"def",
"update_page_name",
"(",
"self",
",",
"object",
",",
"name",
",",
"old",
",",
"new",
")",
":",
"for",
"i",
",",
"value",
"in",
"enumerate",
"(",
"self",
".",
"_uis",
")",
":",
"page",
",",
"ui",
",",
"_",
",",
"_",
"=",
"value",
"if",
"o... | https://github.com/enthought/traitsui/blob/b7c38c7a47bf6ae7971f9ddab70c8a358647dd25/traitsui/qt4/list_editor.py#L626-L646 | ||
DxCx/plugin.video.9anime | 34358c2f701e5ddf19d3276926374a16f63f7b6a | resources/lib/ui/js2py/legecy_translators/nparser.py | python | parseExpression | () | return delegate.markEndIf(expr) | [] | def parseExpression():
expr = None
delegate.markStart()
expr = parseAssignmentExpression()
if match(","):
expr = delegate.createSequenceExpression([expr])
while index < length:
if not match(","):
break
lex()
expr.expressions.append(pars... | [
"def",
"parseExpression",
"(",
")",
":",
"expr",
"=",
"None",
"delegate",
".",
"markStart",
"(",
")",
"expr",
"=",
"parseAssignmentExpression",
"(",
")",
"if",
"match",
"(",
"\",\"",
")",
":",
"expr",
"=",
"delegate",
".",
"createSequenceExpression",
"(",
... | https://github.com/DxCx/plugin.video.9anime/blob/34358c2f701e5ddf19d3276926374a16f63f7b6a/resources/lib/ui/js2py/legecy_translators/nparser.py#L1937-L1948 | |||
facebookresearch/mmf | fb6fe390287e1da12c3bd28d4ab43c5f7dcdfc9f | mmf/utils/logger.py | python | _find_caller | () | Returns:
str: module name of the caller
tuple: a hashable key to be used to identify different callers | Returns:
str: module name of the caller
tuple: a hashable key to be used to identify different callers | [
"Returns",
":",
"str",
":",
"module",
"name",
"of",
"the",
"caller",
"tuple",
":",
"a",
"hashable",
"key",
"to",
"be",
"used",
"to",
"identify",
"different",
"callers"
] | def _find_caller():
"""
Returns:
str: module name of the caller
tuple: a hashable key to be used to identify different callers
"""
frame = sys._getframe(2)
while frame:
code = frame.f_code
if os.path.join("utils", "logger.") not in code.co_filename:
mod_na... | [
"def",
"_find_caller",
"(",
")",
":",
"frame",
"=",
"sys",
".",
"_getframe",
"(",
"2",
")",
"while",
"frame",
":",
"code",
"=",
"frame",
".",
"f_code",
"if",
"os",
".",
"path",
".",
"join",
"(",
"\"utils\"",
",",
"\"logger.\"",
")",
"not",
"in",
"c... | https://github.com/facebookresearch/mmf/blob/fb6fe390287e1da12c3bd28d4ab43c5f7dcdfc9f/mmf/utils/logger.py#L196-L210 | ||
yt-project/yt | dc7b24f9b266703db4c843e329c6c8644d47b824 | yt/frontends/owls/data_structures.py | python | OWLSDataset._set_code_unit_attributes | (self) | [] | def _set_code_unit_attributes(self):
self._set_owls_eagle_units() | [
"def",
"_set_code_unit_attributes",
"(",
"self",
")",
":",
"self",
".",
"_set_owls_eagle_units",
"(",
")"
] | https://github.com/yt-project/yt/blob/dc7b24f9b266703db4c843e329c6c8644d47b824/yt/frontends/owls/data_structures.py#L30-L31 | ||||
Dentosal/python-sc2 | e816cce83772d1aee1291b86b300b69405aa96b4 | sc2/bot_ai.py | python | BotAI.in_pathing_grid | (self, pos: Union[Point2, Point3, Unit]) | return self._game_info.pathing_grid[pos] == 0 | Returns True if a unit can pass through a grid point. | Returns True if a unit can pass through a grid point. | [
"Returns",
"True",
"if",
"a",
"unit",
"can",
"pass",
"through",
"a",
"grid",
"point",
"."
] | def in_pathing_grid(self, pos: Union[Point2, Point3, Unit]) -> bool:
""" Returns True if a unit can pass through a grid point. """
assert isinstance(pos, (Point2, Point3, Unit))
pos = pos.position.to2.rounded
return self._game_info.pathing_grid[pos] == 0 | [
"def",
"in_pathing_grid",
"(",
"self",
",",
"pos",
":",
"Union",
"[",
"Point2",
",",
"Point3",
",",
"Unit",
"]",
")",
"->",
"bool",
":",
"assert",
"isinstance",
"(",
"pos",
",",
"(",
"Point2",
",",
"Point3",
",",
"Unit",
")",
")",
"pos",
"=",
"pos"... | https://github.com/Dentosal/python-sc2/blob/e816cce83772d1aee1291b86b300b69405aa96b4/sc2/bot_ai.py#L482-L486 | |
koaning/scikit-lego | 028597fd0ba9ac387b9faa6f06050a7ee05e6cba | sklego/naive_bayes.py | python | BayesianGaussianMixtureNB.fit | (self, X: np.array, y: np.array) | return self | Fit the model using X, y as training data.
:param X: array-like, shape=(n_columns, n_samples, ) training data.
:param y: array-like, shape=(n_samples, ) training data.
:return: Returns an instance of self. | Fit the model using X, y as training data. | [
"Fit",
"the",
"model",
"using",
"X",
"y",
"as",
"training",
"data",
"."
] | def fit(self, X: np.array, y: np.array) -> "BayesianGaussianMixtureNB":
"""
Fit the model using X, y as training data.
:param X: array-like, shape=(n_columns, n_samples, ) training data.
:param y: array-like, shape=(n_samples, ) training data.
:return: Returns an instance of sel... | [
"def",
"fit",
"(",
"self",
",",
"X",
":",
"np",
".",
"array",
",",
"y",
":",
"np",
".",
"array",
")",
"->",
"\"BayesianGaussianMixtureNB\"",
":",
"X",
",",
"y",
"=",
"check_X_y",
"(",
"X",
",",
"y",
",",
"estimator",
"=",
"self",
",",
"dtype",
"=... | https://github.com/koaning/scikit-lego/blob/028597fd0ba9ac387b9faa6f06050a7ee05e6cba/sklego/naive_bayes.py#L156-L195 | |
chapmanb/bcbb | dbfb52711f0bfcc1d26c5a5b53c9ff4f50dc0027 | nextgen/bcbio/utils.py | python | replace_suffix | (filename, suffix) | return base + suffix | replace the suffix of filename with suffix
example: replace_suffix("/path/to/test.sam", ".bam") ->
"/path/to/test.bam" | replace the suffix of filename with suffix
example: replace_suffix("/path/to/test.sam", ".bam") ->
"/path/to/test.bam" | [
"replace",
"the",
"suffix",
"of",
"filename",
"with",
"suffix",
"example",
":",
"replace_suffix",
"(",
"/",
"path",
"/",
"to",
"/",
"test",
".",
"sam",
".",
"bam",
")",
"-",
">",
"/",
"path",
"/",
"to",
"/",
"test",
".",
"bam"
] | def replace_suffix(filename, suffix):
"""
replace the suffix of filename with suffix
example: replace_suffix("/path/to/test.sam", ".bam") ->
"/path/to/test.bam"
"""
(base, _) = os.path.splitext(filename)
return base + suffix | [
"def",
"replace_suffix",
"(",
"filename",
",",
"suffix",
")",
":",
"(",
"base",
",",
"_",
")",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"return",
"base",
"+",
"suffix"
] | https://github.com/chapmanb/bcbb/blob/dbfb52711f0bfcc1d26c5a5b53c9ff4f50dc0027/nextgen/bcbio/utils.py#L273-L281 | |
aws-samples/aws-kube-codesuite | ab4e5ce45416b83bffb947ab8d234df5437f4fca | src/pkg_resources/__init__.py | python | ResourceManager.postprocess | (self, tempname, filename) | Perform any platform-specific postprocessing of `tempname`
This is where Mac header rewrites should be done; other platforms don't
have anything special they should do.
Resource providers should call this method ONLY after successfully
extracting a compressed resource. They must NOT c... | Perform any platform-specific postprocessing of `tempname` | [
"Perform",
"any",
"platform",
"-",
"specific",
"postprocessing",
"of",
"tempname"
] | def postprocess(self, tempname, filename):
"""Perform any platform-specific postprocessing of `tempname`
This is where Mac header rewrites should be done; other platforms don't
have anything special they should do.
Resource providers should call this method ONLY after successfully
... | [
"def",
"postprocess",
"(",
"self",
",",
"tempname",
",",
"filename",
")",
":",
"if",
"os",
".",
"name",
"==",
"'posix'",
":",
"# Make the resource executable",
"mode",
"=",
"(",
"(",
"os",
".",
"stat",
"(",
"tempname",
")",
".",
"st_mode",
")",
"|",
"0... | https://github.com/aws-samples/aws-kube-codesuite/blob/ab4e5ce45416b83bffb947ab8d234df5437f4fca/src/pkg_resources/__init__.py#L1319-L1337 | ||
nosmokingbandit/Watcher3 | 0217e75158b563bdefc8e01c3be7620008cf3977 | lib/sqlalchemy/ext/mutable.py | python | MutableDict.__setitem__ | (self, key, value) | Detect dictionary set events and emit change events. | Detect dictionary set events and emit change events. | [
"Detect",
"dictionary",
"set",
"events",
"and",
"emit",
"change",
"events",
"."
] | def __setitem__(self, key, value):
"""Detect dictionary set events and emit change events."""
dict.__setitem__(self, key, value)
self.changed() | [
"def",
"__setitem__",
"(",
"self",
",",
"key",
",",
"value",
")",
":",
"dict",
".",
"__setitem__",
"(",
"self",
",",
"key",
",",
"value",
")",
"self",
".",
"changed",
"(",
")"
] | https://github.com/nosmokingbandit/Watcher3/blob/0217e75158b563bdefc8e01c3be7620008cf3977/lib/sqlalchemy/ext/mutable.py#L660-L663 | ||
google/deepvariant | 9cf1c7b0e2342d013180aa153cba3c9331c9aef7 | deepvariant/vcf_stats.py | python | _tstv | (variant, vtype) | return is_transition, is_transversion | Returns a pair of bools indicating Transition, Transversion status. | Returns a pair of bools indicating Transition, Transversion status. | [
"Returns",
"a",
"pair",
"of",
"bools",
"indicating",
"Transition",
"Transversion",
"status",
"."
] | def _tstv(variant, vtype):
"""Returns a pair of bools indicating Transition, Transversion status."""
if vtype == BIALLELIC_SNP:
is_transition = variant_utils.is_transition(variant.reference_bases,
variant.alternate_bases[0])
is_transversion = not is_transition... | [
"def",
"_tstv",
"(",
"variant",
",",
"vtype",
")",
":",
"if",
"vtype",
"==",
"BIALLELIC_SNP",
":",
"is_transition",
"=",
"variant_utils",
".",
"is_transition",
"(",
"variant",
".",
"reference_bases",
",",
"variant",
".",
"alternate_bases",
"[",
"0",
"]",
")"... | https://github.com/google/deepvariant/blob/9cf1c7b0e2342d013180aa153cba3c9331c9aef7/deepvariant/vcf_stats.py#L89-L98 | |
pymedusa/Medusa | 1405fbb6eb8ef4d20fcca24c32ddca52b11f0f38 | ext/pint/compat/tokenize.py | python | _get_normal_name | (orig_enc) | return orig_enc | Imitates get_normal_name in tokenizer.c. | Imitates get_normal_name in tokenizer.c. | [
"Imitates",
"get_normal_name",
"in",
"tokenizer",
".",
"c",
"."
] | def _get_normal_name(orig_enc):
"""Imitates get_normal_name in tokenizer.c."""
# Only care about the first 12 characters.
enc = orig_enc[:12].lower().replace("_", "-")
if enc == "utf-8" or enc.startswith("utf-8-"):
return "utf-8"
if enc in ("latin-1", "iso-8859-1", "iso-latin-1") or \
... | [
"def",
"_get_normal_name",
"(",
"orig_enc",
")",
":",
"# Only care about the first 12 characters.",
"enc",
"=",
"orig_enc",
"[",
":",
"12",
"]",
".",
"lower",
"(",
")",
".",
"replace",
"(",
"\"_\"",
",",
"\"-\"",
")",
"if",
"enc",
"==",
"\"utf-8\"",
"or",
... | https://github.com/pymedusa/Medusa/blob/1405fbb6eb8ef4d20fcca24c32ddca52b11f0f38/ext/pint/compat/tokenize.py#L343-L352 | |
asappresearch/sru | 9ddc8da12f067125c2cfdd4f3b28a87c02889681 | sru/modules.py | python | SRUCell.forward | (self,
input: Tensor,
c0: Optional[Tensor] = None,
mask_pad: Optional[Tensor] = None) | return h, c | The forward method of the SRU layer. | The forward method of the SRU layer. | [
"The",
"forward",
"method",
"of",
"the",
"SRU",
"layer",
"."
] | def forward(self,
input: Tensor,
c0: Optional[Tensor] = None,
mask_pad: Optional[Tensor] = None) -> Tuple[Tensor, Tensor]:
"""The forward method of the SRU layer.
"""
if input.dim() != 2 and input.dim() != 3:
raise ValueError("Input mu... | [
"def",
"forward",
"(",
"self",
",",
"input",
":",
"Tensor",
",",
"c0",
":",
"Optional",
"[",
"Tensor",
"]",
"=",
"None",
",",
"mask_pad",
":",
"Optional",
"[",
"Tensor",
"]",
"=",
"None",
")",
"->",
"Tuple",
"[",
"Tensor",
",",
"Tensor",
"]",
":",
... | https://github.com/asappresearch/sru/blob/9ddc8da12f067125c2cfdd4f3b28a87c02889681/sru/modules.py#L235-L281 | |
silverapp/silver | a59dbc7216733ab49dca2fae525d229bdba04420 | silver/api/views/payment_method_views.py | python | PaymentMethodList.get_queryset | (self) | return PaymentMethod.objects.filter(customer=self.customer) | [] | def get_queryset(self):
return PaymentMethod.objects.filter(customer=self.customer) | [
"def",
"get_queryset",
"(",
"self",
")",
":",
"return",
"PaymentMethod",
".",
"objects",
".",
"filter",
"(",
"customer",
"=",
"self",
".",
"customer",
")"
] | https://github.com/silverapp/silver/blob/a59dbc7216733ab49dca2fae525d229bdba04420/silver/api/views/payment_method_views.py#L61-L62 | |||
mesonbuild/meson | a22d0f9a0a787df70ce79b05d0c45de90a970048 | mesonbuild/backend/vs2010backend.py | python | Vs2010Backend.get_vcvars_command | (self) | return '' | [] | def get_vcvars_command(self):
has_arch_values = 'VSCMD_ARG_TGT_ARCH' in os.environ and 'VSCMD_ARG_HOST_ARCH' in os.environ
# Use vcvarsall.bat if we found it.
if 'VCINSTALLDIR' in os.environ:
vs_version = os.environ['VisualStudioVersion'] \
if 'VisualStudioVersion' i... | [
"def",
"get_vcvars_command",
"(",
"self",
")",
":",
"has_arch_values",
"=",
"'VSCMD_ARG_TGT_ARCH'",
"in",
"os",
".",
"environ",
"and",
"'VSCMD_ARG_HOST_ARCH'",
"in",
"os",
".",
"environ",
"# Use vcvarsall.bat if we found it.",
"if",
"'VCINSTALLDIR'",
"in",
"os",
".",
... | https://github.com/mesonbuild/meson/blob/a22d0f9a0a787df70ce79b05d0c45de90a970048/mesonbuild/backend/vs2010backend.py#L239-L264 | |||
ranahanocka/MeshCNN | 15b83cc6a4db968baf6cf595df78995a9c2dcee3 | models/layers/mesh_prepare.py | python | build_gemm | (mesh, faces, face_areas) | gemm_edges: array (#E x 4) of the 4 one-ring neighbors for each edge
sides: array (#E x 4) indices (values of: 0,1,2,3) indicating where an edge is in the gemm_edge entry of the 4 neighboring edges
for example edge i -> gemm_edges[gemm_edges[i], sides[i]] == [i, i, i, i] | gemm_edges: array (#E x 4) of the 4 one-ring neighbors for each edge
sides: array (#E x 4) indices (values of: 0,1,2,3) indicating where an edge is in the gemm_edge entry of the 4 neighboring edges
for example edge i -> gemm_edges[gemm_edges[i], sides[i]] == [i, i, i, i] | [
"gemm_edges",
":",
"array",
"(",
"#E",
"x",
"4",
")",
"of",
"the",
"4",
"one",
"-",
"ring",
"neighbors",
"for",
"each",
"edge",
"sides",
":",
"array",
"(",
"#E",
"x",
"4",
")",
"indices",
"(",
"values",
"of",
":",
"0",
"1",
"2",
"3",
")",
"indi... | def build_gemm(mesh, faces, face_areas):
"""
gemm_edges: array (#E x 4) of the 4 one-ring neighbors for each edge
sides: array (#E x 4) indices (values of: 0,1,2,3) indicating where an edge is in the gemm_edge entry of the 4 neighboring edges
for example edge i -> gemm_edges[gemm_edges[i], sides[i]] == ... | [
"def",
"build_gemm",
"(",
"mesh",
",",
"faces",
",",
"face_areas",
")",
":",
"mesh",
".",
"ve",
"=",
"[",
"[",
"]",
"for",
"_",
"in",
"mesh",
".",
"vs",
"]",
"edge_nb",
"=",
"[",
"]",
"sides",
"=",
"[",
"]",
"edge2key",
"=",
"dict",
"(",
")",
... | https://github.com/ranahanocka/MeshCNN/blob/15b83cc6a4db968baf6cf595df78995a9c2dcee3/models/layers/mesh_prepare.py#L116-L161 | ||
openshift/openshift-tools | 1188778e728a6e4781acf728123e5b356380fe6f | openshift/installer/vendored/openshift-ansible-3.9.40/roles/lib_utils/src/generate.py | python | main | () | combine the necessary files to create the ansible module | combine the necessary files to create the ansible module | [
"combine",
"the",
"necessary",
"files",
"to",
"create",
"the",
"ansible",
"module"
] | def main():
''' combine the necessary files to create the ansible module '''
args = parse_args()
if args.verify:
verify()
for fname, parts in get_sources().items():
data = generate(parts)
fname = os.path.join(LIBRARY, fname)
with open(fname, 'w') as afd:
afd.... | [
"def",
"main",
"(",
")",
":",
"args",
"=",
"parse_args",
"(",
")",
"if",
"args",
".",
"verify",
":",
"verify",
"(",
")",
"for",
"fname",
",",
"parts",
"in",
"get_sources",
"(",
")",
".",
"items",
"(",
")",
":",
"data",
"=",
"generate",
"(",
"part... | https://github.com/openshift/openshift-tools/blob/1188778e728a6e4781acf728123e5b356380fe6f/openshift/installer/vendored/openshift-ansible-3.9.40/roles/lib_utils/src/generate.py#L95-L106 | ||
zzzeek/sqlalchemy | fc5c54fcd4d868c2a4c7ac19668d72f506fe821e | lib/sqlalchemy/dialects/__init__.py | python | _auto_fn | (name) | default dialect importer.
plugs into the :class:`.PluginLoader`
as a first-hit system. | default dialect importer. | [
"default",
"dialect",
"importer",
"."
] | def _auto_fn(name):
"""default dialect importer.
plugs into the :class:`.PluginLoader`
as a first-hit system.
"""
if "." in name:
dialect, driver = name.split(".")
else:
dialect = name
driver = "base"
try:
if dialect == "mariadb":
# it's "OK" fo... | [
"def",
"_auto_fn",
"(",
"name",
")",
":",
"if",
"\".\"",
"in",
"name",
":",
"dialect",
",",
"driver",
"=",
"name",
".",
"split",
"(",
"\".\"",
")",
"else",
":",
"dialect",
"=",
"name",
"driver",
"=",
"\"base\"",
"try",
":",
"if",
"dialect",
"==",
"... | https://github.com/zzzeek/sqlalchemy/blob/fc5c54fcd4d868c2a4c7ac19668d72f506fe821e/lib/sqlalchemy/dialects/__init__.py#L14-L47 | ||
meduza-corp/interstellar | 40a801ccd7856491726f5a126621d9318cabe2e1 | gsutil/third_party/boto/boto/s3/bucketlistresultset.py | python | bucket_lister | (bucket, prefix='', delimiter='', marker='', headers=None,
encoding_type=None) | A generator function for listing keys in a bucket. | A generator function for listing keys in a bucket. | [
"A",
"generator",
"function",
"for",
"listing",
"keys",
"in",
"a",
"bucket",
"."
] | def bucket_lister(bucket, prefix='', delimiter='', marker='', headers=None,
encoding_type=None):
"""
A generator function for listing keys in a bucket.
"""
more_results = True
k = None
while more_results:
rs = bucket.get_all_keys(prefix=prefix, marker=marker,
... | [
"def",
"bucket_lister",
"(",
"bucket",
",",
"prefix",
"=",
"''",
",",
"delimiter",
"=",
"''",
",",
"marker",
"=",
"''",
",",
"headers",
"=",
"None",
",",
"encoding_type",
"=",
"None",
")",
":",
"more_results",
"=",
"True",
"k",
"=",
"None",
"while",
... | https://github.com/meduza-corp/interstellar/blob/40a801ccd7856491726f5a126621d9318cabe2e1/gsutil/third_party/boto/boto/s3/bucketlistresultset.py#L22-L37 | ||
zhl2008/awd-platform | 0416b31abea29743387b10b3914581fbe8e7da5e | web_hxb2/lib/python3.5/site-packages/django/utils/deprecation.py | python | RenameMethodsBase.__new__ | (cls, name, bases, attrs) | return new_class | [] | def __new__(cls, name, bases, attrs):
new_class = super(RenameMethodsBase, cls).__new__(cls, name, bases, attrs)
for base in inspect.getmro(new_class):
class_name = base.__name__
for renamed_method in cls.renamed_methods:
old_method_name = renamed_method[0]
... | [
"def",
"__new__",
"(",
"cls",
",",
"name",
",",
"bases",
",",
"attrs",
")",
":",
"new_class",
"=",
"super",
"(",
"RenameMethodsBase",
",",
"cls",
")",
".",
"__new__",
"(",
"cls",
",",
"name",
",",
"bases",
",",
"attrs",
")",
"for",
"base",
"in",
"i... | https://github.com/zhl2008/awd-platform/blob/0416b31abea29743387b10b3914581fbe8e7da5e/web_hxb2/lib/python3.5/site-packages/django/utils/deprecation.py#L46-L72 | |||
JacquesLucke/animation_nodes | b1e3ace8dcb0a771fd882fc3ac4e490b009fa0d1 | animation_nodes/ui/selection_pie.py | python | SelectionPie.drawTop | (self, layout) | [] | def drawTop(self, layout):
layout.operator("an.frame_active_network") | [
"def",
"drawTop",
"(",
"self",
",",
"layout",
")",
":",
"layout",
".",
"operator",
"(",
"\"an.frame_active_network\"",
")"
] | https://github.com/JacquesLucke/animation_nodes/blob/b1e3ace8dcb0a771fd882fc3ac4e490b009fa0d1/animation_nodes/ui/selection_pie.py#L24-L25 | ||||
AstusRush/AMaDiA | e2ad87318d9dd30bc24428e05c29cb32a29c83aa | External_Libraries/keyboard_master/keyboard/__init__.py | python | on_release | (callback, suppress=False) | return hook(lambda e: e.event_type == KEY_DOWN or callback(e), suppress=suppress) | Invokes `callback` for every KEY_UP event. For details see `hook`. | Invokes `callback` for every KEY_UP event. For details see `hook`. | [
"Invokes",
"callback",
"for",
"every",
"KEY_UP",
"event",
".",
"For",
"details",
"see",
"hook",
"."
] | def on_release(callback, suppress=False):
"""
Invokes `callback` for every KEY_UP event. For details see `hook`.
"""
return hook(lambda e: e.event_type == KEY_DOWN or callback(e), suppress=suppress) | [
"def",
"on_release",
"(",
"callback",
",",
"suppress",
"=",
"False",
")",
":",
"return",
"hook",
"(",
"lambda",
"e",
":",
"e",
".",
"event_type",
"==",
"KEY_DOWN",
"or",
"callback",
"(",
"e",
")",
",",
"suppress",
"=",
"suppress",
")"
] | https://github.com/AstusRush/AMaDiA/blob/e2ad87318d9dd30bc24428e05c29cb32a29c83aa/External_Libraries/keyboard_master/keyboard/__init__.py#L476-L480 | |
cloudera/impyla | 0c736af4cad2bade9b8e313badc08ec50e81c948 | impala/_thrift_gen/hive_metastore/ThriftHiveMetastore.py | python | drop_partition_args.write | (self, oprot) | [] | def write(self, oprot):
if oprot._fast_encode is not None and self.thrift_spec is not None:
oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec]))
return
oprot.writeStructBegin('drop_partition_args')
if self.db_name is not None:
oprot.... | [
"def",
"write",
"(",
"self",
",",
"oprot",
")",
":",
"if",
"oprot",
".",
"_fast_encode",
"is",
"not",
"None",
"and",
"self",
".",
"thrift_spec",
"is",
"not",
"None",
":",
"oprot",
".",
"trans",
".",
"write",
"(",
"oprot",
".",
"_fast_encode",
"(",
"s... | https://github.com/cloudera/impyla/blob/0c736af4cad2bade9b8e313badc08ec50e81c948/impala/_thrift_gen/hive_metastore/ThriftHiveMetastore.py#L18131-L18156 | ||||
scikit-hep/awkward-0.x | dd885bef15814f588b58944d2505296df4aaae0e | awkward0/array/base.py | python | AwkwardArray.BitMaskedArray | (self) | return awkward0.array.masked.BitMaskedArray | [] | def BitMaskedArray(self):
import awkward0.array.masked
return awkward0.array.masked.BitMaskedArray | [
"def",
"BitMaskedArray",
"(",
"self",
")",
":",
"import",
"awkward0",
".",
"array",
".",
"masked",
"return",
"awkward0",
".",
"array",
".",
"masked",
".",
"BitMaskedArray"
] | https://github.com/scikit-hep/awkward-0.x/blob/dd885bef15814f588b58944d2505296df4aaae0e/awkward0/array/base.py#L337-L339 | |||
volatilityfoundation/volatility | a438e768194a9e05eb4d9ee9338b881c0fa25937 | volatility/obj.py | python | Profile.clear | (self) | Clears out the input vtypes and object_classes, and only the base object types | Clears out the input vtypes and object_classes, and only the base object types | [
"Clears",
"out",
"the",
"input",
"vtypes",
"and",
"object_classes",
"and",
"only",
"the",
"base",
"object",
"types"
] | def clear(self):
""" Clears out the input vtypes and object_classes, and only the base object types """
# Prepopulate object_classes with base classes
self.object_classes = {'BitField': BitField,
'Pointer': Pointer,
'Pointer32':Pointe... | [
"def",
"clear",
"(",
"self",
")",
":",
"# Prepopulate object_classes with base classes",
"self",
".",
"object_classes",
"=",
"{",
"'BitField'",
":",
"BitField",
",",
"'Pointer'",
":",
"Pointer",
",",
"'Pointer32'",
":",
"Pointer32",
",",
"'Void'",
":",
"Void",
"... | https://github.com/volatilityfoundation/volatility/blob/a438e768194a9e05eb4d9ee9338b881c0fa25937/volatility/obj.py#L868-L881 | ||
algorhythms/LeetCode | 3fb14aeea62a960442e47dfde9f964c7ffce32be | 354 Russian Doll Envelopes.py | python | Solution.maxEnvelopesTLE | (self, A) | return max(F) | LIS
O(n^2)
:type A: List[List[int]]
:rtype: int | LIS
O(n^2)
:type A: List[List[int]]
:rtype: int | [
"LIS",
"O",
"(",
"n^2",
")",
":",
"type",
"A",
":",
"List",
"[",
"List",
"[",
"int",
"]]",
":",
"rtype",
":",
"int"
] | def maxEnvelopesTLE(self, A):
"""
LIS
O(n^2)
:type A: List[List[int]]
:rtype: int
"""
if not A: return 0
predicate = lambda a, b: b[0] > a[0] and b[1] > a[1]
A.sort()
n = len(A)
F = [1 for _ in xrange(n)]
for i in xrange(1,... | [
"def",
"maxEnvelopesTLE",
"(",
"self",
",",
"A",
")",
":",
"if",
"not",
"A",
":",
"return",
"0",
"predicate",
"=",
"lambda",
"a",
",",
"b",
":",
"b",
"[",
"0",
"]",
">",
"a",
"[",
"0",
"]",
"and",
"b",
"[",
"1",
"]",
">",
"a",
"[",
"1",
"... | https://github.com/algorhythms/LeetCode/blob/3fb14aeea62a960442e47dfde9f964c7ffce32be/354 Russian Doll Envelopes.py#L41-L59 | |
pywren/pywren | d898af6418a2d915c3984152de07d95e32c9789a | pywren/storage/storage.py | python | Storage.put_func | (self, key, func) | return self.backend_handler.put_object(key, func) | Put serialized function into storage.
:param key: function key
:param func: serialized function
:return: None | Put serialized function into storage.
:param key: function key
:param func: serialized function
:return: None | [
"Put",
"serialized",
"function",
"into",
"storage",
".",
":",
"param",
"key",
":",
"function",
"key",
":",
"param",
"func",
":",
"serialized",
"function",
":",
"return",
":",
"None"
] | def put_func(self, key, func):
"""
Put serialized function into storage.
:param key: function key
:param func: serialized function
:return: None
"""
return self.backend_handler.put_object(key, func) | [
"def",
"put_func",
"(",
"self",
",",
"key",
",",
"func",
")",
":",
"return",
"self",
".",
"backend_handler",
".",
"put_object",
"(",
"key",
",",
"func",
")"
] | https://github.com/pywren/pywren/blob/d898af6418a2d915c3984152de07d95e32c9789a/pywren/storage/storage.py#L62-L69 | |
androguard/androguard | 8d091cbb309c0c50bf239f805cc1e0931b8dcddc | androguard/core/bytecodes/dvm.py | python | ClassDefItem.get_static_values_off | (self) | return self.static_values_off | Return the offset from the start of the file to the list of initial values for static fields,
or 0 if there are none (and all static fields are to be initialized with 0 or null)
:rtype: int | Return the offset from the start of the file to the list of initial values for static fields,
or 0 if there are none (and all static fields are to be initialized with 0 or null) | [
"Return",
"the",
"offset",
"from",
"the",
"start",
"of",
"the",
"file",
"to",
"the",
"list",
"of",
"initial",
"values",
"for",
"static",
"fields",
"or",
"0",
"if",
"there",
"are",
"none",
"(",
"and",
"all",
"static",
"fields",
"are",
"to",
"be",
"initi... | def get_static_values_off(self):
"""
Return the offset from the start of the file to the list of initial values for static fields,
or 0 if there are none (and all static fields are to be initialized with 0 or null)
:rtype: int
"""
return self.static_values_off | [
"def",
"get_static_values_off",
"(",
"self",
")",
":",
"return",
"self",
".",
"static_values_off"
] | https://github.com/androguard/androguard/blob/8d091cbb309c0c50bf239f805cc1e0931b8dcddc/androguard/core/bytecodes/dvm.py#L3753-L3760 | |
oaubert/python-vlc | 908ffdbd0844dc1849728c456e147788798c99da | generated/3.0/vlc.py | python | Instance.playlist_play | (self, i_id, i_options, ppsz_options) | return libvlc_playlist_play(self, i_id, i_options, ppsz_options) | Start playing (if there is any item in the playlist).
Additionnal playlist item options can be specified for addition to the
item before it is played.
@param i_id: the item to play. If this is a negative number, the next item will be selected. Otherwise, the item with the given ID will be played... | Start playing (if there is any item in the playlist).
Additionnal playlist item options can be specified for addition to the
item before it is played. | [
"Start",
"playing",
"(",
"if",
"there",
"is",
"any",
"item",
"in",
"the",
"playlist",
")",
".",
"Additionnal",
"playlist",
"item",
"options",
"can",
"be",
"specified",
"for",
"addition",
"to",
"the",
"item",
"before",
"it",
"is",
"played",
"."
] | def playlist_play(self, i_id, i_options, ppsz_options):
'''Start playing (if there is any item in the playlist).
Additionnal playlist item options can be specified for addition to the
item before it is played.
@param i_id: the item to play. If this is a negative number, the next item wil... | [
"def",
"playlist_play",
"(",
"self",
",",
"i_id",
",",
"i_options",
",",
"ppsz_options",
")",
":",
"return",
"libvlc_playlist_play",
"(",
"self",
",",
"i_id",
",",
"i_options",
",",
"ppsz_options",
")"
] | https://github.com/oaubert/python-vlc/blob/908ffdbd0844dc1849728c456e147788798c99da/generated/3.0/vlc.py#L1976-L1984 | |
evennia/evennia | fa79110ba6b219932f22297838e8ac72ebc0be0e | evennia/locks/lockhandler.py | python | LockHandler.get | (self, access_type=None) | return str(self) | Get the full lockstring or the lockstring of a particular
access type.
Args:
access_type (str, optional):
Returns:
lockstring (str): The matched lockstring, or the full
lockstring if no access_type was given. | Get the full lockstring or the lockstring of a particular
access type. | [
"Get",
"the",
"full",
"lockstring",
"or",
"the",
"lockstring",
"of",
"a",
"particular",
"access",
"type",
"."
] | def get(self, access_type=None):
"""
Get the full lockstring or the lockstring of a particular
access type.
Args:
access_type (str, optional):
Returns:
lockstring (str): The matched lockstring, or the full
lockstring if no access_type was... | [
"def",
"get",
"(",
"self",
",",
"access_type",
"=",
"None",
")",
":",
"if",
"access_type",
":",
"return",
"self",
".",
"locks",
".",
"get",
"(",
"access_type",
",",
"[",
"\"\"",
",",
"\"\"",
",",
"\"\"",
"]",
")",
"[",
"2",
"]",
"return",
"str",
... | https://github.com/evennia/evennia/blob/fa79110ba6b219932f22297838e8ac72ebc0be0e/evennia/locks/lockhandler.py#L417-L432 | |
hacktoolspack/hack-tools | c2b1e324f4b24a2c5b4f111e7ef84e9c547159c2 | wifite2-master/wifite/util/process.py | python | Process.exists | (program) | return True | Checks if program is installed on this system | Checks if program is installed on this system | [
"Checks",
"if",
"program",
"is",
"installed",
"on",
"this",
"system"
] | def exists(program):
''' Checks if program is installed on this system '''
p = Process(['which', program])
stdout = p.stdout().strip()
stderr = p.stderr().strip()
if stdout == '' and stderr == '':
return False
return True | [
"def",
"exists",
"(",
"program",
")",
":",
"p",
"=",
"Process",
"(",
"[",
"'which'",
",",
"program",
"]",
")",
"stdout",
"=",
"p",
".",
"stdout",
"(",
")",
".",
"strip",
"(",
")",
"stderr",
"=",
"p",
".",
"stderr",
"(",
")",
".",
"strip",
"(",
... | https://github.com/hacktoolspack/hack-tools/blob/c2b1e324f4b24a2c5b4f111e7ef84e9c547159c2/wifite2-master/wifite/util/process.py#L55-L64 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.