repo
stringlengths 7
55
| path
stringlengths 4
223
| func_name
stringlengths 1
134
| original_string
stringlengths 75
104k
| language
stringclasses 1
value | code
stringlengths 75
104k
| code_tokens
listlengths 19
28.4k
| docstring
stringlengths 1
46.9k
| docstring_tokens
listlengths 1
1.97k
| sha
stringlengths 40
40
| url
stringlengths 87
315
| partition
stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
nion-software/nionswift
|
nion/swift/HistogramPanel.py
|
ColorMapCanvasItem._repaint
|
def _repaint(self, drawing_context: DrawingContext.DrawingContext):
"""Repaint the canvas item. This will occur on a thread."""
# canvas size
canvas_width = self.canvas_size.width
canvas_height = self.canvas_size.height
with drawing_context.saver():
if self.__color_map_data is not None:
rgba_image = numpy.empty((4,) + self.__color_map_data.shape[:-1], dtype=numpy.uint32)
Image.get_rgb_view(rgba_image)[:] = self.__color_map_data[numpy.newaxis, :, :] # scalar data assigned to each component of rgb view
Image.get_alpha_view(rgba_image)[:] = 255
drawing_context.draw_image(rgba_image, 0, 0, canvas_width, canvas_height)
|
python
|
def _repaint(self, drawing_context: DrawingContext.DrawingContext):
"""Repaint the canvas item. This will occur on a thread."""
# canvas size
canvas_width = self.canvas_size.width
canvas_height = self.canvas_size.height
with drawing_context.saver():
if self.__color_map_data is not None:
rgba_image = numpy.empty((4,) + self.__color_map_data.shape[:-1], dtype=numpy.uint32)
Image.get_rgb_view(rgba_image)[:] = self.__color_map_data[numpy.newaxis, :, :] # scalar data assigned to each component of rgb view
Image.get_alpha_view(rgba_image)[:] = 255
drawing_context.draw_image(rgba_image, 0, 0, canvas_width, canvas_height)
|
[
"def",
"_repaint",
"(",
"self",
",",
"drawing_context",
":",
"DrawingContext",
".",
"DrawingContext",
")",
":",
"# canvas size",
"canvas_width",
"=",
"self",
".",
"canvas_size",
".",
"width",
"canvas_height",
"=",
"self",
".",
"canvas_size",
".",
"height",
"with",
"drawing_context",
".",
"saver",
"(",
")",
":",
"if",
"self",
".",
"__color_map_data",
"is",
"not",
"None",
":",
"rgba_image",
"=",
"numpy",
".",
"empty",
"(",
"(",
"4",
",",
")",
"+",
"self",
".",
"__color_map_data",
".",
"shape",
"[",
":",
"-",
"1",
"]",
",",
"dtype",
"=",
"numpy",
".",
"uint32",
")",
"Image",
".",
"get_rgb_view",
"(",
"rgba_image",
")",
"[",
":",
"]",
"=",
"self",
".",
"__color_map_data",
"[",
"numpy",
".",
"newaxis",
",",
":",
",",
":",
"]",
"# scalar data assigned to each component of rgb view",
"Image",
".",
"get_alpha_view",
"(",
"rgba_image",
")",
"[",
":",
"]",
"=",
"255",
"drawing_context",
".",
"draw_image",
"(",
"rgba_image",
",",
"0",
",",
"0",
",",
"canvas_width",
",",
"canvas_height",
")"
] |
Repaint the canvas item. This will occur on a thread.
|
[
"Repaint",
"the",
"canvas",
"item",
".",
"This",
"will",
"occur",
"on",
"a",
"thread",
"."
] |
d43693eaf057b8683b9638e575000f055fede452
|
https://github.com/nion-software/nionswift/blob/d43693eaf057b8683b9638e575000f055fede452/nion/swift/HistogramPanel.py#L179-L191
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.get_params
|
def get_params(self):
"Parameters used to initialize the class"
import inspect
a = inspect.getargspec(self.__init__)[0]
out = dict()
for key in a[1:]:
value = getattr(self, "_%s" % key, None)
out[key] = value
return out
|
python
|
def get_params(self):
"Parameters used to initialize the class"
import inspect
a = inspect.getargspec(self.__init__)[0]
out = dict()
for key in a[1:]:
value = getattr(self, "_%s" % key, None)
out[key] = value
return out
|
[
"def",
"get_params",
"(",
"self",
")",
":",
"import",
"inspect",
"a",
"=",
"inspect",
".",
"getargspec",
"(",
"self",
".",
"__init__",
")",
"[",
"0",
"]",
"out",
"=",
"dict",
"(",
")",
"for",
"key",
"in",
"a",
"[",
"1",
":",
"]",
":",
"value",
"=",
"getattr",
"(",
"self",
",",
"\"_%s\"",
"%",
"key",
",",
"None",
")",
"out",
"[",
"key",
"]",
"=",
"value",
"return",
"out"
] |
Parameters used to initialize the class
|
[
"Parameters",
"used",
"to",
"initialize",
"the",
"class"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L137-L145
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.signature
|
def signature(self):
"Instance file name"
kw = self.get_params()
keys = sorted(kw.keys())
l = []
for k in keys:
n = k[0] + k[-1]
v = kw[k]
if k == 'function_set':
v = "_".join([x.__name__[0] +
x.__name__[-1] +
str(x.nargs) for x in kw[k]])
elif k == 'population_class':
v = kw[k].__name__
else:
v = str(v)
l.append('{0}_{1}'.format(n, v))
return '-'.join(l)
|
python
|
def signature(self):
"Instance file name"
kw = self.get_params()
keys = sorted(kw.keys())
l = []
for k in keys:
n = k[0] + k[-1]
v = kw[k]
if k == 'function_set':
v = "_".join([x.__name__[0] +
x.__name__[-1] +
str(x.nargs) for x in kw[k]])
elif k == 'population_class':
v = kw[k].__name__
else:
v = str(v)
l.append('{0}_{1}'.format(n, v))
return '-'.join(l)
|
[
"def",
"signature",
"(",
"self",
")",
":",
"kw",
"=",
"self",
".",
"get_params",
"(",
")",
"keys",
"=",
"sorted",
"(",
"kw",
".",
"keys",
"(",
")",
")",
"l",
"=",
"[",
"]",
"for",
"k",
"in",
"keys",
":",
"n",
"=",
"k",
"[",
"0",
"]",
"+",
"k",
"[",
"-",
"1",
"]",
"v",
"=",
"kw",
"[",
"k",
"]",
"if",
"k",
"==",
"'function_set'",
":",
"v",
"=",
"\"_\"",
".",
"join",
"(",
"[",
"x",
".",
"__name__",
"[",
"0",
"]",
"+",
"x",
".",
"__name__",
"[",
"-",
"1",
"]",
"+",
"str",
"(",
"x",
".",
"nargs",
")",
"for",
"x",
"in",
"kw",
"[",
"k",
"]",
"]",
")",
"elif",
"k",
"==",
"'population_class'",
":",
"v",
"=",
"kw",
"[",
"k",
"]",
".",
"__name__",
"else",
":",
"v",
"=",
"str",
"(",
"v",
")",
"l",
".",
"append",
"(",
"'{0}_{1}'",
".",
"format",
"(",
"n",
",",
"v",
")",
")",
"return",
"'-'",
".",
"join",
"(",
"l",
")"
] |
Instance file name
|
[
"Instance",
"file",
"name"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L157-L174
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.population
|
def population(self):
"Class containing the population and all the individuals generated"
try:
return self._p
except AttributeError:
self._p = self._population_class(base=self,
tournament_size=self._tournament_size,
classifier=self.classifier,
labels=self._labels,
es_extra_test=self.es_extra_test,
popsize=self._popsize,
random_generations=self._random_generations,
negative_selection=self._negative_selection)
return self._p
|
python
|
def population(self):
"Class containing the population and all the individuals generated"
try:
return self._p
except AttributeError:
self._p = self._population_class(base=self,
tournament_size=self._tournament_size,
classifier=self.classifier,
labels=self._labels,
es_extra_test=self.es_extra_test,
popsize=self._popsize,
random_generations=self._random_generations,
negative_selection=self._negative_selection)
return self._p
|
[
"def",
"population",
"(",
"self",
")",
":",
"try",
":",
"return",
"self",
".",
"_p",
"except",
"AttributeError",
":",
"self",
".",
"_p",
"=",
"self",
".",
"_population_class",
"(",
"base",
"=",
"self",
",",
"tournament_size",
"=",
"self",
".",
"_tournament_size",
",",
"classifier",
"=",
"self",
".",
"classifier",
",",
"labels",
"=",
"self",
".",
"_labels",
",",
"es_extra_test",
"=",
"self",
".",
"es_extra_test",
",",
"popsize",
"=",
"self",
".",
"_popsize",
",",
"random_generations",
"=",
"self",
".",
"_random_generations",
",",
"negative_selection",
"=",
"self",
".",
"_negative_selection",
")",
"return",
"self",
".",
"_p"
] |
Class containing the population and all the individuals generated
|
[
"Class",
"containing",
"the",
"population",
"and",
"all",
"the",
"individuals",
"generated"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L262-L275
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.random_leaf
|
def random_leaf(self):
"Returns a random variable with the associated weight"
for i in range(self._number_tries_feasible_ind):
var = np.random.randint(self.nvar)
v = self._random_leaf(var)
if v is None:
continue
return v
raise RuntimeError("Could not find a suitable random leaf")
|
python
|
def random_leaf(self):
"Returns a random variable with the associated weight"
for i in range(self._number_tries_feasible_ind):
var = np.random.randint(self.nvar)
v = self._random_leaf(var)
if v is None:
continue
return v
raise RuntimeError("Could not find a suitable random leaf")
|
[
"def",
"random_leaf",
"(",
"self",
")",
":",
"for",
"i",
"in",
"range",
"(",
"self",
".",
"_number_tries_feasible_ind",
")",
":",
"var",
"=",
"np",
".",
"random",
".",
"randint",
"(",
"self",
".",
"nvar",
")",
"v",
"=",
"self",
".",
"_random_leaf",
"(",
"var",
")",
"if",
"v",
"is",
"None",
":",
"continue",
"return",
"v",
"raise",
"RuntimeError",
"(",
"\"Could not find a suitable random leaf\"",
")"
] |
Returns a random variable with the associated weight
|
[
"Returns",
"a",
"random",
"variable",
"with",
"the",
"associated",
"weight"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L300-L308
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.random_offspring
|
def random_offspring(self):
"Returns an offspring with the associated weight(s)"
function_set = self.function_set
function_selection = self._function_selection_ins
function_selection.density = self.population.density
function_selection.unfeasible_functions.clear()
for i in range(self._number_tries_feasible_ind):
if self._function_selection:
func_index = function_selection.tournament()
else:
func_index = function_selection.random_function()
func = function_set[func_index]
args = self.get_args(func)
if args is None:
continue
args = [self.population.population[x].position for x in args]
f = self._random_offspring(func, args)
if f is None:
function_selection.unfeasible_functions.add(func_index)
continue
function_selection[func_index] = f.fitness
return f
raise RuntimeError("Could not find a suitable random offpsring")
|
python
|
def random_offspring(self):
"Returns an offspring with the associated weight(s)"
function_set = self.function_set
function_selection = self._function_selection_ins
function_selection.density = self.population.density
function_selection.unfeasible_functions.clear()
for i in range(self._number_tries_feasible_ind):
if self._function_selection:
func_index = function_selection.tournament()
else:
func_index = function_selection.random_function()
func = function_set[func_index]
args = self.get_args(func)
if args is None:
continue
args = [self.population.population[x].position for x in args]
f = self._random_offspring(func, args)
if f is None:
function_selection.unfeasible_functions.add(func_index)
continue
function_selection[func_index] = f.fitness
return f
raise RuntimeError("Could not find a suitable random offpsring")
|
[
"def",
"random_offspring",
"(",
"self",
")",
":",
"function_set",
"=",
"self",
".",
"function_set",
"function_selection",
"=",
"self",
".",
"_function_selection_ins",
"function_selection",
".",
"density",
"=",
"self",
".",
"population",
".",
"density",
"function_selection",
".",
"unfeasible_functions",
".",
"clear",
"(",
")",
"for",
"i",
"in",
"range",
"(",
"self",
".",
"_number_tries_feasible_ind",
")",
":",
"if",
"self",
".",
"_function_selection",
":",
"func_index",
"=",
"function_selection",
".",
"tournament",
"(",
")",
"else",
":",
"func_index",
"=",
"function_selection",
".",
"random_function",
"(",
")",
"func",
"=",
"function_set",
"[",
"func_index",
"]",
"args",
"=",
"self",
".",
"get_args",
"(",
"func",
")",
"if",
"args",
"is",
"None",
":",
"continue",
"args",
"=",
"[",
"self",
".",
"population",
".",
"population",
"[",
"x",
"]",
".",
"position",
"for",
"x",
"in",
"args",
"]",
"f",
"=",
"self",
".",
"_random_offspring",
"(",
"func",
",",
"args",
")",
"if",
"f",
"is",
"None",
":",
"function_selection",
".",
"unfeasible_functions",
".",
"add",
"(",
"func_index",
")",
"continue",
"function_selection",
"[",
"func_index",
"]",
"=",
"f",
".",
"fitness",
"return",
"f",
"raise",
"RuntimeError",
"(",
"\"Could not find a suitable random offpsring\"",
")"
] |
Returns an offspring with the associated weight(s)
|
[
"Returns",
"an",
"offspring",
"with",
"the",
"associated",
"weight",
"(",
"s",
")"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L454-L476
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.stopping_criteria
|
def stopping_criteria(self):
"Test whether the stopping criteria has been achieved."
if self.stopping_criteria_tl():
return True
if self.generations < np.inf:
inds = self.popsize * self.generations
flag = inds <= len(self.population.hist)
else:
flag = False
if flag:
return True
est = self.population.estopping
if self._tr_fraction < 1:
if est is not None and est.fitness_vs == 0:
return True
esr = self._early_stopping_rounds
if self._tr_fraction < 1 and esr is not None and est is not None:
position = self.population.estopping.position
if position < self.init_popsize:
position = self.init_popsize
return (len(self.population.hist) +
self._unfeasible_counter -
position) > esr
return flag
|
python
|
def stopping_criteria(self):
"Test whether the stopping criteria has been achieved."
if self.stopping_criteria_tl():
return True
if self.generations < np.inf:
inds = self.popsize * self.generations
flag = inds <= len(self.population.hist)
else:
flag = False
if flag:
return True
est = self.population.estopping
if self._tr_fraction < 1:
if est is not None and est.fitness_vs == 0:
return True
esr = self._early_stopping_rounds
if self._tr_fraction < 1 and esr is not None and est is not None:
position = self.population.estopping.position
if position < self.init_popsize:
position = self.init_popsize
return (len(self.population.hist) +
self._unfeasible_counter -
position) > esr
return flag
|
[
"def",
"stopping_criteria",
"(",
"self",
")",
":",
"if",
"self",
".",
"stopping_criteria_tl",
"(",
")",
":",
"return",
"True",
"if",
"self",
".",
"generations",
"<",
"np",
".",
"inf",
":",
"inds",
"=",
"self",
".",
"popsize",
"*",
"self",
".",
"generations",
"flag",
"=",
"inds",
"<=",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
"else",
":",
"flag",
"=",
"False",
"if",
"flag",
":",
"return",
"True",
"est",
"=",
"self",
".",
"population",
".",
"estopping",
"if",
"self",
".",
"_tr_fraction",
"<",
"1",
":",
"if",
"est",
"is",
"not",
"None",
"and",
"est",
".",
"fitness_vs",
"==",
"0",
":",
"return",
"True",
"esr",
"=",
"self",
".",
"_early_stopping_rounds",
"if",
"self",
".",
"_tr_fraction",
"<",
"1",
"and",
"esr",
"is",
"not",
"None",
"and",
"est",
"is",
"not",
"None",
":",
"position",
"=",
"self",
".",
"population",
".",
"estopping",
".",
"position",
"if",
"position",
"<",
"self",
".",
"init_popsize",
":",
"position",
"=",
"self",
".",
"init_popsize",
"return",
"(",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
"+",
"self",
".",
"_unfeasible_counter",
"-",
"position",
")",
">",
"esr",
"return",
"flag"
] |
Test whether the stopping criteria has been achieved.
|
[
"Test",
"whether",
"the",
"stopping",
"criteria",
"has",
"been",
"achieved",
"."
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L488-L511
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.nclasses
|
def nclasses(self, v):
"Number of classes of v, also sets the labes"
if not self.classifier:
return 0
if isinstance(v, list):
self._labels = np.arange(len(v))
return
if not isinstance(v, np.ndarray):
v = tonparray(v)
self._labels = np.unique(v)
return self._labels.shape[0]
|
python
|
def nclasses(self, v):
"Number of classes of v, also sets the labes"
if not self.classifier:
return 0
if isinstance(v, list):
self._labels = np.arange(len(v))
return
if not isinstance(v, np.ndarray):
v = tonparray(v)
self._labels = np.unique(v)
return self._labels.shape[0]
|
[
"def",
"nclasses",
"(",
"self",
",",
"v",
")",
":",
"if",
"not",
"self",
".",
"classifier",
":",
"return",
"0",
"if",
"isinstance",
"(",
"v",
",",
"list",
")",
":",
"self",
".",
"_labels",
"=",
"np",
".",
"arange",
"(",
"len",
"(",
"v",
")",
")",
"return",
"if",
"not",
"isinstance",
"(",
"v",
",",
"np",
".",
"ndarray",
")",
":",
"v",
"=",
"tonparray",
"(",
"v",
")",
"self",
".",
"_labels",
"=",
"np",
".",
"unique",
"(",
"v",
")",
"return",
"self",
".",
"_labels",
".",
"shape",
"[",
"0",
"]"
] |
Number of classes of v, also sets the labes
|
[
"Number",
"of",
"classes",
"of",
"v",
"also",
"sets",
"the",
"labes"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L513-L523
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.fit
|
def fit(self, X, y, test_set=None):
"""Evolutive process"""
self._init_time = time.time()
self.X = X
if self._popsize == "nvar":
self._popsize = self.nvar + len(self._input_functions)
if isinstance(test_set, str) and test_set == 'shuffle':
test_set = self.shuffle_tr2ts()
nclasses = self.nclasses(y)
if self.classifier and self._multiple_outputs:
pass
elif nclasses > 2:
assert False
self._multiclass = True
return self.multiclass(X, y, test_set=test_set)
self.y = y
if test_set is not None:
self.Xtest = test_set
for _ in range(self._number_tries_feasible_ind):
self._logger.info("Starting evolution")
try:
self.create_population()
if self.stopping_criteria_tl():
break
except RuntimeError as err:
self._logger.info("Done evolution (RuntimeError (%s), hist: %s)" % (err, len(self.population.hist)))
return self
self._logger.info("Population created (hist: %s)" % len(self.population.hist))
if len(self.population.hist) >= self._tournament_size:
break
if len(self.population.hist) == 0:
raise RuntimeError("Could not find a suitable individual")
if len(self.population.hist) < self._tournament_size:
self._logger.info("Done evolution (hist: %s)" % len(self.population.hist))
return self
if self._remove_raw_inputs:
for x in range(self.nvar):
self._X[x] = None
while not self.stopping_criteria():
try:
a = self.random_offspring()
except RuntimeError as err:
self._logger.info("Done evolution (RuntimeError (%s), hist: %s)" % (err, len(self.population.hist)))
return self
self.replace(a)
self._logger.info("Done evolution (hist: %s)" % len(self.population.hist))
return self
|
python
|
def fit(self, X, y, test_set=None):
"""Evolutive process"""
self._init_time = time.time()
self.X = X
if self._popsize == "nvar":
self._popsize = self.nvar + len(self._input_functions)
if isinstance(test_set, str) and test_set == 'shuffle':
test_set = self.shuffle_tr2ts()
nclasses = self.nclasses(y)
if self.classifier and self._multiple_outputs:
pass
elif nclasses > 2:
assert False
self._multiclass = True
return self.multiclass(X, y, test_set=test_set)
self.y = y
if test_set is not None:
self.Xtest = test_set
for _ in range(self._number_tries_feasible_ind):
self._logger.info("Starting evolution")
try:
self.create_population()
if self.stopping_criteria_tl():
break
except RuntimeError as err:
self._logger.info("Done evolution (RuntimeError (%s), hist: %s)" % (err, len(self.population.hist)))
return self
self._logger.info("Population created (hist: %s)" % len(self.population.hist))
if len(self.population.hist) >= self._tournament_size:
break
if len(self.population.hist) == 0:
raise RuntimeError("Could not find a suitable individual")
if len(self.population.hist) < self._tournament_size:
self._logger.info("Done evolution (hist: %s)" % len(self.population.hist))
return self
if self._remove_raw_inputs:
for x in range(self.nvar):
self._X[x] = None
while not self.stopping_criteria():
try:
a = self.random_offspring()
except RuntimeError as err:
self._logger.info("Done evolution (RuntimeError (%s), hist: %s)" % (err, len(self.population.hist)))
return self
self.replace(a)
self._logger.info("Done evolution (hist: %s)" % len(self.population.hist))
return self
|
[
"def",
"fit",
"(",
"self",
",",
"X",
",",
"y",
",",
"test_set",
"=",
"None",
")",
":",
"self",
".",
"_init_time",
"=",
"time",
".",
"time",
"(",
")",
"self",
".",
"X",
"=",
"X",
"if",
"self",
".",
"_popsize",
"==",
"\"nvar\"",
":",
"self",
".",
"_popsize",
"=",
"self",
".",
"nvar",
"+",
"len",
"(",
"self",
".",
"_input_functions",
")",
"if",
"isinstance",
"(",
"test_set",
",",
"str",
")",
"and",
"test_set",
"==",
"'shuffle'",
":",
"test_set",
"=",
"self",
".",
"shuffle_tr2ts",
"(",
")",
"nclasses",
"=",
"self",
".",
"nclasses",
"(",
"y",
")",
"if",
"self",
".",
"classifier",
"and",
"self",
".",
"_multiple_outputs",
":",
"pass",
"elif",
"nclasses",
">",
"2",
":",
"assert",
"False",
"self",
".",
"_multiclass",
"=",
"True",
"return",
"self",
".",
"multiclass",
"(",
"X",
",",
"y",
",",
"test_set",
"=",
"test_set",
")",
"self",
".",
"y",
"=",
"y",
"if",
"test_set",
"is",
"not",
"None",
":",
"self",
".",
"Xtest",
"=",
"test_set",
"for",
"_",
"in",
"range",
"(",
"self",
".",
"_number_tries_feasible_ind",
")",
":",
"self",
".",
"_logger",
".",
"info",
"(",
"\"Starting evolution\"",
")",
"try",
":",
"self",
".",
"create_population",
"(",
")",
"if",
"self",
".",
"stopping_criteria_tl",
"(",
")",
":",
"break",
"except",
"RuntimeError",
"as",
"err",
":",
"self",
".",
"_logger",
".",
"info",
"(",
"\"Done evolution (RuntimeError (%s), hist: %s)\"",
"%",
"(",
"err",
",",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
")",
")",
"return",
"self",
"self",
".",
"_logger",
".",
"info",
"(",
"\"Population created (hist: %s)\"",
"%",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
")",
"if",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
">=",
"self",
".",
"_tournament_size",
":",
"break",
"if",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
"==",
"0",
":",
"raise",
"RuntimeError",
"(",
"\"Could not find a suitable individual\"",
")",
"if",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
"<",
"self",
".",
"_tournament_size",
":",
"self",
".",
"_logger",
".",
"info",
"(",
"\"Done evolution (hist: %s)\"",
"%",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
")",
"return",
"self",
"if",
"self",
".",
"_remove_raw_inputs",
":",
"for",
"x",
"in",
"range",
"(",
"self",
".",
"nvar",
")",
":",
"self",
".",
"_X",
"[",
"x",
"]",
"=",
"None",
"while",
"not",
"self",
".",
"stopping_criteria",
"(",
")",
":",
"try",
":",
"a",
"=",
"self",
".",
"random_offspring",
"(",
")",
"except",
"RuntimeError",
"as",
"err",
":",
"self",
".",
"_logger",
".",
"info",
"(",
"\"Done evolution (RuntimeError (%s), hist: %s)\"",
"%",
"(",
"err",
",",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
")",
")",
"return",
"self",
"self",
".",
"replace",
"(",
"a",
")",
"self",
".",
"_logger",
".",
"info",
"(",
"\"Done evolution (hist: %s)\"",
"%",
"len",
"(",
"self",
".",
"population",
".",
"hist",
")",
")",
"return",
"self"
] |
Evolutive process
|
[
"Evolutive",
"process"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L537-L583
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.decision_function
|
def decision_function(self, v=None, X=None):
"Decision function i.e. the raw data of the prediction"
m = self.model(v=v)
return m.decision_function(X)
|
python
|
def decision_function(self, v=None, X=None):
"Decision function i.e. the raw data of the prediction"
m = self.model(v=v)
return m.decision_function(X)
|
[
"def",
"decision_function",
"(",
"self",
",",
"v",
"=",
"None",
",",
"X",
"=",
"None",
")",
":",
"m",
"=",
"self",
".",
"model",
"(",
"v",
"=",
"v",
")",
"return",
"m",
".",
"decision_function",
"(",
"X",
")"
] |
Decision function i.e. the raw data of the prediction
|
[
"Decision",
"function",
"i",
".",
"e",
".",
"the",
"raw",
"data",
"of",
"the",
"prediction"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L593-L596
|
train
|
mgraffg/EvoDAG
|
EvoDAG/base.py
|
EvoDAG.predict
|
def predict(self, v=None, X=None):
"""In classification this returns the classes, in
regression it is equivalent to the decision function"""
if X is None:
X = v
v = None
m = self.model(v=v)
return m.predict(X)
|
python
|
def predict(self, v=None, X=None):
"""In classification this returns the classes, in
regression it is equivalent to the decision function"""
if X is None:
X = v
v = None
m = self.model(v=v)
return m.predict(X)
|
[
"def",
"predict",
"(",
"self",
",",
"v",
"=",
"None",
",",
"X",
"=",
"None",
")",
":",
"if",
"X",
"is",
"None",
":",
"X",
"=",
"v",
"v",
"=",
"None",
"m",
"=",
"self",
".",
"model",
"(",
"v",
"=",
"v",
")",
"return",
"m",
".",
"predict",
"(",
"X",
")"
] |
In classification this returns the classes, in
regression it is equivalent to the decision function
|
[
"In",
"classification",
"this",
"returns",
"the",
"classes",
"in",
"regression",
"it",
"is",
"equivalent",
"to",
"the",
"decision",
"function"
] |
e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5
|
https://github.com/mgraffg/EvoDAG/blob/e11fa1fd1ca9e69cca92696c86661a3dc7b3a1d5/EvoDAG/base.py#L598-L605
|
train
|
marrow/WebCore
|
web/server/gevent_.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080):
"""Gevent-based WSGI-HTTP server."""
# Instantiate the server with a host/port configuration and our application.
WSGIServer((host, int(port)), application).serve_forever()
|
python
|
def serve(application, host='127.0.0.1', port=8080):
"""Gevent-based WSGI-HTTP server."""
# Instantiate the server with a host/port configuration and our application.
WSGIServer((host, int(port)), application).serve_forever()
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
")",
":",
"# Instantiate the server with a host/port configuration and our application.",
"WSGIServer",
"(",
"(",
"host",
",",
"int",
"(",
"port",
")",
")",
",",
"application",
")",
".",
"serve_forever",
"(",
")"
] |
Gevent-based WSGI-HTTP server.
|
[
"Gevent",
"-",
"based",
"WSGI",
"-",
"HTTP",
"server",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/gevent_.py#L14-L18
|
train
|
bitlabstudio/django-subscribe
|
subscribe/templatetags/subscribe_tags.py
|
get_subscribers
|
def get_subscribers(obj):
"""
Returns the subscribers for a given object.
:param obj: Any object.
"""
ctype = ContentType.objects.get_for_model(obj)
return Subscription.objects.filter(content_type=ctype, object_id=obj.pk)
|
python
|
def get_subscribers(obj):
"""
Returns the subscribers for a given object.
:param obj: Any object.
"""
ctype = ContentType.objects.get_for_model(obj)
return Subscription.objects.filter(content_type=ctype, object_id=obj.pk)
|
[
"def",
"get_subscribers",
"(",
"obj",
")",
":",
"ctype",
"=",
"ContentType",
".",
"objects",
".",
"get_for_model",
"(",
"obj",
")",
"return",
"Subscription",
".",
"objects",
".",
"filter",
"(",
"content_type",
"=",
"ctype",
",",
"object_id",
"=",
"obj",
".",
"pk",
")"
] |
Returns the subscribers for a given object.
:param obj: Any object.
|
[
"Returns",
"the",
"subscribers",
"for",
"a",
"given",
"object",
"."
] |
313de63fb4acda172e88b65c3327c793f98e8aa9
|
https://github.com/bitlabstudio/django-subscribe/blob/313de63fb4acda172e88b65c3327c793f98e8aa9/subscribe/templatetags/subscribe_tags.py#L23-L31
|
train
|
bitlabstudio/django-subscribe
|
subscribe/templatetags/subscribe_tags.py
|
is_subscribed
|
def is_subscribed(user, obj):
"""
Returns ``True`` if the user is subscribed to the given object.
:param user: A ``User`` instance.
:param obj: Any object.
"""
if not user.is_authenticated():
return False
ctype = ContentType.objects.get_for_model(obj)
try:
Subscription.objects.get(
user=user, content_type=ctype, object_id=obj.pk)
except Subscription.DoesNotExist:
return False
return True
|
python
|
def is_subscribed(user, obj):
"""
Returns ``True`` if the user is subscribed to the given object.
:param user: A ``User`` instance.
:param obj: Any object.
"""
if not user.is_authenticated():
return False
ctype = ContentType.objects.get_for_model(obj)
try:
Subscription.objects.get(
user=user, content_type=ctype, object_id=obj.pk)
except Subscription.DoesNotExist:
return False
return True
|
[
"def",
"is_subscribed",
"(",
"user",
",",
"obj",
")",
":",
"if",
"not",
"user",
".",
"is_authenticated",
"(",
")",
":",
"return",
"False",
"ctype",
"=",
"ContentType",
".",
"objects",
".",
"get_for_model",
"(",
"obj",
")",
"try",
":",
"Subscription",
".",
"objects",
".",
"get",
"(",
"user",
"=",
"user",
",",
"content_type",
"=",
"ctype",
",",
"object_id",
"=",
"obj",
".",
"pk",
")",
"except",
"Subscription",
".",
"DoesNotExist",
":",
"return",
"False",
"return",
"True"
] |
Returns ``True`` if the user is subscribed to the given object.
:param user: A ``User`` instance.
:param obj: Any object.
|
[
"Returns",
"True",
"if",
"the",
"user",
"is",
"subscribed",
"to",
"the",
"given",
"object",
"."
] |
313de63fb4acda172e88b65c3327c793f98e8aa9
|
https://github.com/bitlabstudio/django-subscribe/blob/313de63fb4acda172e88b65c3327c793f98e8aa9/subscribe/templatetags/subscribe_tags.py#L35-L54
|
train
|
marrow/WebCore
|
web/core/context.py
|
Context._promote
|
def _promote(self, name, instantiate=True):
"""Create a new subclass of Context which incorporates instance attributes and new descriptors.
This promotes an instance and its instance attributes up to being a class with class attributes, then
returns an instance of that class.
"""
metaclass = type(self.__class__)
contents = self.__dict__.copy()
cls = metaclass(str(name), (self.__class__, ), contents)
if instantiate:
return cls()
return cls
|
python
|
def _promote(self, name, instantiate=True):
"""Create a new subclass of Context which incorporates instance attributes and new descriptors.
This promotes an instance and its instance attributes up to being a class with class attributes, then
returns an instance of that class.
"""
metaclass = type(self.__class__)
contents = self.__dict__.copy()
cls = metaclass(str(name), (self.__class__, ), contents)
if instantiate:
return cls()
return cls
|
[
"def",
"_promote",
"(",
"self",
",",
"name",
",",
"instantiate",
"=",
"True",
")",
":",
"metaclass",
"=",
"type",
"(",
"self",
".",
"__class__",
")",
"contents",
"=",
"self",
".",
"__dict__",
".",
"copy",
"(",
")",
"cls",
"=",
"metaclass",
"(",
"str",
"(",
"name",
")",
",",
"(",
"self",
".",
"__class__",
",",
")",
",",
"contents",
")",
"if",
"instantiate",
":",
"return",
"cls",
"(",
")",
"return",
"cls"
] |
Create a new subclass of Context which incorporates instance attributes and new descriptors.
This promotes an instance and its instance attributes up to being a class with class attributes, then
returns an instance of that class.
|
[
"Create",
"a",
"new",
"subclass",
"of",
"Context",
"which",
"incorporates",
"instance",
"attributes",
"and",
"new",
"descriptors",
".",
"This",
"promotes",
"an",
"instance",
"and",
"its",
"instance",
"attributes",
"up",
"to",
"being",
"a",
"class",
"with",
"class",
"attributes",
"then",
"returns",
"an",
"instance",
"of",
"that",
"class",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/core/context.py#L23-L37
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/tune/NeuroMLController.py
|
run_individual
|
def run_individual(sim_var,
reference,
neuroml_file,
nml_doc,
still_included,
generate_dir,
target,
sim_time,
dt,
simulator,
cleanup = True,
show=False):
"""
Run an individual simulation.
The candidate data has been flattened into the sim_var dict. The
sim_var dict contains parameter:value key value pairs, which are
applied to the model before it is simulated.
"""
for var_name in sim_var.keys():
individual_var_names = var_name.split('+')
for individual_var_name in individual_var_names:
words = individual_var_name.split('/')
type, id1 = words[0].split(':')
if ':' in words[1]:
variable, id2 = words[1].split(':')
else:
variable = words[1]
id2 = None
units = words[2]
value = sim_var[var_name]
pyneuroml.pynml.print_comment_v(' Changing value of %s (%s) in %s (%s) to: %s %s'%(variable, id2, type, id1, value, units))
if type == 'channel':
channel = nml_doc.get_by_id(id1)
if channel:
print("Setting channel %s"%(channel))
if variable == 'vShift':
channel.v_shift = '%s %s'%(value, units)
else:
pyneuroml.pynml.print_comment_v('Could not find channel with id %s from expression: %s'%(id1, individual_var_name))
exit()
elif type == 'cell':
cell = None
for c in nml_doc.cells:
if c.id == id1:
cell = c
if variable == 'channelDensity':
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_densities + cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.id == id2:
chanDens = cd
chanDens.cond_density = '%s %s'%(value, units)
elif variable == 'vShift_channelDensity':
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.id == id2:
chanDens = cd
chanDens.v_shift = '%s %s'%(value, units)
elif variable == 'channelDensityNernst':
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_density_nernsts:
if cd.id == id2:
chanDens = cd
chanDens.cond_density = '%s %s'%(value, units)
elif variable == 'erev_id': # change all values of erev in channelDensity elements with only this id
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_densities + cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.id == id2:
chanDens = cd
chanDens.erev = '%s %s'%(value, units)
elif variable == 'erev_ion': # change all values of erev in channelDensity elements with this ion
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_densities + cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.ion == id2:
chanDens = cd
chanDens.erev = '%s %s'%(value, units)
elif variable == 'specificCapacitance':
specCap = None
for sc in cell.biophysical_properties.membrane_properties.specific_capacitances:
if (sc.segment_groups == None and id2 == 'all') or sc.segment_groups == id2 :
specCap = sc
specCap.value = '%s %s'%(value, units)
elif variable == 'resistivity':
resistivity = None
for rs in cell.biophysical_properties.intracellular_properties.resistivities:
if (rs.segment_groups == None and id2 == 'all') or rs.segment_groups == id2 :
resistivity = rs
resistivity.value = '%s %s'%(value, units)
else:
pyneuroml.pynml.print_comment_v('Unknown variable (%s) in variable expression: %s'%(variable, individual_var_name))
exit()
elif type == 'izhikevich2007Cell':
izhcell = None
for c in nml_doc.izhikevich2007_cells:
if c.id == id1:
izhcell = c
izhcell.__setattr__(variable, '%s %s'%(value, units))
else:
pyneuroml.pynml.print_comment_v('Unknown type (%s) in variable expression: %s'%(type, individual_var_name))
new_neuroml_file = '%s/%s'%(generate_dir,os.path.basename(neuroml_file))
if new_neuroml_file == neuroml_file:
pyneuroml.pynml.print_comment_v('Cannot use a directory for generating into (%s) which is the same location of the NeuroML file (%s)!'% \
(neuroml_file, generate_dir))
pyneuroml.pynml.write_neuroml2_file(nml_doc, new_neuroml_file)
for include in still_included:
inc_loc = '%s/%s'%(os.path.dirname(os.path.abspath(neuroml_file)),include)
pyneuroml.pynml.print_comment_v("Copying non included file %s to %s (%s) beside %s"%(inc_loc, generate_dir,os.path.abspath(generate_dir), new_neuroml_file))
shutil.copy(inc_loc, generate_dir)
from pyneuroml.tune.NeuroMLSimulation import NeuroMLSimulation
sim = NeuroMLSimulation(reference,
neuroml_file = new_neuroml_file,
target = target,
sim_time = sim_time,
dt = dt,
simulator = simulator,
generate_dir = generate_dir,
cleanup = cleanup,
nml_doc = nml_doc)
sim.go()
if show:
sim.show()
return sim.t, sim.volts
|
python
|
def run_individual(sim_var,
reference,
neuroml_file,
nml_doc,
still_included,
generate_dir,
target,
sim_time,
dt,
simulator,
cleanup = True,
show=False):
"""
Run an individual simulation.
The candidate data has been flattened into the sim_var dict. The
sim_var dict contains parameter:value key value pairs, which are
applied to the model before it is simulated.
"""
for var_name in sim_var.keys():
individual_var_names = var_name.split('+')
for individual_var_name in individual_var_names:
words = individual_var_name.split('/')
type, id1 = words[0].split(':')
if ':' in words[1]:
variable, id2 = words[1].split(':')
else:
variable = words[1]
id2 = None
units = words[2]
value = sim_var[var_name]
pyneuroml.pynml.print_comment_v(' Changing value of %s (%s) in %s (%s) to: %s %s'%(variable, id2, type, id1, value, units))
if type == 'channel':
channel = nml_doc.get_by_id(id1)
if channel:
print("Setting channel %s"%(channel))
if variable == 'vShift':
channel.v_shift = '%s %s'%(value, units)
else:
pyneuroml.pynml.print_comment_v('Could not find channel with id %s from expression: %s'%(id1, individual_var_name))
exit()
elif type == 'cell':
cell = None
for c in nml_doc.cells:
if c.id == id1:
cell = c
if variable == 'channelDensity':
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_densities + cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.id == id2:
chanDens = cd
chanDens.cond_density = '%s %s'%(value, units)
elif variable == 'vShift_channelDensity':
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.id == id2:
chanDens = cd
chanDens.v_shift = '%s %s'%(value, units)
elif variable == 'channelDensityNernst':
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_density_nernsts:
if cd.id == id2:
chanDens = cd
chanDens.cond_density = '%s %s'%(value, units)
elif variable == 'erev_id': # change all values of erev in channelDensity elements with only this id
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_densities + cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.id == id2:
chanDens = cd
chanDens.erev = '%s %s'%(value, units)
elif variable == 'erev_ion': # change all values of erev in channelDensity elements with this ion
chanDens = None
for cd in cell.biophysical_properties.membrane_properties.channel_densities + cell.biophysical_properties.membrane_properties.channel_density_v_shifts:
if cd.ion == id2:
chanDens = cd
chanDens.erev = '%s %s'%(value, units)
elif variable == 'specificCapacitance':
specCap = None
for sc in cell.biophysical_properties.membrane_properties.specific_capacitances:
if (sc.segment_groups == None and id2 == 'all') or sc.segment_groups == id2 :
specCap = sc
specCap.value = '%s %s'%(value, units)
elif variable == 'resistivity':
resistivity = None
for rs in cell.biophysical_properties.intracellular_properties.resistivities:
if (rs.segment_groups == None and id2 == 'all') or rs.segment_groups == id2 :
resistivity = rs
resistivity.value = '%s %s'%(value, units)
else:
pyneuroml.pynml.print_comment_v('Unknown variable (%s) in variable expression: %s'%(variable, individual_var_name))
exit()
elif type == 'izhikevich2007Cell':
izhcell = None
for c in nml_doc.izhikevich2007_cells:
if c.id == id1:
izhcell = c
izhcell.__setattr__(variable, '%s %s'%(value, units))
else:
pyneuroml.pynml.print_comment_v('Unknown type (%s) in variable expression: %s'%(type, individual_var_name))
new_neuroml_file = '%s/%s'%(generate_dir,os.path.basename(neuroml_file))
if new_neuroml_file == neuroml_file:
pyneuroml.pynml.print_comment_v('Cannot use a directory for generating into (%s) which is the same location of the NeuroML file (%s)!'% \
(neuroml_file, generate_dir))
pyneuroml.pynml.write_neuroml2_file(nml_doc, new_neuroml_file)
for include in still_included:
inc_loc = '%s/%s'%(os.path.dirname(os.path.abspath(neuroml_file)),include)
pyneuroml.pynml.print_comment_v("Copying non included file %s to %s (%s) beside %s"%(inc_loc, generate_dir,os.path.abspath(generate_dir), new_neuroml_file))
shutil.copy(inc_loc, generate_dir)
from pyneuroml.tune.NeuroMLSimulation import NeuroMLSimulation
sim = NeuroMLSimulation(reference,
neuroml_file = new_neuroml_file,
target = target,
sim_time = sim_time,
dt = dt,
simulator = simulator,
generate_dir = generate_dir,
cleanup = cleanup,
nml_doc = nml_doc)
sim.go()
if show:
sim.show()
return sim.t, sim.volts
|
[
"def",
"run_individual",
"(",
"sim_var",
",",
"reference",
",",
"neuroml_file",
",",
"nml_doc",
",",
"still_included",
",",
"generate_dir",
",",
"target",
",",
"sim_time",
",",
"dt",
",",
"simulator",
",",
"cleanup",
"=",
"True",
",",
"show",
"=",
"False",
")",
":",
"for",
"var_name",
"in",
"sim_var",
".",
"keys",
"(",
")",
":",
"individual_var_names",
"=",
"var_name",
".",
"split",
"(",
"'+'",
")",
"for",
"individual_var_name",
"in",
"individual_var_names",
":",
"words",
"=",
"individual_var_name",
".",
"split",
"(",
"'/'",
")",
"type",
",",
"id1",
"=",
"words",
"[",
"0",
"]",
".",
"split",
"(",
"':'",
")",
"if",
"':'",
"in",
"words",
"[",
"1",
"]",
":",
"variable",
",",
"id2",
"=",
"words",
"[",
"1",
"]",
".",
"split",
"(",
"':'",
")",
"else",
":",
"variable",
"=",
"words",
"[",
"1",
"]",
"id2",
"=",
"None",
"units",
"=",
"words",
"[",
"2",
"]",
"value",
"=",
"sim_var",
"[",
"var_name",
"]",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"' Changing value of %s (%s) in %s (%s) to: %s %s'",
"%",
"(",
"variable",
",",
"id2",
",",
"type",
",",
"id1",
",",
"value",
",",
"units",
")",
")",
"if",
"type",
"==",
"'channel'",
":",
"channel",
"=",
"nml_doc",
".",
"get_by_id",
"(",
"id1",
")",
"if",
"channel",
":",
"print",
"(",
"\"Setting channel %s\"",
"%",
"(",
"channel",
")",
")",
"if",
"variable",
"==",
"'vShift'",
":",
"channel",
".",
"v_shift",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"else",
":",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'Could not find channel with id %s from expression: %s'",
"%",
"(",
"id1",
",",
"individual_var_name",
")",
")",
"exit",
"(",
")",
"elif",
"type",
"==",
"'cell'",
":",
"cell",
"=",
"None",
"for",
"c",
"in",
"nml_doc",
".",
"cells",
":",
"if",
"c",
".",
"id",
"==",
"id1",
":",
"cell",
"=",
"c",
"if",
"variable",
"==",
"'channelDensity'",
":",
"chanDens",
"=",
"None",
"for",
"cd",
"in",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_densities",
"+",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_density_v_shifts",
":",
"if",
"cd",
".",
"id",
"==",
"id2",
":",
"chanDens",
"=",
"cd",
"chanDens",
".",
"cond_density",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"elif",
"variable",
"==",
"'vShift_channelDensity'",
":",
"chanDens",
"=",
"None",
"for",
"cd",
"in",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_density_v_shifts",
":",
"if",
"cd",
".",
"id",
"==",
"id2",
":",
"chanDens",
"=",
"cd",
"chanDens",
".",
"v_shift",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"elif",
"variable",
"==",
"'channelDensityNernst'",
":",
"chanDens",
"=",
"None",
"for",
"cd",
"in",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_density_nernsts",
":",
"if",
"cd",
".",
"id",
"==",
"id2",
":",
"chanDens",
"=",
"cd",
"chanDens",
".",
"cond_density",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"elif",
"variable",
"==",
"'erev_id'",
":",
"# change all values of erev in channelDensity elements with only this id",
"chanDens",
"=",
"None",
"for",
"cd",
"in",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_densities",
"+",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_density_v_shifts",
":",
"if",
"cd",
".",
"id",
"==",
"id2",
":",
"chanDens",
"=",
"cd",
"chanDens",
".",
"erev",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"elif",
"variable",
"==",
"'erev_ion'",
":",
"# change all values of erev in channelDensity elements with this ion",
"chanDens",
"=",
"None",
"for",
"cd",
"in",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_densities",
"+",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"channel_density_v_shifts",
":",
"if",
"cd",
".",
"ion",
"==",
"id2",
":",
"chanDens",
"=",
"cd",
"chanDens",
".",
"erev",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"elif",
"variable",
"==",
"'specificCapacitance'",
":",
"specCap",
"=",
"None",
"for",
"sc",
"in",
"cell",
".",
"biophysical_properties",
".",
"membrane_properties",
".",
"specific_capacitances",
":",
"if",
"(",
"sc",
".",
"segment_groups",
"==",
"None",
"and",
"id2",
"==",
"'all'",
")",
"or",
"sc",
".",
"segment_groups",
"==",
"id2",
":",
"specCap",
"=",
"sc",
"specCap",
".",
"value",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"elif",
"variable",
"==",
"'resistivity'",
":",
"resistivity",
"=",
"None",
"for",
"rs",
"in",
"cell",
".",
"biophysical_properties",
".",
"intracellular_properties",
".",
"resistivities",
":",
"if",
"(",
"rs",
".",
"segment_groups",
"==",
"None",
"and",
"id2",
"==",
"'all'",
")",
"or",
"rs",
".",
"segment_groups",
"==",
"id2",
":",
"resistivity",
"=",
"rs",
"resistivity",
".",
"value",
"=",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
"else",
":",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'Unknown variable (%s) in variable expression: %s'",
"%",
"(",
"variable",
",",
"individual_var_name",
")",
")",
"exit",
"(",
")",
"elif",
"type",
"==",
"'izhikevich2007Cell'",
":",
"izhcell",
"=",
"None",
"for",
"c",
"in",
"nml_doc",
".",
"izhikevich2007_cells",
":",
"if",
"c",
".",
"id",
"==",
"id1",
":",
"izhcell",
"=",
"c",
"izhcell",
".",
"__setattr__",
"(",
"variable",
",",
"'%s %s'",
"%",
"(",
"value",
",",
"units",
")",
")",
"else",
":",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'Unknown type (%s) in variable expression: %s'",
"%",
"(",
"type",
",",
"individual_var_name",
")",
")",
"new_neuroml_file",
"=",
"'%s/%s'",
"%",
"(",
"generate_dir",
",",
"os",
".",
"path",
".",
"basename",
"(",
"neuroml_file",
")",
")",
"if",
"new_neuroml_file",
"==",
"neuroml_file",
":",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'Cannot use a directory for generating into (%s) which is the same location of the NeuroML file (%s)!'",
"%",
"(",
"neuroml_file",
",",
"generate_dir",
")",
")",
"pyneuroml",
".",
"pynml",
".",
"write_neuroml2_file",
"(",
"nml_doc",
",",
"new_neuroml_file",
")",
"for",
"include",
"in",
"still_included",
":",
"inc_loc",
"=",
"'%s/%s'",
"%",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"neuroml_file",
")",
")",
",",
"include",
")",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"\"Copying non included file %s to %s (%s) beside %s\"",
"%",
"(",
"inc_loc",
",",
"generate_dir",
",",
"os",
".",
"path",
".",
"abspath",
"(",
"generate_dir",
")",
",",
"new_neuroml_file",
")",
")",
"shutil",
".",
"copy",
"(",
"inc_loc",
",",
"generate_dir",
")",
"from",
"pyneuroml",
".",
"tune",
".",
"NeuroMLSimulation",
"import",
"NeuroMLSimulation",
"sim",
"=",
"NeuroMLSimulation",
"(",
"reference",
",",
"neuroml_file",
"=",
"new_neuroml_file",
",",
"target",
"=",
"target",
",",
"sim_time",
"=",
"sim_time",
",",
"dt",
"=",
"dt",
",",
"simulator",
"=",
"simulator",
",",
"generate_dir",
"=",
"generate_dir",
",",
"cleanup",
"=",
"cleanup",
",",
"nml_doc",
"=",
"nml_doc",
")",
"sim",
".",
"go",
"(",
")",
"if",
"show",
":",
"sim",
".",
"show",
"(",
")",
"return",
"sim",
".",
"t",
",",
"sim",
".",
"volts"
] |
Run an individual simulation.
The candidate data has been flattened into the sim_var dict. The
sim_var dict contains parameter:value key value pairs, which are
applied to the model before it is simulated.
|
[
"Run",
"an",
"individual",
"simulation",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/tune/NeuroMLController.py#L154-L323
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/tune/NeuroMLController.py
|
NeuroMLController.run
|
def run(self,candidates,parameters):
"""
Run simulation for each candidate
This run method will loop through each candidate and run the simulation
corresponding to its parameter values. It will populate an array called
traces with the resulting voltage traces for the simulation and return it.
"""
traces = []
start_time = time.time()
if self.num_parallel_evaluations == 1:
for candidate_i in range(len(candidates)):
candidate = candidates[candidate_i]
sim_var = dict(zip(parameters,candidate))
pyneuroml.pynml.print_comment_v('\n\n - RUN %i (%i/%i); variables: %s\n'%(self.count,candidate_i+1,len(candidates),sim_var))
self.count+=1
t,v = self.run_individual(sim_var)
traces.append([t,v])
else:
import pp
ppservers = ()
job_server = pp.Server(self.num_parallel_evaluations, ppservers=ppservers)
pyneuroml.pynml.print_comment_v('Running %i candidates across %i local processes'%(len(candidates),job_server.get_ncpus()))
jobs = []
for candidate_i in range(len(candidates)):
candidate = candidates[candidate_i]
sim_var = dict(zip(parameters,candidate))
pyneuroml.pynml.print_comment_v('\n\n - PARALLEL RUN %i (%i/%i of curr candidates); variables: %s\n'%(self.count,candidate_i+1,len(candidates),sim_var))
self.count+=1
cand_dir = self.generate_dir+"/CANDIDATE_%s"%candidate_i
if not os.path.exists(cand_dir):
os.mkdir(cand_dir)
vars = (sim_var,
self.ref,
self.neuroml_file,
self.nml_doc,
self.still_included,
cand_dir,
self.target,
self.sim_time,
self.dt,
self.simulator)
job = job_server.submit(run_individual, vars, (), ("pyneuroml.pynml",'pyneuroml.tune.NeuroMLSimulation','shutil','neuroml'))
jobs.append(job)
for job_i in range(len(jobs)):
job = jobs[job_i]
pyneuroml.pynml.print_comment_v("Checking parallel job %i/%i; set running so far: %i"%(job_i,len(jobs),self.count))
t,v = job()
traces.append([t,v])
#pyneuroml.pynml.print_comment_v("Obtained: %s"%result)
####job_server.print_stats()
job_server.destroy()
print("-------------------------------------------")
end_time = time.time()
tot = (end_time-start_time)
pyneuroml.pynml.print_comment_v('Ran %i candidates in %s seconds (~%ss per job)'%(len(candidates),tot,tot/len(candidates)))
return traces
|
python
|
def run(self,candidates,parameters):
"""
Run simulation for each candidate
This run method will loop through each candidate and run the simulation
corresponding to its parameter values. It will populate an array called
traces with the resulting voltage traces for the simulation and return it.
"""
traces = []
start_time = time.time()
if self.num_parallel_evaluations == 1:
for candidate_i in range(len(candidates)):
candidate = candidates[candidate_i]
sim_var = dict(zip(parameters,candidate))
pyneuroml.pynml.print_comment_v('\n\n - RUN %i (%i/%i); variables: %s\n'%(self.count,candidate_i+1,len(candidates),sim_var))
self.count+=1
t,v = self.run_individual(sim_var)
traces.append([t,v])
else:
import pp
ppservers = ()
job_server = pp.Server(self.num_parallel_evaluations, ppservers=ppservers)
pyneuroml.pynml.print_comment_v('Running %i candidates across %i local processes'%(len(candidates),job_server.get_ncpus()))
jobs = []
for candidate_i in range(len(candidates)):
candidate = candidates[candidate_i]
sim_var = dict(zip(parameters,candidate))
pyneuroml.pynml.print_comment_v('\n\n - PARALLEL RUN %i (%i/%i of curr candidates); variables: %s\n'%(self.count,candidate_i+1,len(candidates),sim_var))
self.count+=1
cand_dir = self.generate_dir+"/CANDIDATE_%s"%candidate_i
if not os.path.exists(cand_dir):
os.mkdir(cand_dir)
vars = (sim_var,
self.ref,
self.neuroml_file,
self.nml_doc,
self.still_included,
cand_dir,
self.target,
self.sim_time,
self.dt,
self.simulator)
job = job_server.submit(run_individual, vars, (), ("pyneuroml.pynml",'pyneuroml.tune.NeuroMLSimulation','shutil','neuroml'))
jobs.append(job)
for job_i in range(len(jobs)):
job = jobs[job_i]
pyneuroml.pynml.print_comment_v("Checking parallel job %i/%i; set running so far: %i"%(job_i,len(jobs),self.count))
t,v = job()
traces.append([t,v])
#pyneuroml.pynml.print_comment_v("Obtained: %s"%result)
####job_server.print_stats()
job_server.destroy()
print("-------------------------------------------")
end_time = time.time()
tot = (end_time-start_time)
pyneuroml.pynml.print_comment_v('Ran %i candidates in %s seconds (~%ss per job)'%(len(candidates),tot,tot/len(candidates)))
return traces
|
[
"def",
"run",
"(",
"self",
",",
"candidates",
",",
"parameters",
")",
":",
"traces",
"=",
"[",
"]",
"start_time",
"=",
"time",
".",
"time",
"(",
")",
"if",
"self",
".",
"num_parallel_evaluations",
"==",
"1",
":",
"for",
"candidate_i",
"in",
"range",
"(",
"len",
"(",
"candidates",
")",
")",
":",
"candidate",
"=",
"candidates",
"[",
"candidate_i",
"]",
"sim_var",
"=",
"dict",
"(",
"zip",
"(",
"parameters",
",",
"candidate",
")",
")",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'\\n\\n - RUN %i (%i/%i); variables: %s\\n'",
"%",
"(",
"self",
".",
"count",
",",
"candidate_i",
"+",
"1",
",",
"len",
"(",
"candidates",
")",
",",
"sim_var",
")",
")",
"self",
".",
"count",
"+=",
"1",
"t",
",",
"v",
"=",
"self",
".",
"run_individual",
"(",
"sim_var",
")",
"traces",
".",
"append",
"(",
"[",
"t",
",",
"v",
"]",
")",
"else",
":",
"import",
"pp",
"ppservers",
"=",
"(",
")",
"job_server",
"=",
"pp",
".",
"Server",
"(",
"self",
".",
"num_parallel_evaluations",
",",
"ppservers",
"=",
"ppservers",
")",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'Running %i candidates across %i local processes'",
"%",
"(",
"len",
"(",
"candidates",
")",
",",
"job_server",
".",
"get_ncpus",
"(",
")",
")",
")",
"jobs",
"=",
"[",
"]",
"for",
"candidate_i",
"in",
"range",
"(",
"len",
"(",
"candidates",
")",
")",
":",
"candidate",
"=",
"candidates",
"[",
"candidate_i",
"]",
"sim_var",
"=",
"dict",
"(",
"zip",
"(",
"parameters",
",",
"candidate",
")",
")",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'\\n\\n - PARALLEL RUN %i (%i/%i of curr candidates); variables: %s\\n'",
"%",
"(",
"self",
".",
"count",
",",
"candidate_i",
"+",
"1",
",",
"len",
"(",
"candidates",
")",
",",
"sim_var",
")",
")",
"self",
".",
"count",
"+=",
"1",
"cand_dir",
"=",
"self",
".",
"generate_dir",
"+",
"\"/CANDIDATE_%s\"",
"%",
"candidate_i",
"if",
"not",
"os",
".",
"path",
".",
"exists",
"(",
"cand_dir",
")",
":",
"os",
".",
"mkdir",
"(",
"cand_dir",
")",
"vars",
"=",
"(",
"sim_var",
",",
"self",
".",
"ref",
",",
"self",
".",
"neuroml_file",
",",
"self",
".",
"nml_doc",
",",
"self",
".",
"still_included",
",",
"cand_dir",
",",
"self",
".",
"target",
",",
"self",
".",
"sim_time",
",",
"self",
".",
"dt",
",",
"self",
".",
"simulator",
")",
"job",
"=",
"job_server",
".",
"submit",
"(",
"run_individual",
",",
"vars",
",",
"(",
")",
",",
"(",
"\"pyneuroml.pynml\"",
",",
"'pyneuroml.tune.NeuroMLSimulation'",
",",
"'shutil'",
",",
"'neuroml'",
")",
")",
"jobs",
".",
"append",
"(",
"job",
")",
"for",
"job_i",
"in",
"range",
"(",
"len",
"(",
"jobs",
")",
")",
":",
"job",
"=",
"jobs",
"[",
"job_i",
"]",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"\"Checking parallel job %i/%i; set running so far: %i\"",
"%",
"(",
"job_i",
",",
"len",
"(",
"jobs",
")",
",",
"self",
".",
"count",
")",
")",
"t",
",",
"v",
"=",
"job",
"(",
")",
"traces",
".",
"append",
"(",
"[",
"t",
",",
"v",
"]",
")",
"#pyneuroml.pynml.print_comment_v(\"Obtained: %s\"%result) ",
"####job_server.print_stats()",
"job_server",
".",
"destroy",
"(",
")",
"print",
"(",
"\"-------------------------------------------\"",
")",
"end_time",
"=",
"time",
".",
"time",
"(",
")",
"tot",
"=",
"(",
"end_time",
"-",
"start_time",
")",
"pyneuroml",
".",
"pynml",
".",
"print_comment_v",
"(",
"'Ran %i candidates in %s seconds (~%ss per job)'",
"%",
"(",
"len",
"(",
"candidates",
")",
",",
"tot",
",",
"tot",
"/",
"len",
"(",
"candidates",
")",
")",
")",
"return",
"traces"
] |
Run simulation for each candidate
This run method will loop through each candidate and run the simulation
corresponding to its parameter values. It will populate an array called
traces with the resulting voltage traces for the simulation and return it.
|
[
"Run",
"simulation",
"for",
"each",
"candidate",
"This",
"run",
"method",
"will",
"loop",
"through",
"each",
"candidate",
"and",
"run",
"the",
"simulation",
"corresponding",
"to",
"its",
"parameter",
"values",
".",
"It",
"will",
"populate",
"an",
"array",
"called",
"traces",
"with",
"the",
"resulting",
"voltage",
"traces",
"for",
"the",
"simulation",
"and",
"return",
"it",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/tune/NeuroMLController.py#L64-L135
|
train
|
marrow/WebCore
|
web/ext/local.py
|
ThreadLocalExtension.prepare
|
def prepare(self, context):
"""Executed prior to processing a request."""
if __debug__:
log.debug("Assigning thread local request context.")
self.local.context = context
|
python
|
def prepare(self, context):
"""Executed prior to processing a request."""
if __debug__:
log.debug("Assigning thread local request context.")
self.local.context = context
|
[
"def",
"prepare",
"(",
"self",
",",
"context",
")",
":",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"Assigning thread local request context.\"",
")",
"self",
".",
"local",
".",
"context",
"=",
"context"
] |
Executed prior to processing a request.
|
[
"Executed",
"prior",
"to",
"processing",
"a",
"request",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/local.py#L79-L84
|
train
|
marrow/WebCore
|
web/core/view.py
|
WebViews.register
|
def register(self, kind, handler):
"""Register a handler for a given type, class, interface, or abstract base class.
View registration should happen within the `start` callback of an extension. For example, to register the
previous `json` view example:
class JSONExtension:
def start(self, context):
context.view.register(tuple, json)
The approach of explicitly referencing a view handler isn't very easy to override without also replacing the
extension originally adding it, however there is another approach. Using named handlers registered as discrete
plugins (via the `entry_point` argument in `setup.py`) allows the extension to easily ask "what's my handler?"
class JSONExtension:
def start(self, context):
context.view.register(
tuple,
context.view.json
)
Otherwise unknown attributes of the view registry will attempt to look up a handler plugin by that name.
"""
if __debug__: # In production this logging is completely skipped, regardless of logging level.
if py3 and not pypy: # Where possible, we shorten things to just the cannonical name.
log.debug("Registering view handler.", extra=dict(type=name(kind), handler=name(handler)))
else: # Canonical name lookup is not entirely reliable on some combinations.
log.debug("Registering view handler.", extra=dict(type=repr(kind), handler=repr(handler)))
# Add the handler to the pool of candidates. This adds to a list instead of replacing the "dictionary item".
self._map.add(kind, handler)
return handler
|
python
|
def register(self, kind, handler):
"""Register a handler for a given type, class, interface, or abstract base class.
View registration should happen within the `start` callback of an extension. For example, to register the
previous `json` view example:
class JSONExtension:
def start(self, context):
context.view.register(tuple, json)
The approach of explicitly referencing a view handler isn't very easy to override without also replacing the
extension originally adding it, however there is another approach. Using named handlers registered as discrete
plugins (via the `entry_point` argument in `setup.py`) allows the extension to easily ask "what's my handler?"
class JSONExtension:
def start(self, context):
context.view.register(
tuple,
context.view.json
)
Otherwise unknown attributes of the view registry will attempt to look up a handler plugin by that name.
"""
if __debug__: # In production this logging is completely skipped, regardless of logging level.
if py3 and not pypy: # Where possible, we shorten things to just the cannonical name.
log.debug("Registering view handler.", extra=dict(type=name(kind), handler=name(handler)))
else: # Canonical name lookup is not entirely reliable on some combinations.
log.debug("Registering view handler.", extra=dict(type=repr(kind), handler=repr(handler)))
# Add the handler to the pool of candidates. This adds to a list instead of replacing the "dictionary item".
self._map.add(kind, handler)
return handler
|
[
"def",
"register",
"(",
"self",
",",
"kind",
",",
"handler",
")",
":",
"if",
"__debug__",
":",
"# In production this logging is completely skipped, regardless of logging level.",
"if",
"py3",
"and",
"not",
"pypy",
":",
"# Where possible, we shorten things to just the cannonical name.",
"log",
".",
"debug",
"(",
"\"Registering view handler.\"",
",",
"extra",
"=",
"dict",
"(",
"type",
"=",
"name",
"(",
"kind",
")",
",",
"handler",
"=",
"name",
"(",
"handler",
")",
")",
")",
"else",
":",
"# Canonical name lookup is not entirely reliable on some combinations.",
"log",
".",
"debug",
"(",
"\"Registering view handler.\"",
",",
"extra",
"=",
"dict",
"(",
"type",
"=",
"repr",
"(",
"kind",
")",
",",
"handler",
"=",
"repr",
"(",
"handler",
")",
")",
")",
"# Add the handler to the pool of candidates. This adds to a list instead of replacing the \"dictionary item\".",
"self",
".",
"_map",
".",
"add",
"(",
"kind",
",",
"handler",
")",
"return",
"handler"
] |
Register a handler for a given type, class, interface, or abstract base class.
View registration should happen within the `start` callback of an extension. For example, to register the
previous `json` view example:
class JSONExtension:
def start(self, context):
context.view.register(tuple, json)
The approach of explicitly referencing a view handler isn't very easy to override without also replacing the
extension originally adding it, however there is another approach. Using named handlers registered as discrete
plugins (via the `entry_point` argument in `setup.py`) allows the extension to easily ask "what's my handler?"
class JSONExtension:
def start(self, context):
context.view.register(
tuple,
context.view.json
)
Otherwise unknown attributes of the view registry will attempt to look up a handler plugin by that name.
|
[
"Register",
"a",
"handler",
"for",
"a",
"given",
"type",
"class",
"interface",
"or",
"abstract",
"base",
"class",
".",
"View",
"registration",
"should",
"happen",
"within",
"the",
"start",
"callback",
"of",
"an",
"extension",
".",
"For",
"example",
"to",
"register",
"the",
"previous",
"json",
"view",
"example",
":",
"class",
"JSONExtension",
":",
"def",
"start",
"(",
"self",
"context",
")",
":",
"context",
".",
"view",
".",
"register",
"(",
"tuple",
"json",
")",
"The",
"approach",
"of",
"explicitly",
"referencing",
"a",
"view",
"handler",
"isn",
"t",
"very",
"easy",
"to",
"override",
"without",
"also",
"replacing",
"the",
"extension",
"originally",
"adding",
"it",
"however",
"there",
"is",
"another",
"approach",
".",
"Using",
"named",
"handlers",
"registered",
"as",
"discrete",
"plugins",
"(",
"via",
"the",
"entry_point",
"argument",
"in",
"setup",
".",
"py",
")",
"allows",
"the",
"extension",
"to",
"easily",
"ask",
"what",
"s",
"my",
"handler?",
"class",
"JSONExtension",
":",
"def",
"start",
"(",
"self",
"context",
")",
":",
"context",
".",
"view",
".",
"register",
"(",
"tuple",
"context",
".",
"view",
".",
"json",
")",
"Otherwise",
"unknown",
"attributes",
"of",
"the",
"view",
"registry",
"will",
"attempt",
"to",
"look",
"up",
"a",
"handler",
"plugin",
"by",
"that",
"name",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/core/view.py#L112-L144
|
train
|
marrow/WebCore
|
web/app/static.py
|
static
|
def static(base, mapping=None, far=('js', 'css', 'gif', 'jpg', 'jpeg', 'png', 'ttf', 'woff')):
"""Serve files from disk.
This utility endpoint factory is meant primarily for use in development environments; in production environments
it is better (more efficient, secure, etc.) to serve your static content using a front end load balancer such as
Nginx.
The first argument, `base`, represents the base path to serve files from. Paths below the attachment point for
the generated endpoint will combine this base path with the remaining path elements to determine the file to
serve.
The second argument is an optional dictionary mapping filename extensions to template engines, for cooperation
with the TemplateExtension. (See: https://github.com/marrow/template) The result of attempting to serve a
mapped path is a 2-tuple of `("{mapping}:{path}", dict())`. For example, to render all `.html` files as Mako
templates, you would attach something like the following:
class Root:
page = static('/path/to/static/pages', dict(html='mako'))
By default the "usual culprits" are served with far-futures cache expiry headers. If you wish to change the
extensions searched just assign a new `far` iterable. To disable, assign any falsy value.
"""
base = abspath(base)
@staticmethod
def static_handler(context, *parts, **kw):
path = normpath(pathjoin(base, *parts))
if __debug__:
log.debug("Attempting to serve static file.", extra=dict(
request = id(context),
base = base,
path = path
))
if not path.startswith(base): # Ensure we only serve files from the allowed path.
raise HTTPForbidden("Cowardly refusing to violate base path policy." if __debug__ else None)
if not exists(path): # Do the right thing if the file doesn't actually exist.
raise HTTPNotFound()
if not isfile(path): # Only serve normal files; no UNIX domain sockets, FIFOs, etc., etc.
raise HTTPForbidden("Cowardly refusing to open a non-file." if __debug__ else None)
if far and path.rpartition('.')[2] in far:
context.response.cache_expires = 60*60*24*365
if mapping: # Handle the mapping of filename extensions to 2-tuples. 'Cause why not?
_, _, extension = basename(path).partition('.')
if extension in mapping:
return mapping[extension] + ':' + path, dict()
return open(path, 'rb')
return static_handler
|
python
|
def static(base, mapping=None, far=('js', 'css', 'gif', 'jpg', 'jpeg', 'png', 'ttf', 'woff')):
"""Serve files from disk.
This utility endpoint factory is meant primarily for use in development environments; in production environments
it is better (more efficient, secure, etc.) to serve your static content using a front end load balancer such as
Nginx.
The first argument, `base`, represents the base path to serve files from. Paths below the attachment point for
the generated endpoint will combine this base path with the remaining path elements to determine the file to
serve.
The second argument is an optional dictionary mapping filename extensions to template engines, for cooperation
with the TemplateExtension. (See: https://github.com/marrow/template) The result of attempting to serve a
mapped path is a 2-tuple of `("{mapping}:{path}", dict())`. For example, to render all `.html` files as Mako
templates, you would attach something like the following:
class Root:
page = static('/path/to/static/pages', dict(html='mako'))
By default the "usual culprits" are served with far-futures cache expiry headers. If you wish to change the
extensions searched just assign a new `far` iterable. To disable, assign any falsy value.
"""
base = abspath(base)
@staticmethod
def static_handler(context, *parts, **kw):
path = normpath(pathjoin(base, *parts))
if __debug__:
log.debug("Attempting to serve static file.", extra=dict(
request = id(context),
base = base,
path = path
))
if not path.startswith(base): # Ensure we only serve files from the allowed path.
raise HTTPForbidden("Cowardly refusing to violate base path policy." if __debug__ else None)
if not exists(path): # Do the right thing if the file doesn't actually exist.
raise HTTPNotFound()
if not isfile(path): # Only serve normal files; no UNIX domain sockets, FIFOs, etc., etc.
raise HTTPForbidden("Cowardly refusing to open a non-file." if __debug__ else None)
if far and path.rpartition('.')[2] in far:
context.response.cache_expires = 60*60*24*365
if mapping: # Handle the mapping of filename extensions to 2-tuples. 'Cause why not?
_, _, extension = basename(path).partition('.')
if extension in mapping:
return mapping[extension] + ':' + path, dict()
return open(path, 'rb')
return static_handler
|
[
"def",
"static",
"(",
"base",
",",
"mapping",
"=",
"None",
",",
"far",
"=",
"(",
"'js'",
",",
"'css'",
",",
"'gif'",
",",
"'jpg'",
",",
"'jpeg'",
",",
"'png'",
",",
"'ttf'",
",",
"'woff'",
")",
")",
":",
"base",
"=",
"abspath",
"(",
"base",
")",
"@",
"staticmethod",
"def",
"static_handler",
"(",
"context",
",",
"*",
"parts",
",",
"*",
"*",
"kw",
")",
":",
"path",
"=",
"normpath",
"(",
"pathjoin",
"(",
"base",
",",
"*",
"parts",
")",
")",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"Attempting to serve static file.\"",
",",
"extra",
"=",
"dict",
"(",
"request",
"=",
"id",
"(",
"context",
")",
",",
"base",
"=",
"base",
",",
"path",
"=",
"path",
")",
")",
"if",
"not",
"path",
".",
"startswith",
"(",
"base",
")",
":",
"# Ensure we only serve files from the allowed path.",
"raise",
"HTTPForbidden",
"(",
"\"Cowardly refusing to violate base path policy.\"",
"if",
"__debug__",
"else",
"None",
")",
"if",
"not",
"exists",
"(",
"path",
")",
":",
"# Do the right thing if the file doesn't actually exist.",
"raise",
"HTTPNotFound",
"(",
")",
"if",
"not",
"isfile",
"(",
"path",
")",
":",
"# Only serve normal files; no UNIX domain sockets, FIFOs, etc., etc.",
"raise",
"HTTPForbidden",
"(",
"\"Cowardly refusing to open a non-file.\"",
"if",
"__debug__",
"else",
"None",
")",
"if",
"far",
"and",
"path",
".",
"rpartition",
"(",
"'.'",
")",
"[",
"2",
"]",
"in",
"far",
":",
"context",
".",
"response",
".",
"cache_expires",
"=",
"60",
"*",
"60",
"*",
"24",
"*",
"365",
"if",
"mapping",
":",
"# Handle the mapping of filename extensions to 2-tuples. 'Cause why not?",
"_",
",",
"_",
",",
"extension",
"=",
"basename",
"(",
"path",
")",
".",
"partition",
"(",
"'.'",
")",
"if",
"extension",
"in",
"mapping",
":",
"return",
"mapping",
"[",
"extension",
"]",
"+",
"':'",
"+",
"path",
",",
"dict",
"(",
")",
"return",
"open",
"(",
"path",
",",
"'rb'",
")",
"return",
"static_handler"
] |
Serve files from disk.
This utility endpoint factory is meant primarily for use in development environments; in production environments
it is better (more efficient, secure, etc.) to serve your static content using a front end load balancer such as
Nginx.
The first argument, `base`, represents the base path to serve files from. Paths below the attachment point for
the generated endpoint will combine this base path with the remaining path elements to determine the file to
serve.
The second argument is an optional dictionary mapping filename extensions to template engines, for cooperation
with the TemplateExtension. (See: https://github.com/marrow/template) The result of attempting to serve a
mapped path is a 2-tuple of `("{mapping}:{path}", dict())`. For example, to render all `.html` files as Mako
templates, you would attach something like the following:
class Root:
page = static('/path/to/static/pages', dict(html='mako'))
By default the "usual culprits" are served with far-futures cache expiry headers. If you wish to change the
extensions searched just assign a new `far` iterable. To disable, assign any falsy value.
|
[
"Serve",
"files",
"from",
"disk",
".",
"This",
"utility",
"endpoint",
"factory",
"is",
"meant",
"primarily",
"for",
"use",
"in",
"development",
"environments",
";",
"in",
"production",
"environments",
"it",
"is",
"better",
"(",
"more",
"efficient",
"secure",
"etc",
".",
")",
"to",
"serve",
"your",
"static",
"content",
"using",
"a",
"front",
"end",
"load",
"balancer",
"such",
"as",
"Nginx",
".",
"The",
"first",
"argument",
"base",
"represents",
"the",
"base",
"path",
"to",
"serve",
"files",
"from",
".",
"Paths",
"below",
"the",
"attachment",
"point",
"for",
"the",
"generated",
"endpoint",
"will",
"combine",
"this",
"base",
"path",
"with",
"the",
"remaining",
"path",
"elements",
"to",
"determine",
"the",
"file",
"to",
"serve",
".",
"The",
"second",
"argument",
"is",
"an",
"optional",
"dictionary",
"mapping",
"filename",
"extensions",
"to",
"template",
"engines",
"for",
"cooperation",
"with",
"the",
"TemplateExtension",
".",
"(",
"See",
":",
"https",
":",
"//",
"github",
".",
"com",
"/",
"marrow",
"/",
"template",
")",
"The",
"result",
"of",
"attempting",
"to",
"serve",
"a",
"mapped",
"path",
"is",
"a",
"2",
"-",
"tuple",
"of",
"(",
"{",
"mapping",
"}",
":",
"{",
"path",
"}",
"dict",
"()",
")",
".",
"For",
"example",
"to",
"render",
"all",
".",
"html",
"files",
"as",
"Mako",
"templates",
"you",
"would",
"attach",
"something",
"like",
"the",
"following",
":",
"class",
"Root",
":",
"page",
"=",
"static",
"(",
"/",
"path",
"/",
"to",
"/",
"static",
"/",
"pages",
"dict",
"(",
"html",
"=",
"mako",
"))",
"By",
"default",
"the",
"usual",
"culprits",
"are",
"served",
"with",
"far",
"-",
"futures",
"cache",
"expiry",
"headers",
".",
"If",
"you",
"wish",
"to",
"change",
"the",
"extensions",
"searched",
"just",
"assign",
"a",
"new",
"far",
"iterable",
".",
"To",
"disable",
"assign",
"any",
"falsy",
"value",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/app/static.py#L21-L76
|
train
|
marrow/WebCore
|
web/server/diesel_.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080):
"""Diesel-based (greenlet) WSGI-HTTP server.
As a minor note, this is crazy. Diesel includes Flask, too.
"""
# Instantiate the server with a host/port configuration and our application.
WSGIApplication(application, port=int(port), iface=host).run()
|
python
|
def serve(application, host='127.0.0.1', port=8080):
"""Diesel-based (greenlet) WSGI-HTTP server.
As a minor note, this is crazy. Diesel includes Flask, too.
"""
# Instantiate the server with a host/port configuration and our application.
WSGIApplication(application, port=int(port), iface=host).run()
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
")",
":",
"# Instantiate the server with a host/port configuration and our application.",
"WSGIApplication",
"(",
"application",
",",
"port",
"=",
"int",
"(",
"port",
")",
",",
"iface",
"=",
"host",
")",
".",
"run",
"(",
")"
] |
Diesel-based (greenlet) WSGI-HTTP server.
As a minor note, this is crazy. Diesel includes Flask, too.
|
[
"Diesel",
"-",
"based",
"(",
"greenlet",
")",
"WSGI",
"-",
"HTTP",
"server",
".",
"As",
"a",
"minor",
"note",
"this",
"is",
"crazy",
".",
"Diesel",
"includes",
"Flask",
"too",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/diesel_.py#L14-L21
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/plot/PlotSpikes.py
|
process_args
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(description="A script for plotting files containing spike time data")
parser.add_argument('spiketimeFiles',
type=str,
metavar='<spiketime file>',
help='List of text file containing spike times',
nargs='+')
parser.add_argument('-format',
type=str,
metavar='<format>',
default=DEFAULTS['format'],
help='How the spiketimes are represented on each line of file:\n'+\
'id_t: id of cell, space(s)/tab(s), time of spike (default);\n'+\
't_id: time of spike, space(s)/tab(s), id of cell;\n'+\
'sonata: SONATA format HDF5 file containing spike times')
parser.add_argument('-rates',
action='store_true',
default=DEFAULTS['rates'],
help='Show a plot of rates')
parser.add_argument('-showPlotsAlready',
action='store_true',
default=DEFAULTS['show_plots_already'],
help='Show plots once generated')
parser.add_argument('-saveSpikePlotTo',
type=str,
metavar='<spiketime plot filename>',
default=DEFAULTS['save_spike_plot_to'],
help='Name of file in which to save spiketime plot')
parser.add_argument('-rateWindow',
type=int,
metavar='<rate window>',
default=DEFAULTS['rate_window'],
help='Window for rate calculation in ms')
parser.add_argument('-rateBins',
type=int,
metavar='<rate bins>',
default=DEFAULTS['rate_bins'],
help='Number of bins for rate histogram')
return parser.parse_args()
|
python
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(description="A script for plotting files containing spike time data")
parser.add_argument('spiketimeFiles',
type=str,
metavar='<spiketime file>',
help='List of text file containing spike times',
nargs='+')
parser.add_argument('-format',
type=str,
metavar='<format>',
default=DEFAULTS['format'],
help='How the spiketimes are represented on each line of file:\n'+\
'id_t: id of cell, space(s)/tab(s), time of spike (default);\n'+\
't_id: time of spike, space(s)/tab(s), id of cell;\n'+\
'sonata: SONATA format HDF5 file containing spike times')
parser.add_argument('-rates',
action='store_true',
default=DEFAULTS['rates'],
help='Show a plot of rates')
parser.add_argument('-showPlotsAlready',
action='store_true',
default=DEFAULTS['show_plots_already'],
help='Show plots once generated')
parser.add_argument('-saveSpikePlotTo',
type=str,
metavar='<spiketime plot filename>',
default=DEFAULTS['save_spike_plot_to'],
help='Name of file in which to save spiketime plot')
parser.add_argument('-rateWindow',
type=int,
metavar='<rate window>',
default=DEFAULTS['rate_window'],
help='Window for rate calculation in ms')
parser.add_argument('-rateBins',
type=int,
metavar='<rate bins>',
default=DEFAULTS['rate_bins'],
help='Number of bins for rate histogram')
return parser.parse_args()
|
[
"def",
"process_args",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"\"A script for plotting files containing spike time data\"",
")",
"parser",
".",
"add_argument",
"(",
"'spiketimeFiles'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<spiketime file>'",
",",
"help",
"=",
"'List of text file containing spike times'",
",",
"nargs",
"=",
"'+'",
")",
"parser",
".",
"add_argument",
"(",
"'-format'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<format>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'format'",
"]",
",",
"help",
"=",
"'How the spiketimes are represented on each line of file:\\n'",
"+",
"'id_t: id of cell, space(s)/tab(s), time of spike (default);\\n'",
"+",
"'t_id: time of spike, space(s)/tab(s), id of cell;\\n'",
"+",
"'sonata: SONATA format HDF5 file containing spike times'",
")",
"parser",
".",
"add_argument",
"(",
"'-rates'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'rates'",
"]",
",",
"help",
"=",
"'Show a plot of rates'",
")",
"parser",
".",
"add_argument",
"(",
"'-showPlotsAlready'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'show_plots_already'",
"]",
",",
"help",
"=",
"'Show plots once generated'",
")",
"parser",
".",
"add_argument",
"(",
"'-saveSpikePlotTo'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<spiketime plot filename>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'save_spike_plot_to'",
"]",
",",
"help",
"=",
"'Name of file in which to save spiketime plot'",
")",
"parser",
".",
"add_argument",
"(",
"'-rateWindow'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<rate window>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'rate_window'",
"]",
",",
"help",
"=",
"'Window for rate calculation in ms'",
")",
"parser",
".",
"add_argument",
"(",
"'-rateBins'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<rate bins>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'rate_bins'",
"]",
",",
"help",
"=",
"'Number of bins for rate histogram'",
")",
"return",
"parser",
".",
"parse_args",
"(",
")"
] |
Parse command-line arguments.
|
[
"Parse",
"command",
"-",
"line",
"arguments",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/plot/PlotSpikes.py#L56-L105
|
train
|
marrow/WebCore
|
web/server/stdlib.py
|
simple
|
def simple(application, host='127.0.0.1', port=8080):
"""Python-standard WSGI-HTTP server for testing purposes.
The additional work performed here is to match the default startup output of "waitress".
This is not a production quality interface and will be have badly under load.
"""
# Try to be handy as many terminals allow clicking links.
print("serving on http://{0}:{1}".format(host, port))
# Bind and launch the server; this is a blocking operation.
make_server(host, int(port), application).serve_forever()
|
python
|
def simple(application, host='127.0.0.1', port=8080):
"""Python-standard WSGI-HTTP server for testing purposes.
The additional work performed here is to match the default startup output of "waitress".
This is not a production quality interface and will be have badly under load.
"""
# Try to be handy as many terminals allow clicking links.
print("serving on http://{0}:{1}".format(host, port))
# Bind and launch the server; this is a blocking operation.
make_server(host, int(port), application).serve_forever()
|
[
"def",
"simple",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
")",
":",
"# Try to be handy as many terminals allow clicking links.",
"print",
"(",
"\"serving on http://{0}:{1}\"",
".",
"format",
"(",
"host",
",",
"port",
")",
")",
"# Bind and launch the server; this is a blocking operation.",
"make_server",
"(",
"host",
",",
"int",
"(",
"port",
")",
",",
"application",
")",
".",
"serve_forever",
"(",
")"
] |
Python-standard WSGI-HTTP server for testing purposes.
The additional work performed here is to match the default startup output of "waitress".
This is not a production quality interface and will be have badly under load.
|
[
"Python",
"-",
"standard",
"WSGI",
"-",
"HTTP",
"server",
"for",
"testing",
"purposes",
".",
"The",
"additional",
"work",
"performed",
"here",
"is",
"to",
"match",
"the",
"default",
"startup",
"output",
"of",
"waitress",
".",
"This",
"is",
"not",
"a",
"production",
"quality",
"interface",
"and",
"will",
"be",
"have",
"badly",
"under",
"load",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/stdlib.py#L23-L35
|
train
|
marrow/WebCore
|
web/server/stdlib.py
|
iiscgi
|
def iiscgi(application):
"""A specialized version of the reference WSGI-CGI server to adapt to Microsoft IIS quirks.
This is not a production quality interface and will behave badly under load.
"""
try:
from wsgiref.handlers import IISCGIHandler
except ImportError:
print("Python 3.2 or newer is required.")
if not __debug__:
warnings.warn("Interactive debugging and other persistence-based processes will not work.")
IISCGIHandler().run(application)
|
python
|
def iiscgi(application):
"""A specialized version of the reference WSGI-CGI server to adapt to Microsoft IIS quirks.
This is not a production quality interface and will behave badly under load.
"""
try:
from wsgiref.handlers import IISCGIHandler
except ImportError:
print("Python 3.2 or newer is required.")
if not __debug__:
warnings.warn("Interactive debugging and other persistence-based processes will not work.")
IISCGIHandler().run(application)
|
[
"def",
"iiscgi",
"(",
"application",
")",
":",
"try",
":",
"from",
"wsgiref",
".",
"handlers",
"import",
"IISCGIHandler",
"except",
"ImportError",
":",
"print",
"(",
"\"Python 3.2 or newer is required.\"",
")",
"if",
"not",
"__debug__",
":",
"warnings",
".",
"warn",
"(",
"\"Interactive debugging and other persistence-based processes will not work.\"",
")",
"IISCGIHandler",
"(",
")",
".",
"run",
"(",
"application",
")"
] |
A specialized version of the reference WSGI-CGI server to adapt to Microsoft IIS quirks.
This is not a production quality interface and will behave badly under load.
|
[
"A",
"specialized",
"version",
"of",
"the",
"reference",
"WSGI",
"-",
"CGI",
"server",
"to",
"adapt",
"to",
"Microsoft",
"IIS",
"quirks",
".",
"This",
"is",
"not",
"a",
"production",
"quality",
"interface",
"and",
"will",
"behave",
"badly",
"under",
"load",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/stdlib.py#L53-L66
|
train
|
marrow/WebCore
|
web/server/fcgi.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080, socket=None, **options):
"""Basic FastCGI support via flup.
This web server has many, many options. Please see the Flup project documentation for details.
"""
# Allow either on-disk socket (recommended) or TCP/IP socket use.
if not socket:
bindAddress = (host, int(port))
else:
bindAddress = socket
# Bind and start the blocking web server interface.
WSGIServer(application, bindAddress=bindAddress, **options).run()
|
python
|
def serve(application, host='127.0.0.1', port=8080, socket=None, **options):
"""Basic FastCGI support via flup.
This web server has many, many options. Please see the Flup project documentation for details.
"""
# Allow either on-disk socket (recommended) or TCP/IP socket use.
if not socket:
bindAddress = (host, int(port))
else:
bindAddress = socket
# Bind and start the blocking web server interface.
WSGIServer(application, bindAddress=bindAddress, **options).run()
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
",",
"socket",
"=",
"None",
",",
"*",
"*",
"options",
")",
":",
"# Allow either on-disk socket (recommended) or TCP/IP socket use.",
"if",
"not",
"socket",
":",
"bindAddress",
"=",
"(",
"host",
",",
"int",
"(",
"port",
")",
")",
"else",
":",
"bindAddress",
"=",
"socket",
"# Bind and start the blocking web server interface.",
"WSGIServer",
"(",
"application",
",",
"bindAddress",
"=",
"bindAddress",
",",
"*",
"*",
"options",
")",
".",
"run",
"(",
")"
] |
Basic FastCGI support via flup.
This web server has many, many options. Please see the Flup project documentation for details.
|
[
"Basic",
"FastCGI",
"support",
"via",
"flup",
".",
"This",
"web",
"server",
"has",
"many",
"many",
"options",
".",
"Please",
"see",
"the",
"Flup",
"project",
"documentation",
"for",
"details",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/fcgi.py#L18-L31
|
train
|
bitlabstudio/django-subscribe
|
subscribe/forms.py
|
SubscriptionCreateForm._get_method_kwargs
|
def _get_method_kwargs(self):
"""
Helper method. Returns kwargs needed to filter the correct object.
Can also be used to create the correct object.
"""
method_kwargs = {
'user': self.user,
'content_type': self.ctype,
'object_id': self.content_object.pk,
}
return method_kwargs
|
python
|
def _get_method_kwargs(self):
"""
Helper method. Returns kwargs needed to filter the correct object.
Can also be used to create the correct object.
"""
method_kwargs = {
'user': self.user,
'content_type': self.ctype,
'object_id': self.content_object.pk,
}
return method_kwargs
|
[
"def",
"_get_method_kwargs",
"(",
"self",
")",
":",
"method_kwargs",
"=",
"{",
"'user'",
":",
"self",
".",
"user",
",",
"'content_type'",
":",
"self",
".",
"ctype",
",",
"'object_id'",
":",
"self",
".",
"content_object",
".",
"pk",
",",
"}",
"return",
"method_kwargs"
] |
Helper method. Returns kwargs needed to filter the correct object.
Can also be used to create the correct object.
|
[
"Helper",
"method",
".",
"Returns",
"kwargs",
"needed",
"to",
"filter",
"the",
"correct",
"object",
"."
] |
313de63fb4acda172e88b65c3327c793f98e8aa9
|
https://github.com/bitlabstudio/django-subscribe/blob/313de63fb4acda172e88b65c3327c793f98e8aa9/subscribe/forms.py#L15-L27
|
train
|
bitlabstudio/django-subscribe
|
subscribe/forms.py
|
SubscriptionCreateForm.save
|
def save(self, *args, **kwargs):
"""Adds a subscription for the given user to the given object."""
method_kwargs = self._get_method_kwargs()
try:
subscription = Subscription.objects.get(**method_kwargs)
except Subscription.DoesNotExist:
subscription = Subscription.objects.create(**method_kwargs)
return subscription
|
python
|
def save(self, *args, **kwargs):
"""Adds a subscription for the given user to the given object."""
method_kwargs = self._get_method_kwargs()
try:
subscription = Subscription.objects.get(**method_kwargs)
except Subscription.DoesNotExist:
subscription = Subscription.objects.create(**method_kwargs)
return subscription
|
[
"def",
"save",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"method_kwargs",
"=",
"self",
".",
"_get_method_kwargs",
"(",
")",
"try",
":",
"subscription",
"=",
"Subscription",
".",
"objects",
".",
"get",
"(",
"*",
"*",
"method_kwargs",
")",
"except",
"Subscription",
".",
"DoesNotExist",
":",
"subscription",
"=",
"Subscription",
".",
"objects",
".",
"create",
"(",
"*",
"*",
"method_kwargs",
")",
"return",
"subscription"
] |
Adds a subscription for the given user to the given object.
|
[
"Adds",
"a",
"subscription",
"for",
"the",
"given",
"user",
"to",
"the",
"given",
"object",
"."
] |
313de63fb4acda172e88b65c3327c793f98e8aa9
|
https://github.com/bitlabstudio/django-subscribe/blob/313de63fb4acda172e88b65c3327c793f98e8aa9/subscribe/forms.py#L29-L36
|
train
|
marrow/WebCore
|
web/ext/base.py
|
BaseExtension.prepare
|
def prepare(self, context):
"""Add the usual suspects to the context.
This adds `request`, `response`, and `path` to the `RequestContext` instance.
"""
if __debug__:
log.debug("Preparing request context.", extra=dict(request=id(context)))
# Bridge in WebOb `Request` and `Response` objects.
# Extensions shouldn't rely on these, using `environ` where possible instead.
context.request = Request(context.environ)
context.response = Response(request=context.request)
# Record the initial path representing the point where a front-end web server bridged to us.
context.environ['web.base'] = context.request.script_name
# Track the remaining (unprocessed) path elements.
context.request.remainder = context.request.path_info.split('/')
if context.request.remainder and not context.request.remainder[0]:
del context.request.remainder[0]
# Track the "breadcrumb list" of dispatch through distinct controllers.
context.path = Bread()
|
python
|
def prepare(self, context):
"""Add the usual suspects to the context.
This adds `request`, `response`, and `path` to the `RequestContext` instance.
"""
if __debug__:
log.debug("Preparing request context.", extra=dict(request=id(context)))
# Bridge in WebOb `Request` and `Response` objects.
# Extensions shouldn't rely on these, using `environ` where possible instead.
context.request = Request(context.environ)
context.response = Response(request=context.request)
# Record the initial path representing the point where a front-end web server bridged to us.
context.environ['web.base'] = context.request.script_name
# Track the remaining (unprocessed) path elements.
context.request.remainder = context.request.path_info.split('/')
if context.request.remainder and not context.request.remainder[0]:
del context.request.remainder[0]
# Track the "breadcrumb list" of dispatch through distinct controllers.
context.path = Bread()
|
[
"def",
"prepare",
"(",
"self",
",",
"context",
")",
":",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"Preparing request context.\"",
",",
"extra",
"=",
"dict",
"(",
"request",
"=",
"id",
"(",
"context",
")",
")",
")",
"# Bridge in WebOb `Request` and `Response` objects.",
"# Extensions shouldn't rely on these, using `environ` where possible instead.",
"context",
".",
"request",
"=",
"Request",
"(",
"context",
".",
"environ",
")",
"context",
".",
"response",
"=",
"Response",
"(",
"request",
"=",
"context",
".",
"request",
")",
"# Record the initial path representing the point where a front-end web server bridged to us.",
"context",
".",
"environ",
"[",
"'web.base'",
"]",
"=",
"context",
".",
"request",
".",
"script_name",
"# Track the remaining (unprocessed) path elements.",
"context",
".",
"request",
".",
"remainder",
"=",
"context",
".",
"request",
".",
"path_info",
".",
"split",
"(",
"'/'",
")",
"if",
"context",
".",
"request",
".",
"remainder",
"and",
"not",
"context",
".",
"request",
".",
"remainder",
"[",
"0",
"]",
":",
"del",
"context",
".",
"request",
".",
"remainder",
"[",
"0",
"]",
"# Track the \"breadcrumb list\" of dispatch through distinct controllers.",
"context",
".",
"path",
"=",
"Bread",
"(",
")"
] |
Add the usual suspects to the context.
This adds `request`, `response`, and `path` to the `RequestContext` instance.
|
[
"Add",
"the",
"usual",
"suspects",
"to",
"the",
"context",
".",
"This",
"adds",
"request",
"response",
"and",
"path",
"to",
"the",
"RequestContext",
"instance",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/base.py#L83-L106
|
train
|
marrow/WebCore
|
web/ext/base.py
|
BaseExtension.dispatch
|
def dispatch(self, context, consumed, handler, is_endpoint):
"""Called as dispatch descends into a tier.
The base extension uses this to maintain the "current url".
"""
request = context.request
if __debug__:
log.debug("Handling dispatch event.", extra=dict(
request = id(context),
consumed = consumed,
handler = safe_name(handler),
endpoint = is_endpoint
))
# The leading path element (leading slash) requires special treatment.
if not consumed and context.request.path_info_peek() == '':
consumed = ['']
nConsumed = 0
if consumed:
# Migrate path elements consumed from the `PATH_INFO` to `SCRIPT_NAME` WSGI environment variables.
if not isinstance(consumed, (list, tuple)):
consumed = consumed.split('/')
for element in consumed:
if element == context.request.path_info_peek():
context.request.path_info_pop()
nConsumed += 1
else:
break
# Update the breadcrumb list.
context.path.append(Crumb(handler, Path(request.script_name)))
if consumed: # Lastly, update the remaining path element list.
request.remainder = request.remainder[nConsumed:]
|
python
|
def dispatch(self, context, consumed, handler, is_endpoint):
"""Called as dispatch descends into a tier.
The base extension uses this to maintain the "current url".
"""
request = context.request
if __debug__:
log.debug("Handling dispatch event.", extra=dict(
request = id(context),
consumed = consumed,
handler = safe_name(handler),
endpoint = is_endpoint
))
# The leading path element (leading slash) requires special treatment.
if not consumed and context.request.path_info_peek() == '':
consumed = ['']
nConsumed = 0
if consumed:
# Migrate path elements consumed from the `PATH_INFO` to `SCRIPT_NAME` WSGI environment variables.
if not isinstance(consumed, (list, tuple)):
consumed = consumed.split('/')
for element in consumed:
if element == context.request.path_info_peek():
context.request.path_info_pop()
nConsumed += 1
else:
break
# Update the breadcrumb list.
context.path.append(Crumb(handler, Path(request.script_name)))
if consumed: # Lastly, update the remaining path element list.
request.remainder = request.remainder[nConsumed:]
|
[
"def",
"dispatch",
"(",
"self",
",",
"context",
",",
"consumed",
",",
"handler",
",",
"is_endpoint",
")",
":",
"request",
"=",
"context",
".",
"request",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"Handling dispatch event.\"",
",",
"extra",
"=",
"dict",
"(",
"request",
"=",
"id",
"(",
"context",
")",
",",
"consumed",
"=",
"consumed",
",",
"handler",
"=",
"safe_name",
"(",
"handler",
")",
",",
"endpoint",
"=",
"is_endpoint",
")",
")",
"# The leading path element (leading slash) requires special treatment.",
"if",
"not",
"consumed",
"and",
"context",
".",
"request",
".",
"path_info_peek",
"(",
")",
"==",
"''",
":",
"consumed",
"=",
"[",
"''",
"]",
"nConsumed",
"=",
"0",
"if",
"consumed",
":",
"# Migrate path elements consumed from the `PATH_INFO` to `SCRIPT_NAME` WSGI environment variables.",
"if",
"not",
"isinstance",
"(",
"consumed",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"consumed",
"=",
"consumed",
".",
"split",
"(",
"'/'",
")",
"for",
"element",
"in",
"consumed",
":",
"if",
"element",
"==",
"context",
".",
"request",
".",
"path_info_peek",
"(",
")",
":",
"context",
".",
"request",
".",
"path_info_pop",
"(",
")",
"nConsumed",
"+=",
"1",
"else",
":",
"break",
"# Update the breadcrumb list.",
"context",
".",
"path",
".",
"append",
"(",
"Crumb",
"(",
"handler",
",",
"Path",
"(",
"request",
".",
"script_name",
")",
")",
")",
"if",
"consumed",
":",
"# Lastly, update the remaining path element list.",
"request",
".",
"remainder",
"=",
"request",
".",
"remainder",
"[",
"nConsumed",
":",
"]"
] |
Called as dispatch descends into a tier.
The base extension uses this to maintain the "current url".
|
[
"Called",
"as",
"dispatch",
"descends",
"into",
"a",
"tier",
".",
"The",
"base",
"extension",
"uses",
"this",
"to",
"maintain",
"the",
"current",
"url",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/base.py#L108-L145
|
train
|
marrow/WebCore
|
web/ext/base.py
|
BaseExtension.render_none
|
def render_none(self, context, result):
"""Render empty responses."""
context.response.body = b''
del context.response.content_length
return True
|
python
|
def render_none(self, context, result):
"""Render empty responses."""
context.response.body = b''
del context.response.content_length
return True
|
[
"def",
"render_none",
"(",
"self",
",",
"context",
",",
"result",
")",
":",
"context",
".",
"response",
".",
"body",
"=",
"b''",
"del",
"context",
".",
"response",
".",
"content_length",
"return",
"True"
] |
Render empty responses.
|
[
"Render",
"empty",
"responses",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/base.py#L149-L153
|
train
|
marrow/WebCore
|
web/ext/base.py
|
BaseExtension.render_binary
|
def render_binary(self, context, result):
"""Return binary responses unmodified."""
context.response.app_iter = iter((result, )) # This wraps the binary string in a WSGI body iterable.
return True
|
python
|
def render_binary(self, context, result):
"""Return binary responses unmodified."""
context.response.app_iter = iter((result, )) # This wraps the binary string in a WSGI body iterable.
return True
|
[
"def",
"render_binary",
"(",
"self",
",",
"context",
",",
"result",
")",
":",
"context",
".",
"response",
".",
"app_iter",
"=",
"iter",
"(",
"(",
"result",
",",
")",
")",
"# This wraps the binary string in a WSGI body iterable.",
"return",
"True"
] |
Return binary responses unmodified.
|
[
"Return",
"binary",
"responses",
"unmodified",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/base.py#L160-L163
|
train
|
marrow/WebCore
|
web/ext/base.py
|
BaseExtension.render_file
|
def render_file(self, context, result):
"""Perform appropriate metadata wrangling for returned open file handles."""
if __debug__:
log.debug("Processing file-like object.", extra=dict(request=id(context), result=repr(result)))
response = context.response
response.conditional_response = True
modified = mktime(gmtime(getmtime(result.name)))
response.last_modified = datetime.fromtimestamp(modified)
ct, ce = guess_type(result.name)
if not ct: ct = 'application/octet-stream'
response.content_type, response.content_encoding = ct, ce
response.etag = unicode(modified)
result.seek(0, 2) # Seek to the end of the file.
response.content_length = result.tell()
result.seek(0) # Seek back to the start of the file.
response.body_file = result
return True
|
python
|
def render_file(self, context, result):
"""Perform appropriate metadata wrangling for returned open file handles."""
if __debug__:
log.debug("Processing file-like object.", extra=dict(request=id(context), result=repr(result)))
response = context.response
response.conditional_response = True
modified = mktime(gmtime(getmtime(result.name)))
response.last_modified = datetime.fromtimestamp(modified)
ct, ce = guess_type(result.name)
if not ct: ct = 'application/octet-stream'
response.content_type, response.content_encoding = ct, ce
response.etag = unicode(modified)
result.seek(0, 2) # Seek to the end of the file.
response.content_length = result.tell()
result.seek(0) # Seek back to the start of the file.
response.body_file = result
return True
|
[
"def",
"render_file",
"(",
"self",
",",
"context",
",",
"result",
")",
":",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"Processing file-like object.\"",
",",
"extra",
"=",
"dict",
"(",
"request",
"=",
"id",
"(",
"context",
")",
",",
"result",
"=",
"repr",
"(",
"result",
")",
")",
")",
"response",
"=",
"context",
".",
"response",
"response",
".",
"conditional_response",
"=",
"True",
"modified",
"=",
"mktime",
"(",
"gmtime",
"(",
"getmtime",
"(",
"result",
".",
"name",
")",
")",
")",
"response",
".",
"last_modified",
"=",
"datetime",
".",
"fromtimestamp",
"(",
"modified",
")",
"ct",
",",
"ce",
"=",
"guess_type",
"(",
"result",
".",
"name",
")",
"if",
"not",
"ct",
":",
"ct",
"=",
"'application/octet-stream'",
"response",
".",
"content_type",
",",
"response",
".",
"content_encoding",
"=",
"ct",
",",
"ce",
"response",
".",
"etag",
"=",
"unicode",
"(",
"modified",
")",
"result",
".",
"seek",
"(",
"0",
",",
"2",
")",
"# Seek to the end of the file.",
"response",
".",
"content_length",
"=",
"result",
".",
"tell",
"(",
")",
"result",
".",
"seek",
"(",
"0",
")",
"# Seek back to the start of the file.",
"response",
".",
"body_file",
"=",
"result",
"return",
"True"
] |
Perform appropriate metadata wrangling for returned open file handles.
|
[
"Perform",
"appropriate",
"metadata",
"wrangling",
"for",
"returned",
"open",
"file",
"handles",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/base.py#L170-L192
|
train
|
marrow/WebCore
|
web/ext/base.py
|
BaseExtension.render_generator
|
def render_generator(self, context, result):
"""Attempt to serve generator responses through stream encoding.
This allows for direct use of cinje template functions, which are generators, as returned views.
"""
context.response.encoding = 'utf8'
context.response.app_iter = (
(i.encode('utf8') if isinstance(i, unicode) else i) # Stream encode unicode chunks.
for i in result if i is not None # Skip None values.
)
return True
|
python
|
def render_generator(self, context, result):
"""Attempt to serve generator responses through stream encoding.
This allows for direct use of cinje template functions, which are generators, as returned views.
"""
context.response.encoding = 'utf8'
context.response.app_iter = (
(i.encode('utf8') if isinstance(i, unicode) else i) # Stream encode unicode chunks.
for i in result if i is not None # Skip None values.
)
return True
|
[
"def",
"render_generator",
"(",
"self",
",",
"context",
",",
"result",
")",
":",
"context",
".",
"response",
".",
"encoding",
"=",
"'utf8'",
"context",
".",
"response",
".",
"app_iter",
"=",
"(",
"(",
"i",
".",
"encode",
"(",
"'utf8'",
")",
"if",
"isinstance",
"(",
"i",
",",
"unicode",
")",
"else",
"i",
")",
"# Stream encode unicode chunks.",
"for",
"i",
"in",
"result",
"if",
"i",
"is",
"not",
"None",
"# Skip None values.",
")",
"return",
"True"
] |
Attempt to serve generator responses through stream encoding.
This allows for direct use of cinje template functions, which are generators, as returned views.
|
[
"Attempt",
"to",
"serve",
"generator",
"responses",
"through",
"stream",
"encoding",
".",
"This",
"allows",
"for",
"direct",
"use",
"of",
"cinje",
"template",
"functions",
"which",
"are",
"generators",
"as",
"returned",
"views",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/base.py#L194-L204
|
train
|
marrow/WebCore
|
web/server/cherrypy_.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080):
"""CherryPy-based WSGI-HTTP server."""
# Instantiate the server with our configuration and application.
server = CherryPyWSGIServer((host, int(port)), application, server_name=host)
# Try to be handy as many terminals allow clicking links.
print("serving on http://{0}:{1}".format(host, port))
# Bind and launch the server; this is a blocking operation.
try:
server.start()
except KeyboardInterrupt:
server.stop()
|
python
|
def serve(application, host='127.0.0.1', port=8080):
"""CherryPy-based WSGI-HTTP server."""
# Instantiate the server with our configuration and application.
server = CherryPyWSGIServer((host, int(port)), application, server_name=host)
# Try to be handy as many terminals allow clicking links.
print("serving on http://{0}:{1}".format(host, port))
# Bind and launch the server; this is a blocking operation.
try:
server.start()
except KeyboardInterrupt:
server.stop()
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
")",
":",
"# Instantiate the server with our configuration and application.",
"server",
"=",
"CherryPyWSGIServer",
"(",
"(",
"host",
",",
"int",
"(",
"port",
")",
")",
",",
"application",
",",
"server_name",
"=",
"host",
")",
"# Try to be handy as many terminals allow clicking links.",
"print",
"(",
"\"serving on http://{0}:{1}\"",
".",
"format",
"(",
"host",
",",
"port",
")",
")",
"# Bind and launch the server; this is a blocking operation.",
"try",
":",
"server",
".",
"start",
"(",
")",
"except",
"KeyboardInterrupt",
":",
"server",
".",
"stop",
"(",
")"
] |
CherryPy-based WSGI-HTTP server.
|
[
"CherryPy",
"-",
"based",
"WSGI",
"-",
"HTTP",
"server",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/cherrypy_.py#L14-L27
|
train
|
broadinstitute/xtermcolor
|
xtermcolor/ColorMap.py
|
TerminalColorMap.colorize
|
def colorize(self, string, rgb=None, ansi=None, bg=None, ansi_bg=None):
'''Returns the colored string'''
if not isinstance(string, str):
string = str(string)
if rgb is None and ansi is None:
raise TerminalColorMapException(
'colorize: must specify one named parameter: rgb or ansi')
if rgb is not None and ansi is not None:
raise TerminalColorMapException(
'colorize: must specify only one named parameter: rgb or ansi')
if bg is not None and ansi_bg is not None:
raise TerminalColorMapException(
'colorize: must specify only one named parameter: bg or ansi_bg')
if rgb is not None:
(closestAnsi, closestRgb) = self.convert(rgb)
elif ansi is not None:
(closestAnsi, closestRgb) = (ansi, self.colors[ansi])
if bg is None and ansi_bg is None:
return "\033[38;5;{ansiCode:d}m{string:s}\033[0m".format(ansiCode=closestAnsi, string=string)
if bg is not None:
(closestBgAnsi, unused) = self.convert(bg)
elif ansi_bg is not None:
(closestBgAnsi, unused) = (ansi_bg, self.colors[ansi_bg])
return "\033[38;5;{ansiCode:d}m\033[48;5;{bf:d}m{string:s}\033[0m".format(ansiCode=closestAnsi, bf=closestBgAnsi, string=string)
|
python
|
def colorize(self, string, rgb=None, ansi=None, bg=None, ansi_bg=None):
'''Returns the colored string'''
if not isinstance(string, str):
string = str(string)
if rgb is None and ansi is None:
raise TerminalColorMapException(
'colorize: must specify one named parameter: rgb or ansi')
if rgb is not None and ansi is not None:
raise TerminalColorMapException(
'colorize: must specify only one named parameter: rgb or ansi')
if bg is not None and ansi_bg is not None:
raise TerminalColorMapException(
'colorize: must specify only one named parameter: bg or ansi_bg')
if rgb is not None:
(closestAnsi, closestRgb) = self.convert(rgb)
elif ansi is not None:
(closestAnsi, closestRgb) = (ansi, self.colors[ansi])
if bg is None and ansi_bg is None:
return "\033[38;5;{ansiCode:d}m{string:s}\033[0m".format(ansiCode=closestAnsi, string=string)
if bg is not None:
(closestBgAnsi, unused) = self.convert(bg)
elif ansi_bg is not None:
(closestBgAnsi, unused) = (ansi_bg, self.colors[ansi_bg])
return "\033[38;5;{ansiCode:d}m\033[48;5;{bf:d}m{string:s}\033[0m".format(ansiCode=closestAnsi, bf=closestBgAnsi, string=string)
|
[
"def",
"colorize",
"(",
"self",
",",
"string",
",",
"rgb",
"=",
"None",
",",
"ansi",
"=",
"None",
",",
"bg",
"=",
"None",
",",
"ansi_bg",
"=",
"None",
")",
":",
"if",
"not",
"isinstance",
"(",
"string",
",",
"str",
")",
":",
"string",
"=",
"str",
"(",
"string",
")",
"if",
"rgb",
"is",
"None",
"and",
"ansi",
"is",
"None",
":",
"raise",
"TerminalColorMapException",
"(",
"'colorize: must specify one named parameter: rgb or ansi'",
")",
"if",
"rgb",
"is",
"not",
"None",
"and",
"ansi",
"is",
"not",
"None",
":",
"raise",
"TerminalColorMapException",
"(",
"'colorize: must specify only one named parameter: rgb or ansi'",
")",
"if",
"bg",
"is",
"not",
"None",
"and",
"ansi_bg",
"is",
"not",
"None",
":",
"raise",
"TerminalColorMapException",
"(",
"'colorize: must specify only one named parameter: bg or ansi_bg'",
")",
"if",
"rgb",
"is",
"not",
"None",
":",
"(",
"closestAnsi",
",",
"closestRgb",
")",
"=",
"self",
".",
"convert",
"(",
"rgb",
")",
"elif",
"ansi",
"is",
"not",
"None",
":",
"(",
"closestAnsi",
",",
"closestRgb",
")",
"=",
"(",
"ansi",
",",
"self",
".",
"colors",
"[",
"ansi",
"]",
")",
"if",
"bg",
"is",
"None",
"and",
"ansi_bg",
"is",
"None",
":",
"return",
"\"\\033[38;5;{ansiCode:d}m{string:s}\\033[0m\"",
".",
"format",
"(",
"ansiCode",
"=",
"closestAnsi",
",",
"string",
"=",
"string",
")",
"if",
"bg",
"is",
"not",
"None",
":",
"(",
"closestBgAnsi",
",",
"unused",
")",
"=",
"self",
".",
"convert",
"(",
"bg",
")",
"elif",
"ansi_bg",
"is",
"not",
"None",
":",
"(",
"closestBgAnsi",
",",
"unused",
")",
"=",
"(",
"ansi_bg",
",",
"self",
".",
"colors",
"[",
"ansi_bg",
"]",
")",
"return",
"\"\\033[38;5;{ansiCode:d}m\\033[48;5;{bf:d}m{string:s}\\033[0m\"",
".",
"format",
"(",
"ansiCode",
"=",
"closestAnsi",
",",
"bf",
"=",
"closestBgAnsi",
",",
"string",
"=",
"string",
")"
] |
Returns the colored string
|
[
"Returns",
"the",
"colored",
"string"
] |
2615100979eb59744597c9d01e6fd1b711b311e1
|
https://github.com/broadinstitute/xtermcolor/blob/2615100979eb59744597c9d01e6fd1b711b311e1/xtermcolor/ColorMap.py#L27-L54
|
train
|
marrow/WebCore
|
web/ext/serialize.py
|
SerializationExtension.render_serialization
|
def render_serialization(self, context, result):
"""Render serialized responses."""
resp = context.response
serial = context.serialize
match = context.request.accept.best_match(serial.types, default_match=self.default)
result = serial[match](result)
if isinstance(result, str):
result = result.decode('utf-8')
resp.charset = 'utf-8'
resp.content_type = match
resp.text = result
return True
|
python
|
def render_serialization(self, context, result):
"""Render serialized responses."""
resp = context.response
serial = context.serialize
match = context.request.accept.best_match(serial.types, default_match=self.default)
result = serial[match](result)
if isinstance(result, str):
result = result.decode('utf-8')
resp.charset = 'utf-8'
resp.content_type = match
resp.text = result
return True
|
[
"def",
"render_serialization",
"(",
"self",
",",
"context",
",",
"result",
")",
":",
"resp",
"=",
"context",
".",
"response",
"serial",
"=",
"context",
".",
"serialize",
"match",
"=",
"context",
".",
"request",
".",
"accept",
".",
"best_match",
"(",
"serial",
".",
"types",
",",
"default_match",
"=",
"self",
".",
"default",
")",
"result",
"=",
"serial",
"[",
"match",
"]",
"(",
"result",
")",
"if",
"isinstance",
"(",
"result",
",",
"str",
")",
":",
"result",
"=",
"result",
".",
"decode",
"(",
"'utf-8'",
")",
"resp",
".",
"charset",
"=",
"'utf-8'",
"resp",
".",
"content_type",
"=",
"match",
"resp",
".",
"text",
"=",
"result",
"return",
"True"
] |
Render serialized responses.
|
[
"Render",
"serialized",
"responses",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/serialize.py#L91-L106
|
train
|
marrow/WebCore
|
web/server/eventlet_.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080):
"""Eventlet-based WSGI-HTTP server.
For a more fully-featured Eventlet-capable interface, see also [Spawning](http://pypi.python.org/pypi/Spawning/).
"""
# Instantiate the server with a bound port and with our application.
server(listen(host, int(port)), application)
|
python
|
def serve(application, host='127.0.0.1', port=8080):
"""Eventlet-based WSGI-HTTP server.
For a more fully-featured Eventlet-capable interface, see also [Spawning](http://pypi.python.org/pypi/Spawning/).
"""
# Instantiate the server with a bound port and with our application.
server(listen(host, int(port)), application)
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
")",
":",
"# Instantiate the server with a bound port and with our application.",
"server",
"(",
"listen",
"(",
"host",
",",
"int",
"(",
"port",
")",
")",
",",
"application",
")"
] |
Eventlet-based WSGI-HTTP server.
For a more fully-featured Eventlet-capable interface, see also [Spawning](http://pypi.python.org/pypi/Spawning/).
|
[
"Eventlet",
"-",
"based",
"WSGI",
"-",
"HTTP",
"server",
".",
"For",
"a",
"more",
"fully",
"-",
"featured",
"Eventlet",
"-",
"capable",
"interface",
"see",
"also",
"[",
"Spawning",
"]",
"(",
"http",
":",
"//",
"pypi",
".",
"python",
".",
"org",
"/",
"pypi",
"/",
"Spawning",
"/",
")",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/eventlet_.py#L15-L22
|
train
|
NeuroML/pyNeuroML
|
examples/component_evaluation.py
|
main
|
def main(args=None):
"""Main"""
vs = [(v-100)*0.001 for v in range(200)]
for f in ['IM.channel.nml','Kd.channel.nml']:
nml_doc = pynml.read_neuroml2_file(f)
for ct in nml_doc.ComponentType:
ys = []
for v in vs:
req_variables = {'v':'%sV'%v,'vShift':'10mV'}
vals = pynml.evaluate_component(ct,req_variables=req_variables)
print vals
if 'x' in vals:
ys.append(vals['x'])
if 't' in vals:
ys.append(vals['t'])
if 'r' in vals:
ys.append(vals['r'])
ax = pynml.generate_plot([vs],[ys],
"Some traces from %s in %s"%(ct.name,f),
show_plot_already=False )
print vals
plt.show()
|
python
|
def main(args=None):
"""Main"""
vs = [(v-100)*0.001 for v in range(200)]
for f in ['IM.channel.nml','Kd.channel.nml']:
nml_doc = pynml.read_neuroml2_file(f)
for ct in nml_doc.ComponentType:
ys = []
for v in vs:
req_variables = {'v':'%sV'%v,'vShift':'10mV'}
vals = pynml.evaluate_component(ct,req_variables=req_variables)
print vals
if 'x' in vals:
ys.append(vals['x'])
if 't' in vals:
ys.append(vals['t'])
if 'r' in vals:
ys.append(vals['r'])
ax = pynml.generate_plot([vs],[ys],
"Some traces from %s in %s"%(ct.name,f),
show_plot_already=False )
print vals
plt.show()
|
[
"def",
"main",
"(",
"args",
"=",
"None",
")",
":",
"vs",
"=",
"[",
"(",
"v",
"-",
"100",
")",
"*",
"0.001",
"for",
"v",
"in",
"range",
"(",
"200",
")",
"]",
"for",
"f",
"in",
"[",
"'IM.channel.nml'",
",",
"'Kd.channel.nml'",
"]",
":",
"nml_doc",
"=",
"pynml",
".",
"read_neuroml2_file",
"(",
"f",
")",
"for",
"ct",
"in",
"nml_doc",
".",
"ComponentType",
":",
"ys",
"=",
"[",
"]",
"for",
"v",
"in",
"vs",
":",
"req_variables",
"=",
"{",
"'v'",
":",
"'%sV'",
"%",
"v",
",",
"'vShift'",
":",
"'10mV'",
"}",
"vals",
"=",
"pynml",
".",
"evaluate_component",
"(",
"ct",
",",
"req_variables",
"=",
"req_variables",
")",
"print",
"vals",
"if",
"'x'",
"in",
"vals",
":",
"ys",
".",
"append",
"(",
"vals",
"[",
"'x'",
"]",
")",
"if",
"'t'",
"in",
"vals",
":",
"ys",
".",
"append",
"(",
"vals",
"[",
"'t'",
"]",
")",
"if",
"'r'",
"in",
"vals",
":",
"ys",
".",
"append",
"(",
"vals",
"[",
"'r'",
"]",
")",
"ax",
"=",
"pynml",
".",
"generate_plot",
"(",
"[",
"vs",
"]",
",",
"[",
"ys",
"]",
",",
"\"Some traces from %s in %s\"",
"%",
"(",
"ct",
".",
"name",
",",
"f",
")",
",",
"show_plot_already",
"=",
"False",
")",
"print",
"vals",
"plt",
".",
"show",
"(",
")"
] |
Main
|
[
"Main"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/examples/component_evaluation.py#L5-L33
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/analysis/NML2ChannelAnalysis.py
|
process_args
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(
description=("A script which can be run to generate a LEMS "
"file to analyse the behaviour of channels in "
"NeuroML 2"))
parser.add_argument('channelFiles',
type=str,
nargs='+',
metavar='<NeuroML 2 Channel file>',
help="Name of the NeuroML 2 file(s)")
parser.add_argument('-v',
action='store_true',
default=DEFAULTS['v'],
help="Verbose output")
parser.add_argument('-minV',
type=int,
metavar='<min v>',
default=DEFAULTS['minV'],
help="Minimum voltage to test (integer, mV), default: %smV"%DEFAULTS['minV'])
parser.add_argument('-maxV',
type=int,
metavar='<max v>',
default=DEFAULTS['maxV'],
help="Maximum voltage to test (integer, mV), default: %smV"%DEFAULTS['maxV'])
parser.add_argument('-temperature',
type=float,
metavar='<temperature>',
default=DEFAULTS['temperature'],
help="Temperature (float, celsius), default: %sdegC"%DEFAULTS['temperature'])
parser.add_argument('-duration',
type=float,
metavar='<duration>',
default=DEFAULTS['duration'],
help="Duration of simulation in ms, default: %sms"%DEFAULTS['duration'])
parser.add_argument('-clampDelay',
type=float,
metavar='<clamp delay>',
default=DEFAULTS['clampDelay'],
help="Delay before voltage clamp is activated in ms, default: %sms"%DEFAULTS['clampDelay'])
parser.add_argument('-clampDuration',
type=float,
metavar='<clamp duration>',
default=DEFAULTS['clampDuration'],
help="Duration of voltage clamp in ms, default: %sms"%DEFAULTS['clampDuration'])
parser.add_argument('-clampBaseVoltage',
type=float,
metavar='<clamp base voltage>',
default=DEFAULTS['clampBaseVoltage'],
help="Clamp base (starting/finishing) voltage in mV, default: %smV"%DEFAULTS['clampBaseVoltage'])
parser.add_argument('-stepTargetVoltage',
type=float,
metavar='<step target voltage>',
default=DEFAULTS['stepTargetVoltage'],
help=("Voltage in mV through which to step voltage clamps, default: %smV"%DEFAULTS['stepTargetVoltage']))
parser.add_argument('-erev',
type=float,
metavar='<reversal potential>',
default=DEFAULTS['erev'],
help="Reversal potential of channel for currents, default: %smV"%DEFAULTS['erev'])
parser.add_argument('-scaleDt',
type=float,
metavar='<scale dt in generated LEMS>',
default=DEFAULTS['scaleDt'],
help="Scale dt in generated LEMS, default: %s"%DEFAULTS['scaleDt'])
parser.add_argument('-caConc',
type=float,
metavar='<Ca2+ concentration>',
default=DEFAULTS['caConc'],
help=("Internal concentration of Ca2+ (float, "
"concentration in mM), default: %smM"%DEFAULTS['caConc']))
parser.add_argument('-datSuffix',
type=str,
metavar='<dat suffix>',
default=DEFAULTS['datSuffix'],
help="String to add to dat file names (before .dat)")
parser.add_argument('-norun',
action='store_true',
default=DEFAULTS['norun'],
help=("If used, just generate the LEMS file, "
"don't run it"))
parser.add_argument('-nogui',
action='store_true',
default=DEFAULTS['nogui'],
help=("Supress plotting of variables and only save "
"data to file"))
parser.add_argument('-html',
action='store_true',
default=DEFAULTS['html'],
help=("Generate a HTML page featuring the plots for the "
"channel"))
parser.add_argument('-md',
action='store_true',
default=DEFAULTS['md'],
help=("Generate a (GitHub flavoured) Markdown page featuring the plots for the "
"channel"))
parser.add_argument('-ivCurve',
action='store_true',
default=DEFAULTS['ivCurve'],
help=("Save currents through voltage clamp at each "
"level & plot current vs voltage for ion "
"channel"))
return parser.parse_args()
|
python
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(
description=("A script which can be run to generate a LEMS "
"file to analyse the behaviour of channels in "
"NeuroML 2"))
parser.add_argument('channelFiles',
type=str,
nargs='+',
metavar='<NeuroML 2 Channel file>',
help="Name of the NeuroML 2 file(s)")
parser.add_argument('-v',
action='store_true',
default=DEFAULTS['v'],
help="Verbose output")
parser.add_argument('-minV',
type=int,
metavar='<min v>',
default=DEFAULTS['minV'],
help="Minimum voltage to test (integer, mV), default: %smV"%DEFAULTS['minV'])
parser.add_argument('-maxV',
type=int,
metavar='<max v>',
default=DEFAULTS['maxV'],
help="Maximum voltage to test (integer, mV), default: %smV"%DEFAULTS['maxV'])
parser.add_argument('-temperature',
type=float,
metavar='<temperature>',
default=DEFAULTS['temperature'],
help="Temperature (float, celsius), default: %sdegC"%DEFAULTS['temperature'])
parser.add_argument('-duration',
type=float,
metavar='<duration>',
default=DEFAULTS['duration'],
help="Duration of simulation in ms, default: %sms"%DEFAULTS['duration'])
parser.add_argument('-clampDelay',
type=float,
metavar='<clamp delay>',
default=DEFAULTS['clampDelay'],
help="Delay before voltage clamp is activated in ms, default: %sms"%DEFAULTS['clampDelay'])
parser.add_argument('-clampDuration',
type=float,
metavar='<clamp duration>',
default=DEFAULTS['clampDuration'],
help="Duration of voltage clamp in ms, default: %sms"%DEFAULTS['clampDuration'])
parser.add_argument('-clampBaseVoltage',
type=float,
metavar='<clamp base voltage>',
default=DEFAULTS['clampBaseVoltage'],
help="Clamp base (starting/finishing) voltage in mV, default: %smV"%DEFAULTS['clampBaseVoltage'])
parser.add_argument('-stepTargetVoltage',
type=float,
metavar='<step target voltage>',
default=DEFAULTS['stepTargetVoltage'],
help=("Voltage in mV through which to step voltage clamps, default: %smV"%DEFAULTS['stepTargetVoltage']))
parser.add_argument('-erev',
type=float,
metavar='<reversal potential>',
default=DEFAULTS['erev'],
help="Reversal potential of channel for currents, default: %smV"%DEFAULTS['erev'])
parser.add_argument('-scaleDt',
type=float,
metavar='<scale dt in generated LEMS>',
default=DEFAULTS['scaleDt'],
help="Scale dt in generated LEMS, default: %s"%DEFAULTS['scaleDt'])
parser.add_argument('-caConc',
type=float,
metavar='<Ca2+ concentration>',
default=DEFAULTS['caConc'],
help=("Internal concentration of Ca2+ (float, "
"concentration in mM), default: %smM"%DEFAULTS['caConc']))
parser.add_argument('-datSuffix',
type=str,
metavar='<dat suffix>',
default=DEFAULTS['datSuffix'],
help="String to add to dat file names (before .dat)")
parser.add_argument('-norun',
action='store_true',
default=DEFAULTS['norun'],
help=("If used, just generate the LEMS file, "
"don't run it"))
parser.add_argument('-nogui',
action='store_true',
default=DEFAULTS['nogui'],
help=("Supress plotting of variables and only save "
"data to file"))
parser.add_argument('-html',
action='store_true',
default=DEFAULTS['html'],
help=("Generate a HTML page featuring the plots for the "
"channel"))
parser.add_argument('-md',
action='store_true',
default=DEFAULTS['md'],
help=("Generate a (GitHub flavoured) Markdown page featuring the plots for the "
"channel"))
parser.add_argument('-ivCurve',
action='store_true',
default=DEFAULTS['ivCurve'],
help=("Save currents through voltage clamp at each "
"level & plot current vs voltage for ion "
"channel"))
return parser.parse_args()
|
[
"def",
"process_args",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"(",
"\"A script which can be run to generate a LEMS \"",
"\"file to analyse the behaviour of channels in \"",
"\"NeuroML 2\"",
")",
")",
"parser",
".",
"add_argument",
"(",
"'channelFiles'",
",",
"type",
"=",
"str",
",",
"nargs",
"=",
"'+'",
",",
"metavar",
"=",
"'<NeuroML 2 Channel file>'",
",",
"help",
"=",
"\"Name of the NeuroML 2 file(s)\"",
")",
"parser",
".",
"add_argument",
"(",
"'-v'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'v'",
"]",
",",
"help",
"=",
"\"Verbose output\"",
")",
"parser",
".",
"add_argument",
"(",
"'-minV'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<min v>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'minV'",
"]",
",",
"help",
"=",
"\"Minimum voltage to test (integer, mV), default: %smV\"",
"%",
"DEFAULTS",
"[",
"'minV'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-maxV'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<max v>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'maxV'",
"]",
",",
"help",
"=",
"\"Maximum voltage to test (integer, mV), default: %smV\"",
"%",
"DEFAULTS",
"[",
"'maxV'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-temperature'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<temperature>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'temperature'",
"]",
",",
"help",
"=",
"\"Temperature (float, celsius), default: %sdegC\"",
"%",
"DEFAULTS",
"[",
"'temperature'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-duration'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<duration>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'duration'",
"]",
",",
"help",
"=",
"\"Duration of simulation in ms, default: %sms\"",
"%",
"DEFAULTS",
"[",
"'duration'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-clampDelay'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<clamp delay>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'clampDelay'",
"]",
",",
"help",
"=",
"\"Delay before voltage clamp is activated in ms, default: %sms\"",
"%",
"DEFAULTS",
"[",
"'clampDelay'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-clampDuration'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<clamp duration>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'clampDuration'",
"]",
",",
"help",
"=",
"\"Duration of voltage clamp in ms, default: %sms\"",
"%",
"DEFAULTS",
"[",
"'clampDuration'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-clampBaseVoltage'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<clamp base voltage>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'clampBaseVoltage'",
"]",
",",
"help",
"=",
"\"Clamp base (starting/finishing) voltage in mV, default: %smV\"",
"%",
"DEFAULTS",
"[",
"'clampBaseVoltage'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-stepTargetVoltage'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<step target voltage>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'stepTargetVoltage'",
"]",
",",
"help",
"=",
"(",
"\"Voltage in mV through which to step voltage clamps, default: %smV\"",
"%",
"DEFAULTS",
"[",
"'stepTargetVoltage'",
"]",
")",
")",
"parser",
".",
"add_argument",
"(",
"'-erev'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<reversal potential>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'erev'",
"]",
",",
"help",
"=",
"\"Reversal potential of channel for currents, default: %smV\"",
"%",
"DEFAULTS",
"[",
"'erev'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-scaleDt'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<scale dt in generated LEMS>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'scaleDt'",
"]",
",",
"help",
"=",
"\"Scale dt in generated LEMS, default: %s\"",
"%",
"DEFAULTS",
"[",
"'scaleDt'",
"]",
")",
"parser",
".",
"add_argument",
"(",
"'-caConc'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<Ca2+ concentration>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'caConc'",
"]",
",",
"help",
"=",
"(",
"\"Internal concentration of Ca2+ (float, \"",
"\"concentration in mM), default: %smM\"",
"%",
"DEFAULTS",
"[",
"'caConc'",
"]",
")",
")",
"parser",
".",
"add_argument",
"(",
"'-datSuffix'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<dat suffix>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'datSuffix'",
"]",
",",
"help",
"=",
"\"String to add to dat file names (before .dat)\"",
")",
"parser",
".",
"add_argument",
"(",
"'-norun'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'norun'",
"]",
",",
"help",
"=",
"(",
"\"If used, just generate the LEMS file, \"",
"\"don't run it\"",
")",
")",
"parser",
".",
"add_argument",
"(",
"'-nogui'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'nogui'",
"]",
",",
"help",
"=",
"(",
"\"Supress plotting of variables and only save \"",
"\"data to file\"",
")",
")",
"parser",
".",
"add_argument",
"(",
"'-html'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'html'",
"]",
",",
"help",
"=",
"(",
"\"Generate a HTML page featuring the plots for the \"",
"\"channel\"",
")",
")",
"parser",
".",
"add_argument",
"(",
"'-md'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'md'",
"]",
",",
"help",
"=",
"(",
"\"Generate a (GitHub flavoured) Markdown page featuring the plots for the \"",
"\"channel\"",
")",
")",
"parser",
".",
"add_argument",
"(",
"'-ivCurve'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'ivCurve'",
"]",
",",
"help",
"=",
"(",
"\"Save currents through voltage clamp at each \"",
"\"level & plot current vs voltage for ion \"",
"\"channel\"",
")",
")",
"return",
"parser",
".",
"parse_args",
"(",
")"
] |
Parse command-line arguments.
|
[
"Parse",
"command",
"-",
"line",
"arguments",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/analysis/NML2ChannelAnalysis.py#L58-L185
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/analysis/NML2ChannelAnalysis.py
|
plot_iv_curve
|
def plot_iv_curve(a, hold_v, i, *plt_args, **plt_kwargs):
"""A single IV curve"""
grid = plt_kwargs.pop('grid',True)
same_fig = plt_kwargs.pop('same_fig',False)
if not len(plt_args):
plt_args = ('ko-',)
if 'label' not in plt_kwargs:
plt_kwargs['label'] = 'Current'
if not same_fig:
make_iv_curve_fig(a, grid=grid)
if type(i) is dict:
i = [i[v] for v in hold_v]
plt.plot([v*1e3 for v in hold_v], [ii*1e12 for ii in i], *plt_args, **plt_kwargs)
plt.legend(loc=2)
|
python
|
def plot_iv_curve(a, hold_v, i, *plt_args, **plt_kwargs):
"""A single IV curve"""
grid = plt_kwargs.pop('grid',True)
same_fig = plt_kwargs.pop('same_fig',False)
if not len(plt_args):
plt_args = ('ko-',)
if 'label' not in plt_kwargs:
plt_kwargs['label'] = 'Current'
if not same_fig:
make_iv_curve_fig(a, grid=grid)
if type(i) is dict:
i = [i[v] for v in hold_v]
plt.plot([v*1e3 for v in hold_v], [ii*1e12 for ii in i], *plt_args, **plt_kwargs)
plt.legend(loc=2)
|
[
"def",
"plot_iv_curve",
"(",
"a",
",",
"hold_v",
",",
"i",
",",
"*",
"plt_args",
",",
"*",
"*",
"plt_kwargs",
")",
":",
"grid",
"=",
"plt_kwargs",
".",
"pop",
"(",
"'grid'",
",",
"True",
")",
"same_fig",
"=",
"plt_kwargs",
".",
"pop",
"(",
"'same_fig'",
",",
"False",
")",
"if",
"not",
"len",
"(",
"plt_args",
")",
":",
"plt_args",
"=",
"(",
"'ko-'",
",",
")",
"if",
"'label'",
"not",
"in",
"plt_kwargs",
":",
"plt_kwargs",
"[",
"'label'",
"]",
"=",
"'Current'",
"if",
"not",
"same_fig",
":",
"make_iv_curve_fig",
"(",
"a",
",",
"grid",
"=",
"grid",
")",
"if",
"type",
"(",
"i",
")",
"is",
"dict",
":",
"i",
"=",
"[",
"i",
"[",
"v",
"]",
"for",
"v",
"in",
"hold_v",
"]",
"plt",
".",
"plot",
"(",
"[",
"v",
"*",
"1e3",
"for",
"v",
"in",
"hold_v",
"]",
",",
"[",
"ii",
"*",
"1e12",
"for",
"ii",
"in",
"i",
"]",
",",
"*",
"plt_args",
",",
"*",
"*",
"plt_kwargs",
")",
"plt",
".",
"legend",
"(",
"loc",
"=",
"2",
")"
] |
A single IV curve
|
[
"A",
"single",
"IV",
"curve"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/analysis/NML2ChannelAnalysis.py#L581-L594
|
train
|
marrow/WebCore
|
example/stream.py
|
root
|
def root(context):
"""Multipart AJAX request example.
See: http://test.getify.com/mpAjax/description.html
"""
response = context.response
parts = []
for i in range(12):
for j in range(12):
parts.append(executor.submit(mul, i, j))
def stream(parts, timeout=None):
try:
for future in as_completed(parts, timeout):
mime, result = future.result()
result = result.encode('utf8')
yield "!!!!!!=_NextPart_{num}\nContent-Type: {mime}\nContent-Length: {length}\n\n".format(
num = randint(100000000, 999999999),
mime = mime,
length = len(result)
).encode('utf8') + result
except TimeoutError:
for future in parts:
future.cancel()
response.content_length = None
response.app_iter = stream(parts, 0.2)
return response
|
python
|
def root(context):
"""Multipart AJAX request example.
See: http://test.getify.com/mpAjax/description.html
"""
response = context.response
parts = []
for i in range(12):
for j in range(12):
parts.append(executor.submit(mul, i, j))
def stream(parts, timeout=None):
try:
for future in as_completed(parts, timeout):
mime, result = future.result()
result = result.encode('utf8')
yield "!!!!!!=_NextPart_{num}\nContent-Type: {mime}\nContent-Length: {length}\n\n".format(
num = randint(100000000, 999999999),
mime = mime,
length = len(result)
).encode('utf8') + result
except TimeoutError:
for future in parts:
future.cancel()
response.content_length = None
response.app_iter = stream(parts, 0.2)
return response
|
[
"def",
"root",
"(",
"context",
")",
":",
"response",
"=",
"context",
".",
"response",
"parts",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"12",
")",
":",
"for",
"j",
"in",
"range",
"(",
"12",
")",
":",
"parts",
".",
"append",
"(",
"executor",
".",
"submit",
"(",
"mul",
",",
"i",
",",
"j",
")",
")",
"def",
"stream",
"(",
"parts",
",",
"timeout",
"=",
"None",
")",
":",
"try",
":",
"for",
"future",
"in",
"as_completed",
"(",
"parts",
",",
"timeout",
")",
":",
"mime",
",",
"result",
"=",
"future",
".",
"result",
"(",
")",
"result",
"=",
"result",
".",
"encode",
"(",
"'utf8'",
")",
"yield",
"\"!!!!!!=_NextPart_{num}\\nContent-Type: {mime}\\nContent-Length: {length}\\n\\n\"",
".",
"format",
"(",
"num",
"=",
"randint",
"(",
"100000000",
",",
"999999999",
")",
",",
"mime",
"=",
"mime",
",",
"length",
"=",
"len",
"(",
"result",
")",
")",
".",
"encode",
"(",
"'utf8'",
")",
"+",
"result",
"except",
"TimeoutError",
":",
"for",
"future",
"in",
"parts",
":",
"future",
".",
"cancel",
"(",
")",
"response",
".",
"content_length",
"=",
"None",
"response",
".",
"app_iter",
"=",
"stream",
"(",
"parts",
",",
"0.2",
")",
"return",
"response"
] |
Multipart AJAX request example.
See: http://test.getify.com/mpAjax/description.html
|
[
"Multipart",
"AJAX",
"request",
"example",
".",
"See",
":",
"http",
":",
"//",
"test",
".",
"getify",
".",
"com",
"/",
"mpAjax",
"/",
"description",
".",
"html"
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/example/stream.py#L16-L49
|
train
|
contraslash/django-crud-generator
|
django_crud_generator/django_crud_generator.py
|
render_template_with_args_in_file
|
def render_template_with_args_in_file(file, template_file_name, **kwargs):
"""
Get a file and render the content of the template_file_name with kwargs in a file
:param file: A File Stream to write
:param template_file_name: path to route with template name
:param **kwargs: Args to be rendered in template
"""
template_file_content = "".join(
codecs.open(
template_file_name,
encoding='UTF-8'
).readlines()
)
template_rendered = string.Template(template_file_content).safe_substitute(**kwargs)
file.write(template_rendered)
|
python
|
def render_template_with_args_in_file(file, template_file_name, **kwargs):
"""
Get a file and render the content of the template_file_name with kwargs in a file
:param file: A File Stream to write
:param template_file_name: path to route with template name
:param **kwargs: Args to be rendered in template
"""
template_file_content = "".join(
codecs.open(
template_file_name,
encoding='UTF-8'
).readlines()
)
template_rendered = string.Template(template_file_content).safe_substitute(**kwargs)
file.write(template_rendered)
|
[
"def",
"render_template_with_args_in_file",
"(",
"file",
",",
"template_file_name",
",",
"*",
"*",
"kwargs",
")",
":",
"template_file_content",
"=",
"\"\"",
".",
"join",
"(",
"codecs",
".",
"open",
"(",
"template_file_name",
",",
"encoding",
"=",
"'UTF-8'",
")",
".",
"readlines",
"(",
")",
")",
"template_rendered",
"=",
"string",
".",
"Template",
"(",
"template_file_content",
")",
".",
"safe_substitute",
"(",
"*",
"*",
"kwargs",
")",
"file",
".",
"write",
"(",
"template_rendered",
")"
] |
Get a file and render the content of the template_file_name with kwargs in a file
:param file: A File Stream to write
:param template_file_name: path to route with template name
:param **kwargs: Args to be rendered in template
|
[
"Get",
"a",
"file",
"and",
"render",
"the",
"content",
"of",
"the",
"template_file_name",
"with",
"kwargs",
"in",
"a",
"file",
":",
"param",
"file",
":",
"A",
"File",
"Stream",
"to",
"write",
":",
"param",
"template_file_name",
":",
"path",
"to",
"route",
"with",
"template",
"name",
":",
"param",
"**",
"kwargs",
":",
"Args",
"to",
"be",
"rendered",
"in",
"template"
] |
9080a227291a36cf0d93b49d20601e55a616da1e
|
https://github.com/contraslash/django-crud-generator/blob/9080a227291a36cf0d93b49d20601e55a616da1e/django_crud_generator/django_crud_generator.py#L20-L34
|
train
|
contraslash/django-crud-generator
|
django_crud_generator/django_crud_generator.py
|
create_or_open
|
def create_or_open(file_name, initial_template_file_name, args):
"""
Creates a file or open the file with file_name name
:param file_name: String with a filename
:param initial_template_file_name: String with path to initial template
:param args: from console to determine path to save the files
"""
file = None
if not os.path.isfile(
os.path.join(
args['django_application_folder'],
file_name
)
):
# If file_name does not exists, create
file = codecs.open(
os.path.join(
args['django_application_folder'],
file_name
),
'w+',
encoding='UTF-8'
)
print("Creating {}".format(file_name))
if initial_template_file_name:
render_template_with_args_in_file(file, initial_template_file_name, **{})
else:
# If file exists, just load the file
file = codecs.open(
os.path.join(
args['django_application_folder'],
file_name
),
'a+',
encoding='UTF-8'
)
return file
|
python
|
def create_or_open(file_name, initial_template_file_name, args):
"""
Creates a file or open the file with file_name name
:param file_name: String with a filename
:param initial_template_file_name: String with path to initial template
:param args: from console to determine path to save the files
"""
file = None
if not os.path.isfile(
os.path.join(
args['django_application_folder'],
file_name
)
):
# If file_name does not exists, create
file = codecs.open(
os.path.join(
args['django_application_folder'],
file_name
),
'w+',
encoding='UTF-8'
)
print("Creating {}".format(file_name))
if initial_template_file_name:
render_template_with_args_in_file(file, initial_template_file_name, **{})
else:
# If file exists, just load the file
file = codecs.open(
os.path.join(
args['django_application_folder'],
file_name
),
'a+',
encoding='UTF-8'
)
return file
|
[
"def",
"create_or_open",
"(",
"file_name",
",",
"initial_template_file_name",
",",
"args",
")",
":",
"file",
"=",
"None",
"if",
"not",
"os",
".",
"path",
".",
"isfile",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"file_name",
")",
")",
":",
"# If file_name does not exists, create",
"file",
"=",
"codecs",
".",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"file_name",
")",
",",
"'w+'",
",",
"encoding",
"=",
"'UTF-8'",
")",
"print",
"(",
"\"Creating {}\"",
".",
"format",
"(",
"file_name",
")",
")",
"if",
"initial_template_file_name",
":",
"render_template_with_args_in_file",
"(",
"file",
",",
"initial_template_file_name",
",",
"*",
"*",
"{",
"}",
")",
"else",
":",
"# If file exists, just load the file",
"file",
"=",
"codecs",
".",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"file_name",
")",
",",
"'a+'",
",",
"encoding",
"=",
"'UTF-8'",
")",
"return",
"file"
] |
Creates a file or open the file with file_name name
:param file_name: String with a filename
:param initial_template_file_name: String with path to initial template
:param args: from console to determine path to save the files
|
[
"Creates",
"a",
"file",
"or",
"open",
"the",
"file",
"with",
"file_name",
"name",
":",
"param",
"file_name",
":",
"String",
"with",
"a",
"filename",
":",
"param",
"initial_template_file_name",
":",
"String",
"with",
"path",
"to",
"initial",
"template",
":",
"param",
"args",
":",
"from",
"console",
"to",
"determine",
"path",
"to",
"save",
"the",
"files"
] |
9080a227291a36cf0d93b49d20601e55a616da1e
|
https://github.com/contraslash/django-crud-generator/blob/9080a227291a36cf0d93b49d20601e55a616da1e/django_crud_generator/django_crud_generator.py#L37-L74
|
train
|
contraslash/django-crud-generator
|
django_crud_generator/django_crud_generator.py
|
generic_insert_module
|
def generic_insert_module(module_name, args, **kwargs):
"""
In general we have a initial template and then insert new data, so we dont repeat the schema for each module
:param module_name: String with module name
:paran **kwargs: Args to be rendered in template
"""
file = create_or_open(
'{}.py'.format(module_name),
os.path.join(
BASE_TEMPLATES_DIR,
'{}_initial.py.tmpl'.format(module_name)
),
args
)
render_template_with_args_in_file(
file,
os.path.join(
BASE_TEMPLATES_DIR,
'{}.py.tmpl'.format(module_name)
),
**kwargs
)
file.close()
|
python
|
def generic_insert_module(module_name, args, **kwargs):
"""
In general we have a initial template and then insert new data, so we dont repeat the schema for each module
:param module_name: String with module name
:paran **kwargs: Args to be rendered in template
"""
file = create_or_open(
'{}.py'.format(module_name),
os.path.join(
BASE_TEMPLATES_DIR,
'{}_initial.py.tmpl'.format(module_name)
),
args
)
render_template_with_args_in_file(
file,
os.path.join(
BASE_TEMPLATES_DIR,
'{}.py.tmpl'.format(module_name)
),
**kwargs
)
file.close()
|
[
"def",
"generic_insert_module",
"(",
"module_name",
",",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"file",
"=",
"create_or_open",
"(",
"'{}.py'",
".",
"format",
"(",
"module_name",
")",
",",
"os",
".",
"path",
".",
"join",
"(",
"BASE_TEMPLATES_DIR",
",",
"'{}_initial.py.tmpl'",
".",
"format",
"(",
"module_name",
")",
")",
",",
"args",
")",
"render_template_with_args_in_file",
"(",
"file",
",",
"os",
".",
"path",
".",
"join",
"(",
"BASE_TEMPLATES_DIR",
",",
"'{}.py.tmpl'",
".",
"format",
"(",
"module_name",
")",
")",
",",
"*",
"*",
"kwargs",
")",
"file",
".",
"close",
"(",
")"
] |
In general we have a initial template and then insert new data, so we dont repeat the schema for each module
:param module_name: String with module name
:paran **kwargs: Args to be rendered in template
|
[
"In",
"general",
"we",
"have",
"a",
"initial",
"template",
"and",
"then",
"insert",
"new",
"data",
"so",
"we",
"dont",
"repeat",
"the",
"schema",
"for",
"each",
"module",
":",
"param",
"module_name",
":",
"String",
"with",
"module",
"name",
":",
"paran",
"**",
"kwargs",
":",
"Args",
"to",
"be",
"rendered",
"in",
"template"
] |
9080a227291a36cf0d93b49d20601e55a616da1e
|
https://github.com/contraslash/django-crud-generator/blob/9080a227291a36cf0d93b49d20601e55a616da1e/django_crud_generator/django_crud_generator.py#L77-L100
|
train
|
contraslash/django-crud-generator
|
django_crud_generator/django_crud_generator.py
|
sanity_check
|
def sanity_check(args):
"""
Verify if the work folder is a django app.
A valid django app always must have a models.py file
:return: None
"""
if not os.path.isfile(
os.path.join(
args['django_application_folder'],
'models.py'
)
):
print("django_application_folder is not a Django application folder")
sys.exit(1)
|
python
|
def sanity_check(args):
"""
Verify if the work folder is a django app.
A valid django app always must have a models.py file
:return: None
"""
if not os.path.isfile(
os.path.join(
args['django_application_folder'],
'models.py'
)
):
print("django_application_folder is not a Django application folder")
sys.exit(1)
|
[
"def",
"sanity_check",
"(",
"args",
")",
":",
"if",
"not",
"os",
".",
"path",
".",
"isfile",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"'models.py'",
")",
")",
":",
"print",
"(",
"\"django_application_folder is not a Django application folder\"",
")",
"sys",
".",
"exit",
"(",
"1",
")"
] |
Verify if the work folder is a django app.
A valid django app always must have a models.py file
:return: None
|
[
"Verify",
"if",
"the",
"work",
"folder",
"is",
"a",
"django",
"app",
".",
"A",
"valid",
"django",
"app",
"always",
"must",
"have",
"a",
"models",
".",
"py",
"file",
":",
"return",
":",
"None"
] |
9080a227291a36cf0d93b49d20601e55a616da1e
|
https://github.com/contraslash/django-crud-generator/blob/9080a227291a36cf0d93b49d20601e55a616da1e/django_crud_generator/django_crud_generator.py#L103-L116
|
train
|
contraslash/django-crud-generator
|
django_crud_generator/django_crud_generator.py
|
generic_insert_with_folder
|
def generic_insert_with_folder(folder_name, file_name, template_name, args):
"""
In general if we need to put a file on a folder, we use this method
"""
# First we make sure views are a package instead a file
if not os.path.isdir(
os.path.join(
args['django_application_folder'],
folder_name
)
):
os.mkdir(os.path.join(args['django_application_folder'], folder_name))
codecs.open(
os.path.join(
args['django_application_folder'],
folder_name,
'__init__.py'
),
'w+'
)
view_file = create_or_open(
os.path.join(
folder_name,
'{}.py'.format(file_name)
),
'',
args
)
# Load content from template
render_template_with_args_in_file(
view_file,
os.path.join(
BASE_TEMPLATES_DIR,
template_name
),
model_name=args['model_name'],
model_prefix=args['model_prefix'],
model_name_lower=args['model_name'].lower(),
application_name=args['django_application_folder'].split("/")[-1]
)
view_file.close()
|
python
|
def generic_insert_with_folder(folder_name, file_name, template_name, args):
"""
In general if we need to put a file on a folder, we use this method
"""
# First we make sure views are a package instead a file
if not os.path.isdir(
os.path.join(
args['django_application_folder'],
folder_name
)
):
os.mkdir(os.path.join(args['django_application_folder'], folder_name))
codecs.open(
os.path.join(
args['django_application_folder'],
folder_name,
'__init__.py'
),
'w+'
)
view_file = create_or_open(
os.path.join(
folder_name,
'{}.py'.format(file_name)
),
'',
args
)
# Load content from template
render_template_with_args_in_file(
view_file,
os.path.join(
BASE_TEMPLATES_DIR,
template_name
),
model_name=args['model_name'],
model_prefix=args['model_prefix'],
model_name_lower=args['model_name'].lower(),
application_name=args['django_application_folder'].split("/")[-1]
)
view_file.close()
|
[
"def",
"generic_insert_with_folder",
"(",
"folder_name",
",",
"file_name",
",",
"template_name",
",",
"args",
")",
":",
"# First we make sure views are a package instead a file",
"if",
"not",
"os",
".",
"path",
".",
"isdir",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"folder_name",
")",
")",
":",
"os",
".",
"mkdir",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"folder_name",
")",
")",
"codecs",
".",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"args",
"[",
"'django_application_folder'",
"]",
",",
"folder_name",
",",
"'__init__.py'",
")",
",",
"'w+'",
")",
"view_file",
"=",
"create_or_open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"folder_name",
",",
"'{}.py'",
".",
"format",
"(",
"file_name",
")",
")",
",",
"''",
",",
"args",
")",
"# Load content from template",
"render_template_with_args_in_file",
"(",
"view_file",
",",
"os",
".",
"path",
".",
"join",
"(",
"BASE_TEMPLATES_DIR",
",",
"template_name",
")",
",",
"model_name",
"=",
"args",
"[",
"'model_name'",
"]",
",",
"model_prefix",
"=",
"args",
"[",
"'model_prefix'",
"]",
",",
"model_name_lower",
"=",
"args",
"[",
"'model_name'",
"]",
".",
"lower",
"(",
")",
",",
"application_name",
"=",
"args",
"[",
"'django_application_folder'",
"]",
".",
"split",
"(",
"\"/\"",
")",
"[",
"-",
"1",
"]",
")",
"view_file",
".",
"close",
"(",
")"
] |
In general if we need to put a file on a folder, we use this method
|
[
"In",
"general",
"if",
"we",
"need",
"to",
"put",
"a",
"file",
"on",
"a",
"folder",
"we",
"use",
"this",
"method"
] |
9080a227291a36cf0d93b49d20601e55a616da1e
|
https://github.com/contraslash/django-crud-generator/blob/9080a227291a36cf0d93b49d20601e55a616da1e/django_crud_generator/django_crud_generator.py#L119-L160
|
train
|
marrow/WebCore
|
web/server/waitress_.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080, threads=4, **kw):
"""The recommended development HTTP server.
Note that this server performs additional buffering and will not honour chunked encoding breaks.
"""
# Bind and start the server; this is a blocking process.
serve_(application, host=host, port=int(port), threads=int(threads), **kw)
|
python
|
def serve(application, host='127.0.0.1', port=8080, threads=4, **kw):
"""The recommended development HTTP server.
Note that this server performs additional buffering and will not honour chunked encoding breaks.
"""
# Bind and start the server; this is a blocking process.
serve_(application, host=host, port=int(port), threads=int(threads), **kw)
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
",",
"threads",
"=",
"4",
",",
"*",
"*",
"kw",
")",
":",
"# Bind and start the server; this is a blocking process.",
"serve_",
"(",
"application",
",",
"host",
"=",
"host",
",",
"port",
"=",
"int",
"(",
"port",
")",
",",
"threads",
"=",
"int",
"(",
"threads",
")",
",",
"*",
"*",
"kw",
")"
] |
The recommended development HTTP server.
Note that this server performs additional buffering and will not honour chunked encoding breaks.
|
[
"The",
"recommended",
"development",
"HTTP",
"server",
".",
"Note",
"that",
"this",
"server",
"performs",
"additional",
"buffering",
"and",
"will",
"not",
"honour",
"chunked",
"encoding",
"breaks",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/waitress_.py#L18-L25
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/tune/NeuroMLSimulation.py
|
NeuroMLSimulation.show
|
def show(self):
"""
Plot the result of the simulation once it's been intialized
"""
from matplotlib import pyplot as plt
if self.already_run:
for ref in self.volts.keys():
plt.plot(self.t, self.volts[ref], label=ref)
plt.title("Simulation voltage vs time")
plt.legend()
plt.xlabel("Time [ms]")
plt.ylabel("Voltage [mV]")
else:
pynml.print_comment("First you have to 'go()' the simulation.", True)
plt.show()
|
python
|
def show(self):
"""
Plot the result of the simulation once it's been intialized
"""
from matplotlib import pyplot as plt
if self.already_run:
for ref in self.volts.keys():
plt.plot(self.t, self.volts[ref], label=ref)
plt.title("Simulation voltage vs time")
plt.legend()
plt.xlabel("Time [ms]")
plt.ylabel("Voltage [mV]")
else:
pynml.print_comment("First you have to 'go()' the simulation.", True)
plt.show()
|
[
"def",
"show",
"(",
"self",
")",
":",
"from",
"matplotlib",
"import",
"pyplot",
"as",
"plt",
"if",
"self",
".",
"already_run",
":",
"for",
"ref",
"in",
"self",
".",
"volts",
".",
"keys",
"(",
")",
":",
"plt",
".",
"plot",
"(",
"self",
".",
"t",
",",
"self",
".",
"volts",
"[",
"ref",
"]",
",",
"label",
"=",
"ref",
")",
"plt",
".",
"title",
"(",
"\"Simulation voltage vs time\"",
")",
"plt",
".",
"legend",
"(",
")",
"plt",
".",
"xlabel",
"(",
"\"Time [ms]\"",
")",
"plt",
".",
"ylabel",
"(",
"\"Voltage [mV]\"",
")",
"else",
":",
"pynml",
".",
"print_comment",
"(",
"\"First you have to 'go()' the simulation.\"",
",",
"True",
")",
"plt",
".",
"show",
"(",
")"
] |
Plot the result of the simulation once it's been intialized
|
[
"Plot",
"the",
"result",
"of",
"the",
"simulation",
"once",
"it",
"s",
"been",
"intialized"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/tune/NeuroMLSimulation.py#L58-L77
|
train
|
marrow/WebCore
|
example/annotation.py
|
Root.mul
|
def mul(self, a: int = None, b: int = None) -> 'json':
"""Multiply two values together and return the result via JSON.
Python 3 function annotations are used to ensure that the arguments are integers. This requires the
functionality of `web.ext.annotation:AnnotationExtension`.
There are several ways to execute this method:
* POST http://localhost:8080/mul
* GET http://localhost:8080/mul?a=27&b=42
* GET http://localhost:8080/mul/27/42
The latter relies on the fact we can't descend past a callable method so the remaining path elements are
used as positional arguments, whereas the others rely on keyword argument assignment from a form-encoded
request body or query string arguments. (Security note: any data in the request body takes presidence over
query string arguments!)
You can easily test these on the command line using cURL:
curl http://localhost:8080/mul/27/42 # HTTP GET
curl -d a=27 -d b=42 http://localhost:8080/mul # HTTP POST
"""
if not a or not b:
return dict(message="Pass arguments a and b to multiply them together!")
return dict(answer=a * b)
|
python
|
def mul(self, a: int = None, b: int = None) -> 'json':
"""Multiply two values together and return the result via JSON.
Python 3 function annotations are used to ensure that the arguments are integers. This requires the
functionality of `web.ext.annotation:AnnotationExtension`.
There are several ways to execute this method:
* POST http://localhost:8080/mul
* GET http://localhost:8080/mul?a=27&b=42
* GET http://localhost:8080/mul/27/42
The latter relies on the fact we can't descend past a callable method so the remaining path elements are
used as positional arguments, whereas the others rely on keyword argument assignment from a form-encoded
request body or query string arguments. (Security note: any data in the request body takes presidence over
query string arguments!)
You can easily test these on the command line using cURL:
curl http://localhost:8080/mul/27/42 # HTTP GET
curl -d a=27 -d b=42 http://localhost:8080/mul # HTTP POST
"""
if not a or not b:
return dict(message="Pass arguments a and b to multiply them together!")
return dict(answer=a * b)
|
[
"def",
"mul",
"(",
"self",
",",
"a",
":",
"int",
"=",
"None",
",",
"b",
":",
"int",
"=",
"None",
")",
"->",
"'json'",
":",
"if",
"not",
"a",
"or",
"not",
"b",
":",
"return",
"dict",
"(",
"message",
"=",
"\"Pass arguments a and b to multiply them together!\"",
")",
"return",
"dict",
"(",
"answer",
"=",
"a",
"*",
"b",
")"
] |
Multiply two values together and return the result via JSON.
Python 3 function annotations are used to ensure that the arguments are integers. This requires the
functionality of `web.ext.annotation:AnnotationExtension`.
There are several ways to execute this method:
* POST http://localhost:8080/mul
* GET http://localhost:8080/mul?a=27&b=42
* GET http://localhost:8080/mul/27/42
The latter relies on the fact we can't descend past a callable method so the remaining path elements are
used as positional arguments, whereas the others rely on keyword argument assignment from a form-encoded
request body or query string arguments. (Security note: any data in the request body takes presidence over
query string arguments!)
You can easily test these on the command line using cURL:
curl http://localhost:8080/mul/27/42 # HTTP GET
curl -d a=27 -d b=42 http://localhost:8080/mul # HTTP POST
|
[
"Multiply",
"two",
"values",
"together",
"and",
"return",
"the",
"result",
"via",
"JSON",
".",
"Python",
"3",
"function",
"annotations",
"are",
"used",
"to",
"ensure",
"that",
"the",
"arguments",
"are",
"integers",
".",
"This",
"requires",
"the",
"functionality",
"of",
"web",
".",
"ext",
".",
"annotation",
":",
"AnnotationExtension",
".",
"There",
"are",
"several",
"ways",
"to",
"execute",
"this",
"method",
":",
"*",
"POST",
"http",
":",
"//",
"localhost",
":",
"8080",
"/",
"mul",
"*",
"GET",
"http",
":",
"//",
"localhost",
":",
"8080",
"/",
"mul?a",
"=",
"27&b",
"=",
"42",
"*",
"GET",
"http",
":",
"//",
"localhost",
":",
"8080",
"/",
"mul",
"/",
"27",
"/",
"42",
"The",
"latter",
"relies",
"on",
"the",
"fact",
"we",
"can",
"t",
"descend",
"past",
"a",
"callable",
"method",
"so",
"the",
"remaining",
"path",
"elements",
"are",
"used",
"as",
"positional",
"arguments",
"whereas",
"the",
"others",
"rely",
"on",
"keyword",
"argument",
"assignment",
"from",
"a",
"form",
"-",
"encoded",
"request",
"body",
"or",
"query",
"string",
"arguments",
".",
"(",
"Security",
"note",
":",
"any",
"data",
"in",
"the",
"request",
"body",
"takes",
"presidence",
"over",
"query",
"string",
"arguments!",
")",
"You",
"can",
"easily",
"test",
"these",
"on",
"the",
"command",
"line",
"using",
"cURL",
":",
"curl",
"http",
":",
"//",
"localhost",
":",
"8080",
"/",
"mul",
"/",
"27",
"/",
"42",
"#",
"HTTP",
"GET",
"curl",
"-",
"d",
"a",
"=",
"27",
"-",
"d",
"b",
"=",
"42",
"http",
":",
"//",
"localhost",
":",
"8080",
"/",
"mul",
"#",
"HTTP",
"POST"
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/example/annotation.py#L15-L42
|
train
|
broadinstitute/xtermcolor
|
xtermcolor/__init__.py
|
colorize
|
def colorize(string, rgb=None, ansi=None, bg=None, ansi_bg=None, fd=1):
'''Returns the colored string to print on the terminal.
This function detects the terminal type and if it is supported and the
output is not going to a pipe or a file, then it will return the colored
string, otherwise it will return the string without modifications.
string = the string to print. Only accepts strings, unicode strings must
be encoded in advance.
rgb = Rgb color for the text; for example 0xFF0000 is red.
ansi = Ansi for the text
bg = Rgb color for the background
ansi_bg= Ansi color for the background
fd = The file descriptor that will be used by print, by default is the
stdout
'''
#Reinitializes if fd used is different
if colorize.fd != fd:
colorize.init = False
colorize.fd = fd
#Checks if it is on a terminal, and if the terminal is recognized
if not colorize.init:
colorize.init = True
colorize.is_term = isatty(fd)
if 'TERM' in environ:
if environ['TERM'].startswith('xterm'):
colorize.cmap = XTermColorMap()
elif environ['TERM'] == 'vt100':
colorize.cmap = VT100ColorMap()
else:
colorize.is_term = False
else:
colorize.is_term = False
if colorize.is_term:
string = colorize.cmap.colorize(string, rgb, ansi, bg, ansi_bg)
return string
|
python
|
def colorize(string, rgb=None, ansi=None, bg=None, ansi_bg=None, fd=1):
'''Returns the colored string to print on the terminal.
This function detects the terminal type and if it is supported and the
output is not going to a pipe or a file, then it will return the colored
string, otherwise it will return the string without modifications.
string = the string to print. Only accepts strings, unicode strings must
be encoded in advance.
rgb = Rgb color for the text; for example 0xFF0000 is red.
ansi = Ansi for the text
bg = Rgb color for the background
ansi_bg= Ansi color for the background
fd = The file descriptor that will be used by print, by default is the
stdout
'''
#Reinitializes if fd used is different
if colorize.fd != fd:
colorize.init = False
colorize.fd = fd
#Checks if it is on a terminal, and if the terminal is recognized
if not colorize.init:
colorize.init = True
colorize.is_term = isatty(fd)
if 'TERM' in environ:
if environ['TERM'].startswith('xterm'):
colorize.cmap = XTermColorMap()
elif environ['TERM'] == 'vt100':
colorize.cmap = VT100ColorMap()
else:
colorize.is_term = False
else:
colorize.is_term = False
if colorize.is_term:
string = colorize.cmap.colorize(string, rgb, ansi, bg, ansi_bg)
return string
|
[
"def",
"colorize",
"(",
"string",
",",
"rgb",
"=",
"None",
",",
"ansi",
"=",
"None",
",",
"bg",
"=",
"None",
",",
"ansi_bg",
"=",
"None",
",",
"fd",
"=",
"1",
")",
":",
"#Reinitializes if fd used is different",
"if",
"colorize",
".",
"fd",
"!=",
"fd",
":",
"colorize",
".",
"init",
"=",
"False",
"colorize",
".",
"fd",
"=",
"fd",
"#Checks if it is on a terminal, and if the terminal is recognized",
"if",
"not",
"colorize",
".",
"init",
":",
"colorize",
".",
"init",
"=",
"True",
"colorize",
".",
"is_term",
"=",
"isatty",
"(",
"fd",
")",
"if",
"'TERM'",
"in",
"environ",
":",
"if",
"environ",
"[",
"'TERM'",
"]",
".",
"startswith",
"(",
"'xterm'",
")",
":",
"colorize",
".",
"cmap",
"=",
"XTermColorMap",
"(",
")",
"elif",
"environ",
"[",
"'TERM'",
"]",
"==",
"'vt100'",
":",
"colorize",
".",
"cmap",
"=",
"VT100ColorMap",
"(",
")",
"else",
":",
"colorize",
".",
"is_term",
"=",
"False",
"else",
":",
"colorize",
".",
"is_term",
"=",
"False",
"if",
"colorize",
".",
"is_term",
":",
"string",
"=",
"colorize",
".",
"cmap",
".",
"colorize",
"(",
"string",
",",
"rgb",
",",
"ansi",
",",
"bg",
",",
"ansi_bg",
")",
"return",
"string"
] |
Returns the colored string to print on the terminal.
This function detects the terminal type and if it is supported and the
output is not going to a pipe or a file, then it will return the colored
string, otherwise it will return the string without modifications.
string = the string to print. Only accepts strings, unicode strings must
be encoded in advance.
rgb = Rgb color for the text; for example 0xFF0000 is red.
ansi = Ansi for the text
bg = Rgb color for the background
ansi_bg= Ansi color for the background
fd = The file descriptor that will be used by print, by default is the
stdout
|
[
"Returns",
"the",
"colored",
"string",
"to",
"print",
"on",
"the",
"terminal",
"."
] |
2615100979eb59744597c9d01e6fd1b711b311e1
|
https://github.com/broadinstitute/xtermcolor/blob/2615100979eb59744597c9d01e6fd1b711b311e1/xtermcolor/__init__.py#L6-L45
|
train
|
marrow/WebCore
|
web/ext/annotation.py
|
AnnotationExtension.mutate
|
def mutate(self, context, handler, args, kw):
"""Inspect and potentially mutate the given handler's arguments.
The args list and kw dictionary may be freely modified, though invalid arguments to the handler will fail.
"""
def cast(arg, val):
if arg not in annotations:
return
cast = annotations[key]
try:
val = cast(val)
except (ValueError, TypeError) as e:
parts = list(e.args)
parts[0] = parts[0] + " processing argument '{}'".format(arg)
e.args = tuple(parts)
raise
return val
annotations = getattr(handler.__func__ if hasattr(handler, '__func__') else handler, '__annotations__', None)
if not annotations:
return
argspec = getfullargspec(handler)
arglist = list(argspec.args)
if ismethod(handler):
del arglist[0]
for i, value in enumerate(list(args)):
key = arglist[i]
if key in annotations:
args[i] = cast(key, value)
# Convert keyword arguments
for key, value in list(items(kw)):
if key in annotations:
kw[key] = cast(key, value)
|
python
|
def mutate(self, context, handler, args, kw):
"""Inspect and potentially mutate the given handler's arguments.
The args list and kw dictionary may be freely modified, though invalid arguments to the handler will fail.
"""
def cast(arg, val):
if arg not in annotations:
return
cast = annotations[key]
try:
val = cast(val)
except (ValueError, TypeError) as e:
parts = list(e.args)
parts[0] = parts[0] + " processing argument '{}'".format(arg)
e.args = tuple(parts)
raise
return val
annotations = getattr(handler.__func__ if hasattr(handler, '__func__') else handler, '__annotations__', None)
if not annotations:
return
argspec = getfullargspec(handler)
arglist = list(argspec.args)
if ismethod(handler):
del arglist[0]
for i, value in enumerate(list(args)):
key = arglist[i]
if key in annotations:
args[i] = cast(key, value)
# Convert keyword arguments
for key, value in list(items(kw)):
if key in annotations:
kw[key] = cast(key, value)
|
[
"def",
"mutate",
"(",
"self",
",",
"context",
",",
"handler",
",",
"args",
",",
"kw",
")",
":",
"def",
"cast",
"(",
"arg",
",",
"val",
")",
":",
"if",
"arg",
"not",
"in",
"annotations",
":",
"return",
"cast",
"=",
"annotations",
"[",
"key",
"]",
"try",
":",
"val",
"=",
"cast",
"(",
"val",
")",
"except",
"(",
"ValueError",
",",
"TypeError",
")",
"as",
"e",
":",
"parts",
"=",
"list",
"(",
"e",
".",
"args",
")",
"parts",
"[",
"0",
"]",
"=",
"parts",
"[",
"0",
"]",
"+",
"\" processing argument '{}'\"",
".",
"format",
"(",
"arg",
")",
"e",
".",
"args",
"=",
"tuple",
"(",
"parts",
")",
"raise",
"return",
"val",
"annotations",
"=",
"getattr",
"(",
"handler",
".",
"__func__",
"if",
"hasattr",
"(",
"handler",
",",
"'__func__'",
")",
"else",
"handler",
",",
"'__annotations__'",
",",
"None",
")",
"if",
"not",
"annotations",
":",
"return",
"argspec",
"=",
"getfullargspec",
"(",
"handler",
")",
"arglist",
"=",
"list",
"(",
"argspec",
".",
"args",
")",
"if",
"ismethod",
"(",
"handler",
")",
":",
"del",
"arglist",
"[",
"0",
"]",
"for",
"i",
",",
"value",
"in",
"enumerate",
"(",
"list",
"(",
"args",
")",
")",
":",
"key",
"=",
"arglist",
"[",
"i",
"]",
"if",
"key",
"in",
"annotations",
":",
"args",
"[",
"i",
"]",
"=",
"cast",
"(",
"key",
",",
"value",
")",
"# Convert keyword arguments",
"for",
"key",
",",
"value",
"in",
"list",
"(",
"items",
"(",
"kw",
")",
")",
":",
"if",
"key",
"in",
"annotations",
":",
"kw",
"[",
"key",
"]",
"=",
"cast",
"(",
"key",
",",
"value",
")"
] |
Inspect and potentially mutate the given handler's arguments.
The args list and kw dictionary may be freely modified, though invalid arguments to the handler will fail.
|
[
"Inspect",
"and",
"potentially",
"mutate",
"the",
"given",
"handler",
"s",
"arguments",
".",
"The",
"args",
"list",
"and",
"kw",
"dictionary",
"may",
"be",
"freely",
"modified",
"though",
"invalid",
"arguments",
"to",
"the",
"handler",
"will",
"fail",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/annotation.py#L48-L87
|
train
|
marrow/WebCore
|
web/ext/annotation.py
|
AnnotationExtension.transform
|
def transform(self, context, handler, result):
"""Transform the value returned by the controller endpoint.
This extension transforms returned values if the endpoint has a return type annotation.
"""
handler = handler.__func__ if hasattr(handler, '__func__') else handler
annotation = getattr(handler, '__annotations__', {}).get('return', None)
if annotation:
return (annotation, result)
return result
|
python
|
def transform(self, context, handler, result):
"""Transform the value returned by the controller endpoint.
This extension transforms returned values if the endpoint has a return type annotation.
"""
handler = handler.__func__ if hasattr(handler, '__func__') else handler
annotation = getattr(handler, '__annotations__', {}).get('return', None)
if annotation:
return (annotation, result)
return result
|
[
"def",
"transform",
"(",
"self",
",",
"context",
",",
"handler",
",",
"result",
")",
":",
"handler",
"=",
"handler",
".",
"__func__",
"if",
"hasattr",
"(",
"handler",
",",
"'__func__'",
")",
"else",
"handler",
"annotation",
"=",
"getattr",
"(",
"handler",
",",
"'__annotations__'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'return'",
",",
"None",
")",
"if",
"annotation",
":",
"return",
"(",
"annotation",
",",
"result",
")",
"return",
"result"
] |
Transform the value returned by the controller endpoint.
This extension transforms returned values if the endpoint has a return type annotation.
|
[
"Transform",
"the",
"value",
"returned",
"by",
"the",
"controller",
"endpoint",
".",
"This",
"extension",
"transforms",
"returned",
"values",
"if",
"the",
"endpoint",
"has",
"a",
"return",
"type",
"annotation",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/annotation.py#L89-L100
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/tune/NeuroMLTuner.py
|
process_args
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(
description=("A script which can be run to tune a NeuroML 2 model against a number of target properties. Work in progress!"))
parser.add_argument('prefix',
type=str,
metavar='<prefix>',
help="Prefix for optimisation run")
parser.add_argument('neuromlFile',
type=str,
metavar='<neuromlFile>',
help="NeuroML2 file containing model")
parser.add_argument('target',
type=str,
metavar='<target>',
help="Target in NeuroML2 model")
parser.add_argument('parameters',
type=str,
metavar='<parameters>',
help="List of parameter to adjust")
parser.add_argument('maxConstraints',
type=str,
metavar='<max_constraints>',
help="Max values for parameters")
parser.add_argument('minConstraints',
type=str,
metavar='<min_constraints>',
help="Min values for parameters")
parser.add_argument('targetData',
type=str,
metavar='<targetData>',
help="List of name/value pairs for properties extracted from data to judge fitness against")
parser.add_argument('weights',
type=str,
metavar='<weights>',
help="Weights to assign to each target name/value pair")
parser.add_argument('-simTime',
type=float,
metavar='<simTime>',
default=DEFAULTS['simTime'],
help="Simulation duration")
parser.add_argument('-dt',
type=float,
metavar='<dt>',
default=DEFAULTS['dt'],
help="Simulation timestep")
parser.add_argument('-analysisStartTime',
type=float,
metavar='<analysisStartTime>',
default=DEFAULTS['analysisStartTime'],
help="Analysis start time")
parser.add_argument('-populationSize',
type=int,
metavar='<populationSize>',
default=DEFAULTS['populationSize'],
help="Population size")
parser.add_argument('-maxEvaluations',
type=int,
metavar='<maxEvaluations>',
default=DEFAULTS['maxEvaluations'],
help="Maximum evaluations")
parser.add_argument('-numSelected',
type=int,
metavar='<numSelected>',
default=DEFAULTS['numSelected'],
help="Number selected")
parser.add_argument('-numOffspring',
type=int,
metavar='<numOffspring>',
default=DEFAULTS['numOffspring'],
help="Number offspring")
parser.add_argument('-mutationRate',
type=float,
metavar='<mutationRate>',
default=DEFAULTS['mutationRate'],
help="Mutation rate")
parser.add_argument('-numElites',
type=int,
metavar='<numElites>',
default=DEFAULTS['numElites'],
help="Number of elites")
parser.add_argument('-numParallelEvaluations',
type=int,
metavar='<numParallelEvaluations>',
default=DEFAULTS['numParallelEvaluations'],
help="Number of evaluations to run in parallel")
parser.add_argument('-seed',
type=int,
metavar='<seed>',
default=DEFAULTS['seed'],
help="Seed for optimiser")
parser.add_argument('-simulator',
type=str,
metavar='<simulator>',
default=DEFAULTS['simulator'],
help="Simulator to run")
parser.add_argument('-knownTargetValues',
type=str,
metavar='<knownTargetValues>',
help="List of name/value pairs which represent the known values of the target parameters")
parser.add_argument('-nogui',
action='store_true',
default=DEFAULTS['nogui'],
help="Should GUI elements be supressed?")
parser.add_argument('-showPlotAlready',
action='store_true',
default=DEFAULTS['showPlotAlready'],
help="Should generated plots be suppressed until show() called?")
parser.add_argument('-verbose',
action='store_true',
default=DEFAULTS['verbose'],
help="Verbose mode")
parser.add_argument('-dryRun',
action='store_true',
default=DEFAULTS['dryRun'],
help="Dry run; just print setup information")
parser.add_argument('-extraReportInfo',
type=str,
metavar='<extraReportInfo>',
default=DEFAULTS['extraReportInfo'],
help='Extra tag/value pairs can be put into the report.json: -extraReportInfo=["tag":"value"]')
parser.add_argument('-cleanup',
action='store_true',
default=DEFAULTS['cleanup'],
help="Should (some) generated files, e.g. *.dat, be deleted as optimisation progresses?")
return parser.parse_args()
|
python
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(
description=("A script which can be run to tune a NeuroML 2 model against a number of target properties. Work in progress!"))
parser.add_argument('prefix',
type=str,
metavar='<prefix>',
help="Prefix for optimisation run")
parser.add_argument('neuromlFile',
type=str,
metavar='<neuromlFile>',
help="NeuroML2 file containing model")
parser.add_argument('target',
type=str,
metavar='<target>',
help="Target in NeuroML2 model")
parser.add_argument('parameters',
type=str,
metavar='<parameters>',
help="List of parameter to adjust")
parser.add_argument('maxConstraints',
type=str,
metavar='<max_constraints>',
help="Max values for parameters")
parser.add_argument('minConstraints',
type=str,
metavar='<min_constraints>',
help="Min values for parameters")
parser.add_argument('targetData',
type=str,
metavar='<targetData>',
help="List of name/value pairs for properties extracted from data to judge fitness against")
parser.add_argument('weights',
type=str,
metavar='<weights>',
help="Weights to assign to each target name/value pair")
parser.add_argument('-simTime',
type=float,
metavar='<simTime>',
default=DEFAULTS['simTime'],
help="Simulation duration")
parser.add_argument('-dt',
type=float,
metavar='<dt>',
default=DEFAULTS['dt'],
help="Simulation timestep")
parser.add_argument('-analysisStartTime',
type=float,
metavar='<analysisStartTime>',
default=DEFAULTS['analysisStartTime'],
help="Analysis start time")
parser.add_argument('-populationSize',
type=int,
metavar='<populationSize>',
default=DEFAULTS['populationSize'],
help="Population size")
parser.add_argument('-maxEvaluations',
type=int,
metavar='<maxEvaluations>',
default=DEFAULTS['maxEvaluations'],
help="Maximum evaluations")
parser.add_argument('-numSelected',
type=int,
metavar='<numSelected>',
default=DEFAULTS['numSelected'],
help="Number selected")
parser.add_argument('-numOffspring',
type=int,
metavar='<numOffspring>',
default=DEFAULTS['numOffspring'],
help="Number offspring")
parser.add_argument('-mutationRate',
type=float,
metavar='<mutationRate>',
default=DEFAULTS['mutationRate'],
help="Mutation rate")
parser.add_argument('-numElites',
type=int,
metavar='<numElites>',
default=DEFAULTS['numElites'],
help="Number of elites")
parser.add_argument('-numParallelEvaluations',
type=int,
metavar='<numParallelEvaluations>',
default=DEFAULTS['numParallelEvaluations'],
help="Number of evaluations to run in parallel")
parser.add_argument('-seed',
type=int,
metavar='<seed>',
default=DEFAULTS['seed'],
help="Seed for optimiser")
parser.add_argument('-simulator',
type=str,
metavar='<simulator>',
default=DEFAULTS['simulator'],
help="Simulator to run")
parser.add_argument('-knownTargetValues',
type=str,
metavar='<knownTargetValues>',
help="List of name/value pairs which represent the known values of the target parameters")
parser.add_argument('-nogui',
action='store_true',
default=DEFAULTS['nogui'],
help="Should GUI elements be supressed?")
parser.add_argument('-showPlotAlready',
action='store_true',
default=DEFAULTS['showPlotAlready'],
help="Should generated plots be suppressed until show() called?")
parser.add_argument('-verbose',
action='store_true',
default=DEFAULTS['verbose'],
help="Verbose mode")
parser.add_argument('-dryRun',
action='store_true',
default=DEFAULTS['dryRun'],
help="Dry run; just print setup information")
parser.add_argument('-extraReportInfo',
type=str,
metavar='<extraReportInfo>',
default=DEFAULTS['extraReportInfo'],
help='Extra tag/value pairs can be put into the report.json: -extraReportInfo=["tag":"value"]')
parser.add_argument('-cleanup',
action='store_true',
default=DEFAULTS['cleanup'],
help="Should (some) generated files, e.g. *.dat, be deleted as optimisation progresses?")
return parser.parse_args()
|
[
"def",
"process_args",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"(",
"\"A script which can be run to tune a NeuroML 2 model against a number of target properties. Work in progress!\"",
")",
")",
"parser",
".",
"add_argument",
"(",
"'prefix'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<prefix>'",
",",
"help",
"=",
"\"Prefix for optimisation run\"",
")",
"parser",
".",
"add_argument",
"(",
"'neuromlFile'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<neuromlFile>'",
",",
"help",
"=",
"\"NeuroML2 file containing model\"",
")",
"parser",
".",
"add_argument",
"(",
"'target'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<target>'",
",",
"help",
"=",
"\"Target in NeuroML2 model\"",
")",
"parser",
".",
"add_argument",
"(",
"'parameters'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<parameters>'",
",",
"help",
"=",
"\"List of parameter to adjust\"",
")",
"parser",
".",
"add_argument",
"(",
"'maxConstraints'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<max_constraints>'",
",",
"help",
"=",
"\"Max values for parameters\"",
")",
"parser",
".",
"add_argument",
"(",
"'minConstraints'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<min_constraints>'",
",",
"help",
"=",
"\"Min values for parameters\"",
")",
"parser",
".",
"add_argument",
"(",
"'targetData'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<targetData>'",
",",
"help",
"=",
"\"List of name/value pairs for properties extracted from data to judge fitness against\"",
")",
"parser",
".",
"add_argument",
"(",
"'weights'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<weights>'",
",",
"help",
"=",
"\"Weights to assign to each target name/value pair\"",
")",
"parser",
".",
"add_argument",
"(",
"'-simTime'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<simTime>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'simTime'",
"]",
",",
"help",
"=",
"\"Simulation duration\"",
")",
"parser",
".",
"add_argument",
"(",
"'-dt'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<dt>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'dt'",
"]",
",",
"help",
"=",
"\"Simulation timestep\"",
")",
"parser",
".",
"add_argument",
"(",
"'-analysisStartTime'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<analysisStartTime>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'analysisStartTime'",
"]",
",",
"help",
"=",
"\"Analysis start time\"",
")",
"parser",
".",
"add_argument",
"(",
"'-populationSize'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<populationSize>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'populationSize'",
"]",
",",
"help",
"=",
"\"Population size\"",
")",
"parser",
".",
"add_argument",
"(",
"'-maxEvaluations'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<maxEvaluations>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'maxEvaluations'",
"]",
",",
"help",
"=",
"\"Maximum evaluations\"",
")",
"parser",
".",
"add_argument",
"(",
"'-numSelected'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<numSelected>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'numSelected'",
"]",
",",
"help",
"=",
"\"Number selected\"",
")",
"parser",
".",
"add_argument",
"(",
"'-numOffspring'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<numOffspring>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'numOffspring'",
"]",
",",
"help",
"=",
"\"Number offspring\"",
")",
"parser",
".",
"add_argument",
"(",
"'-mutationRate'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<mutationRate>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'mutationRate'",
"]",
",",
"help",
"=",
"\"Mutation rate\"",
")",
"parser",
".",
"add_argument",
"(",
"'-numElites'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<numElites>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'numElites'",
"]",
",",
"help",
"=",
"\"Number of elites\"",
")",
"parser",
".",
"add_argument",
"(",
"'-numParallelEvaluations'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<numParallelEvaluations>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'numParallelEvaluations'",
"]",
",",
"help",
"=",
"\"Number of evaluations to run in parallel\"",
")",
"parser",
".",
"add_argument",
"(",
"'-seed'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<seed>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'seed'",
"]",
",",
"help",
"=",
"\"Seed for optimiser\"",
")",
"parser",
".",
"add_argument",
"(",
"'-simulator'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<simulator>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'simulator'",
"]",
",",
"help",
"=",
"\"Simulator to run\"",
")",
"parser",
".",
"add_argument",
"(",
"'-knownTargetValues'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<knownTargetValues>'",
",",
"help",
"=",
"\"List of name/value pairs which represent the known values of the target parameters\"",
")",
"parser",
".",
"add_argument",
"(",
"'-nogui'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'nogui'",
"]",
",",
"help",
"=",
"\"Should GUI elements be supressed?\"",
")",
"parser",
".",
"add_argument",
"(",
"'-showPlotAlready'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'showPlotAlready'",
"]",
",",
"help",
"=",
"\"Should generated plots be suppressed until show() called?\"",
")",
"parser",
".",
"add_argument",
"(",
"'-verbose'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'verbose'",
"]",
",",
"help",
"=",
"\"Verbose mode\"",
")",
"parser",
".",
"add_argument",
"(",
"'-dryRun'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'dryRun'",
"]",
",",
"help",
"=",
"\"Dry run; just print setup information\"",
")",
"parser",
".",
"add_argument",
"(",
"'-extraReportInfo'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<extraReportInfo>'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'extraReportInfo'",
"]",
",",
"help",
"=",
"'Extra tag/value pairs can be put into the report.json: -extraReportInfo=[\"tag\":\"value\"]'",
")",
"parser",
".",
"add_argument",
"(",
"'-cleanup'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'cleanup'",
"]",
",",
"help",
"=",
"\"Should (some) generated files, e.g. *.dat, be deleted as optimisation progresses?\"",
")",
"return",
"parser",
".",
"parse_args",
"(",
")"
] |
Parse command-line arguments.
|
[
"Parse",
"command",
"-",
"line",
"arguments",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/tune/NeuroMLTuner.py#L52-L209
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/povray/MakeMovie.py
|
process_args
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(description="A file for overlaying POVRay files generated from NeuroML by NeuroML1ToPOVRay.py with cell activity (e.g. as generated from a neuroConstruct simulation)")
parser.add_argument('prefix',
type=str,
metavar='<network prefix>',
help='Prefix for files in PovRay, e.g. use PREFIX is files are PREFIX.pov, PREFIX_net.inc, etc.')
parser.add_argument('-activity',
action='store_true',
default=False,
help="If this is specified, overlay network activity (not tested!!)")
parser.add_argument('-maxV',
type=float,
metavar='<maxV>',
default=50.0,
help='Max voltage for colour scale in mV')
parser.add_argument('-minV',
type=float,
metavar='<minV>',
default=-90.0,
help='Min voltage for colour scale in mV')
parser.add_argument('-startTime',
type=float,
metavar='<startTime>',
default=0,
help='Time in ms at which to start overlaying the simulation activity')
parser.add_argument('-endTime',
type=float,
metavar='<endTime>',
default=100,
help='End time of simulation activity in ms')
parser.add_argument('-title',
type=str,
metavar='<title>',
default='Movie generated from neuroConstruct simulation',
help='Title for movie')
parser.add_argument('-left',
type=str,
metavar='<left info>',
default='',
help='Text on left')
parser.add_argument('-frames',
type=int,
metavar='<frames>',
default=100,
help='Number of frames')
parser.add_argument('-name',
type=str,
metavar='<Movie name>',
default='output',
help='Movie name')
return parser.parse_args()
|
python
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(description="A file for overlaying POVRay files generated from NeuroML by NeuroML1ToPOVRay.py with cell activity (e.g. as generated from a neuroConstruct simulation)")
parser.add_argument('prefix',
type=str,
metavar='<network prefix>',
help='Prefix for files in PovRay, e.g. use PREFIX is files are PREFIX.pov, PREFIX_net.inc, etc.')
parser.add_argument('-activity',
action='store_true',
default=False,
help="If this is specified, overlay network activity (not tested!!)")
parser.add_argument('-maxV',
type=float,
metavar='<maxV>',
default=50.0,
help='Max voltage for colour scale in mV')
parser.add_argument('-minV',
type=float,
metavar='<minV>',
default=-90.0,
help='Min voltage for colour scale in mV')
parser.add_argument('-startTime',
type=float,
metavar='<startTime>',
default=0,
help='Time in ms at which to start overlaying the simulation activity')
parser.add_argument('-endTime',
type=float,
metavar='<endTime>',
default=100,
help='End time of simulation activity in ms')
parser.add_argument('-title',
type=str,
metavar='<title>',
default='Movie generated from neuroConstruct simulation',
help='Title for movie')
parser.add_argument('-left',
type=str,
metavar='<left info>',
default='',
help='Text on left')
parser.add_argument('-frames',
type=int,
metavar='<frames>',
default=100,
help='Number of frames')
parser.add_argument('-name',
type=str,
metavar='<Movie name>',
default='output',
help='Movie name')
return parser.parse_args()
|
[
"def",
"process_args",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"\"A file for overlaying POVRay files generated from NeuroML by NeuroML1ToPOVRay.py with cell activity (e.g. as generated from a neuroConstruct simulation)\"",
")",
"parser",
".",
"add_argument",
"(",
"'prefix'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<network prefix>'",
",",
"help",
"=",
"'Prefix for files in PovRay, e.g. use PREFIX is files are PREFIX.pov, PREFIX_net.inc, etc.'",
")",
"parser",
".",
"add_argument",
"(",
"'-activity'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, overlay network activity (not tested!!)\"",
")",
"parser",
".",
"add_argument",
"(",
"'-maxV'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<maxV>'",
",",
"default",
"=",
"50.0",
",",
"help",
"=",
"'Max voltage for colour scale in mV'",
")",
"parser",
".",
"add_argument",
"(",
"'-minV'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<minV>'",
",",
"default",
"=",
"-",
"90.0",
",",
"help",
"=",
"'Min voltage for colour scale in mV'",
")",
"parser",
".",
"add_argument",
"(",
"'-startTime'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<startTime>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Time in ms at which to start overlaying the simulation activity'",
")",
"parser",
".",
"add_argument",
"(",
"'-endTime'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<endTime>'",
",",
"default",
"=",
"100",
",",
"help",
"=",
"'End time of simulation activity in ms'",
")",
"parser",
".",
"add_argument",
"(",
"'-title'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<title>'",
",",
"default",
"=",
"'Movie generated from neuroConstruct simulation'",
",",
"help",
"=",
"'Title for movie'",
")",
"parser",
".",
"add_argument",
"(",
"'-left'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<left info>'",
",",
"default",
"=",
"''",
",",
"help",
"=",
"'Text on left'",
")",
"parser",
".",
"add_argument",
"(",
"'-frames'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<frames>'",
",",
"default",
"=",
"100",
",",
"help",
"=",
"'Number of frames'",
")",
"parser",
".",
"add_argument",
"(",
"'-name'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<Movie name>'",
",",
"default",
"=",
"'output'",
",",
"help",
"=",
"'Movie name'",
")",
"return",
"parser",
".",
"parse_args",
"(",
")"
] |
Parse command-line arguments.
|
[
"Parse",
"command",
"-",
"line",
"arguments",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/povray/MakeMovie.py#L27-L92
|
train
|
marrow/WebCore
|
web/server/tornado_.py
|
serve
|
def serve(application, host='127.0.0.1', port=8080, **options):
"""Tornado's HTTPServer.
This is a high quality asynchronous server with many options. For details, please visit:
http://www.tornadoweb.org/en/stable/httpserver.html#http-server
"""
# Wrap our our WSGI application (potentially stack) in a Tornado adapter.
container = tornado.wsgi.WSGIContainer(application)
# Spin up a Tornado HTTP server using this container.
http_server = tornado.httpserver.HTTPServer(container, **options)
http_server.listen(int(port), host)
# Start and block on the Tornado IO loop.
tornado.ioloop.IOLoop.instance().start()
|
python
|
def serve(application, host='127.0.0.1', port=8080, **options):
"""Tornado's HTTPServer.
This is a high quality asynchronous server with many options. For details, please visit:
http://www.tornadoweb.org/en/stable/httpserver.html#http-server
"""
# Wrap our our WSGI application (potentially stack) in a Tornado adapter.
container = tornado.wsgi.WSGIContainer(application)
# Spin up a Tornado HTTP server using this container.
http_server = tornado.httpserver.HTTPServer(container, **options)
http_server.listen(int(port), host)
# Start and block on the Tornado IO loop.
tornado.ioloop.IOLoop.instance().start()
|
[
"def",
"serve",
"(",
"application",
",",
"host",
"=",
"'127.0.0.1'",
",",
"port",
"=",
"8080",
",",
"*",
"*",
"options",
")",
":",
"# Wrap our our WSGI application (potentially stack) in a Tornado adapter.",
"container",
"=",
"tornado",
".",
"wsgi",
".",
"WSGIContainer",
"(",
"application",
")",
"# Spin up a Tornado HTTP server using this container.",
"http_server",
"=",
"tornado",
".",
"httpserver",
".",
"HTTPServer",
"(",
"container",
",",
"*",
"*",
"options",
")",
"http_server",
".",
"listen",
"(",
"int",
"(",
"port",
")",
",",
"host",
")",
"# Start and block on the Tornado IO loop.",
"tornado",
".",
"ioloop",
".",
"IOLoop",
".",
"instance",
"(",
")",
".",
"start",
"(",
")"
] |
Tornado's HTTPServer.
This is a high quality asynchronous server with many options. For details, please visit:
http://www.tornadoweb.org/en/stable/httpserver.html#http-server
|
[
"Tornado",
"s",
"HTTPServer",
".",
"This",
"is",
"a",
"high",
"quality",
"asynchronous",
"server",
"with",
"many",
"options",
".",
"For",
"details",
"please",
"visit",
":",
"http",
":",
"//",
"www",
".",
"tornadoweb",
".",
"org",
"/",
"en",
"/",
"stable",
"/",
"httpserver",
".",
"html#http",
"-",
"server"
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/server/tornado_.py#L18-L34
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/pynml.py
|
parse_arguments
|
def parse_arguments():
"""Parse command line arguments"""
import argparse
parser = argparse.ArgumentParser(
description=('pyNeuroML v%s: Python utilities for NeuroML2' % __version__
+ "\n libNeuroML v%s"%(neuroml.__version__)
+ "\n jNeuroML v%s"%JNEUROML_VERSION),
usage=('pynml [-h|--help] [<shared options>] '
'<one of the mutually-exclusive options>'),
formatter_class=argparse.RawTextHelpFormatter
)
shared_options = parser.add_argument_group(
title='Shared options',
description=('These options can be added to any of the '
'mutually-exclusive options')
)
shared_options.add_argument(
'-verbose',
action='store_true',
default=DEFAULTS['v'],
help='Verbose output'
)
shared_options.add_argument(
'-java_max_memory',
metavar='MAX',
default=DEFAULTS['default_java_max_memory'],
help=('Java memory for jNeuroML, e.g. 400M, 2G (used in\n'
'-Xmx argument to java)')
)
shared_options.add_argument(
'-nogui',
action='store_true',
default=DEFAULTS['nogui'],
help=('Suppress GUI,\n'
'i.e. show no plots, just save results')
)
shared_options.add_argument(
'lems_file',
type=str,
metavar='<LEMS/NeuroML 2 file>',
help='LEMS/NeuroML 2 file to process'
)
mut_exc_opts_grp = parser.add_argument_group(
title='Mutually-exclusive options',
description='Only one of these options can be selected'
)
mut_exc_opts = mut_exc_opts_grp.add_mutually_exclusive_group(required=False)
mut_exc_opts.add_argument(
'-sedml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert\n'
'simulation settings (duration, dt, what to save)\n'
'to SED-ML format')
)
mut_exc_opts.add_argument(
'-neuron',
nargs=argparse.REMAINDER,
help=('(Via jNeuroML) Load a LEMS file, and convert it to\n'
'NEURON format.\n'
'The full format of the \'-neuron\' option is:\n'
'-neuron [-nogui] [-run] [-outputdir dir] <LEMS file>\n'
' -nogui\n'
' do not generate gtaphical elements in NEURON,\n'
' just run, save data, and quit\n'
' -run\n'
' compile NMODL files and run the main NEURON\n'
' hoc file (Linux only currently)\n'
' -outputdir <dir>\n'
' generate NEURON files in directory <dir>\n'
' <LEMS file>\n'
' the LEMS file to use')
)
mut_exc_opts.add_argument(
'-svg',
action='store_true',
help=('(Via jNeuroML) Convert NeuroML2 file (network & cells)\n'
'to SVG format view of 3D structure')
)
mut_exc_opts.add_argument(
'-png',
action='store_true',
help=('(Via jNeuroML) Convert NeuroML2 file (network & cells)\n'
'to PNG format view of 3D structure')
)
mut_exc_opts.add_argument(
'-dlems',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to dLEMS format, a distilled form of LEMS in JSON')
)
mut_exc_opts.add_argument(
'-vertex',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to VERTEX format')
)
mut_exc_opts.add_argument(
'-xpp',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to XPPAUT format')
)
mut_exc_opts.add_argument(
'-dnsim',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to DNsim format')
)
mut_exc_opts.add_argument(
'-brian',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to Brian format')
)
mut_exc_opts.add_argument(
'-sbml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to SBML format')
)
mut_exc_opts.add_argument(
'-matlab',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to MATLAB format')
)
mut_exc_opts.add_argument(
'-cvode',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to C format using CVODE package')
)
mut_exc_opts.add_argument(
'-nineml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to NineML format')
)
mut_exc_opts.add_argument(
'-spineml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to SpineML format')
)
mut_exc_opts.add_argument(
'-sbml-import',
metavar=('<SBML file>', 'duration', 'dt'),
nargs=3,
help=('(Via jNeuroML) Load a SBML file, and convert it\n'
'toLEMS format using values for duration & dt\n'
'in ms (ignoring SBML units)')
)
mut_exc_opts.add_argument(
'-sbml-import-units',
metavar=('<SBML file>', 'duration', 'dt'),
nargs=3,
help=('(Via jNeuroML) Load a SBML file, and convert it\n'
'to LEMS format using values for duration & dt\n'
'in ms (attempt to extract SBML units; ensure units\n'
'are valid in the SBML!)')
)
mut_exc_opts.add_argument(
'-vhdl',
metavar=('neuronid', '<LEMS file>'),
nargs=2,
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to VHDL format')
)
mut_exc_opts.add_argument(
'-graph',
metavar=('level'),
nargs=1,
help=('Load a NeuroML file, and convert it to a graph using\n'
'GraphViz. Detail is set by level (1, 2, etc.)')
)
mut_exc_opts.add_argument(
'-matrix',
metavar=('level'),
nargs=1,
help=('Load a NeuroML file, and convert it to a matrix displaying\n'
'connectivity. Detail is set by level (1, 2, etc.)')
)
mut_exc_opts.add_argument(
'-validate',
action='store_true',
help=('(Via jNeuroML) Validate NeuroML2 file(s) against the\n'
'latest Schema')
)
mut_exc_opts.add_argument(
'-validatev1',
action='store_true',
help=('(Via jNeuroML) Validate NeuroML file(s) against the\n'
'v1.8.1 Schema')
)
return parser.parse_args()
|
python
|
def parse_arguments():
"""Parse command line arguments"""
import argparse
parser = argparse.ArgumentParser(
description=('pyNeuroML v%s: Python utilities for NeuroML2' % __version__
+ "\n libNeuroML v%s"%(neuroml.__version__)
+ "\n jNeuroML v%s"%JNEUROML_VERSION),
usage=('pynml [-h|--help] [<shared options>] '
'<one of the mutually-exclusive options>'),
formatter_class=argparse.RawTextHelpFormatter
)
shared_options = parser.add_argument_group(
title='Shared options',
description=('These options can be added to any of the '
'mutually-exclusive options')
)
shared_options.add_argument(
'-verbose',
action='store_true',
default=DEFAULTS['v'],
help='Verbose output'
)
shared_options.add_argument(
'-java_max_memory',
metavar='MAX',
default=DEFAULTS['default_java_max_memory'],
help=('Java memory for jNeuroML, e.g. 400M, 2G (used in\n'
'-Xmx argument to java)')
)
shared_options.add_argument(
'-nogui',
action='store_true',
default=DEFAULTS['nogui'],
help=('Suppress GUI,\n'
'i.e. show no plots, just save results')
)
shared_options.add_argument(
'lems_file',
type=str,
metavar='<LEMS/NeuroML 2 file>',
help='LEMS/NeuroML 2 file to process'
)
mut_exc_opts_grp = parser.add_argument_group(
title='Mutually-exclusive options',
description='Only one of these options can be selected'
)
mut_exc_opts = mut_exc_opts_grp.add_mutually_exclusive_group(required=False)
mut_exc_opts.add_argument(
'-sedml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert\n'
'simulation settings (duration, dt, what to save)\n'
'to SED-ML format')
)
mut_exc_opts.add_argument(
'-neuron',
nargs=argparse.REMAINDER,
help=('(Via jNeuroML) Load a LEMS file, and convert it to\n'
'NEURON format.\n'
'The full format of the \'-neuron\' option is:\n'
'-neuron [-nogui] [-run] [-outputdir dir] <LEMS file>\n'
' -nogui\n'
' do not generate gtaphical elements in NEURON,\n'
' just run, save data, and quit\n'
' -run\n'
' compile NMODL files and run the main NEURON\n'
' hoc file (Linux only currently)\n'
' -outputdir <dir>\n'
' generate NEURON files in directory <dir>\n'
' <LEMS file>\n'
' the LEMS file to use')
)
mut_exc_opts.add_argument(
'-svg',
action='store_true',
help=('(Via jNeuroML) Convert NeuroML2 file (network & cells)\n'
'to SVG format view of 3D structure')
)
mut_exc_opts.add_argument(
'-png',
action='store_true',
help=('(Via jNeuroML) Convert NeuroML2 file (network & cells)\n'
'to PNG format view of 3D structure')
)
mut_exc_opts.add_argument(
'-dlems',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to dLEMS format, a distilled form of LEMS in JSON')
)
mut_exc_opts.add_argument(
'-vertex',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to VERTEX format')
)
mut_exc_opts.add_argument(
'-xpp',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to XPPAUT format')
)
mut_exc_opts.add_argument(
'-dnsim',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to DNsim format')
)
mut_exc_opts.add_argument(
'-brian',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to Brian format')
)
mut_exc_opts.add_argument(
'-sbml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to SBML format')
)
mut_exc_opts.add_argument(
'-matlab',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to MATLAB format')
)
mut_exc_opts.add_argument(
'-cvode',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to C format using CVODE package')
)
mut_exc_opts.add_argument(
'-nineml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to NineML format')
)
mut_exc_opts.add_argument(
'-spineml',
action='store_true',
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to SpineML format')
)
mut_exc_opts.add_argument(
'-sbml-import',
metavar=('<SBML file>', 'duration', 'dt'),
nargs=3,
help=('(Via jNeuroML) Load a SBML file, and convert it\n'
'toLEMS format using values for duration & dt\n'
'in ms (ignoring SBML units)')
)
mut_exc_opts.add_argument(
'-sbml-import-units',
metavar=('<SBML file>', 'duration', 'dt'),
nargs=3,
help=('(Via jNeuroML) Load a SBML file, and convert it\n'
'to LEMS format using values for duration & dt\n'
'in ms (attempt to extract SBML units; ensure units\n'
'are valid in the SBML!)')
)
mut_exc_opts.add_argument(
'-vhdl',
metavar=('neuronid', '<LEMS file>'),
nargs=2,
help=('(Via jNeuroML) Load a LEMS file, and convert it\n'
'to VHDL format')
)
mut_exc_opts.add_argument(
'-graph',
metavar=('level'),
nargs=1,
help=('Load a NeuroML file, and convert it to a graph using\n'
'GraphViz. Detail is set by level (1, 2, etc.)')
)
mut_exc_opts.add_argument(
'-matrix',
metavar=('level'),
nargs=1,
help=('Load a NeuroML file, and convert it to a matrix displaying\n'
'connectivity. Detail is set by level (1, 2, etc.)')
)
mut_exc_opts.add_argument(
'-validate',
action='store_true',
help=('(Via jNeuroML) Validate NeuroML2 file(s) against the\n'
'latest Schema')
)
mut_exc_opts.add_argument(
'-validatev1',
action='store_true',
help=('(Via jNeuroML) Validate NeuroML file(s) against the\n'
'v1.8.1 Schema')
)
return parser.parse_args()
|
[
"def",
"parse_arguments",
"(",
")",
":",
"import",
"argparse",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"(",
"'pyNeuroML v%s: Python utilities for NeuroML2'",
"%",
"__version__",
"+",
"\"\\n libNeuroML v%s\"",
"%",
"(",
"neuroml",
".",
"__version__",
")",
"+",
"\"\\n jNeuroML v%s\"",
"%",
"JNEUROML_VERSION",
")",
",",
"usage",
"=",
"(",
"'pynml [-h|--help] [<shared options>] '",
"'<one of the mutually-exclusive options>'",
")",
",",
"formatter_class",
"=",
"argparse",
".",
"RawTextHelpFormatter",
")",
"shared_options",
"=",
"parser",
".",
"add_argument_group",
"(",
"title",
"=",
"'Shared options'",
",",
"description",
"=",
"(",
"'These options can be added to any of the '",
"'mutually-exclusive options'",
")",
")",
"shared_options",
".",
"add_argument",
"(",
"'-verbose'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'v'",
"]",
",",
"help",
"=",
"'Verbose output'",
")",
"shared_options",
".",
"add_argument",
"(",
"'-java_max_memory'",
",",
"metavar",
"=",
"'MAX'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'default_java_max_memory'",
"]",
",",
"help",
"=",
"(",
"'Java memory for jNeuroML, e.g. 400M, 2G (used in\\n'",
"'-Xmx argument to java)'",
")",
")",
"shared_options",
".",
"add_argument",
"(",
"'-nogui'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"DEFAULTS",
"[",
"'nogui'",
"]",
",",
"help",
"=",
"(",
"'Suppress GUI,\\n'",
"'i.e. show no plots, just save results'",
")",
")",
"shared_options",
".",
"add_argument",
"(",
"'lems_file'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<LEMS/NeuroML 2 file>'",
",",
"help",
"=",
"'LEMS/NeuroML 2 file to process'",
")",
"mut_exc_opts_grp",
"=",
"parser",
".",
"add_argument_group",
"(",
"title",
"=",
"'Mutually-exclusive options'",
",",
"description",
"=",
"'Only one of these options can be selected'",
")",
"mut_exc_opts",
"=",
"mut_exc_opts_grp",
".",
"add_mutually_exclusive_group",
"(",
"required",
"=",
"False",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-sedml'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert\\n'",
"'simulation settings (duration, dt, what to save)\\n'",
"'to SED-ML format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-neuron'",
",",
"nargs",
"=",
"argparse",
".",
"REMAINDER",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it to\\n'",
"'NEURON format.\\n'",
"'The full format of the \\'-neuron\\' option is:\\n'",
"'-neuron [-nogui] [-run] [-outputdir dir] <LEMS file>\\n'",
"' -nogui\\n'",
"' do not generate gtaphical elements in NEURON,\\n'",
"' just run, save data, and quit\\n'",
"' -run\\n'",
"' compile NMODL files and run the main NEURON\\n'",
"' hoc file (Linux only currently)\\n'",
"' -outputdir <dir>\\n'",
"' generate NEURON files in directory <dir>\\n'",
"' <LEMS file>\\n'",
"' the LEMS file to use'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-svg'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Convert NeuroML2 file (network & cells)\\n'",
"'to SVG format view of 3D structure'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-png'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Convert NeuroML2 file (network & cells)\\n'",
"'to PNG format view of 3D structure'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-dlems'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to dLEMS format, a distilled form of LEMS in JSON'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-vertex'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to VERTEX format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-xpp'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to XPPAUT format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-dnsim'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to DNsim format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-brian'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to Brian format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-sbml'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to SBML format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-matlab'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to MATLAB format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-cvode'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to C format using CVODE package'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-nineml'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to NineML format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-spineml'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to SpineML format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-sbml-import'",
",",
"metavar",
"=",
"(",
"'<SBML file>'",
",",
"'duration'",
",",
"'dt'",
")",
",",
"nargs",
"=",
"3",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a SBML file, and convert it\\n'",
"'toLEMS format using values for duration & dt\\n'",
"'in ms (ignoring SBML units)'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-sbml-import-units'",
",",
"metavar",
"=",
"(",
"'<SBML file>'",
",",
"'duration'",
",",
"'dt'",
")",
",",
"nargs",
"=",
"3",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a SBML file, and convert it\\n'",
"'to LEMS format using values for duration & dt\\n'",
"'in ms (attempt to extract SBML units; ensure units\\n'",
"'are valid in the SBML!)'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-vhdl'",
",",
"metavar",
"=",
"(",
"'neuronid'",
",",
"'<LEMS file>'",
")",
",",
"nargs",
"=",
"2",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Load a LEMS file, and convert it\\n'",
"'to VHDL format'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-graph'",
",",
"metavar",
"=",
"(",
"'level'",
")",
",",
"nargs",
"=",
"1",
",",
"help",
"=",
"(",
"'Load a NeuroML file, and convert it to a graph using\\n'",
"'GraphViz. Detail is set by level (1, 2, etc.)'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-matrix'",
",",
"metavar",
"=",
"(",
"'level'",
")",
",",
"nargs",
"=",
"1",
",",
"help",
"=",
"(",
"'Load a NeuroML file, and convert it to a matrix displaying\\n'",
"'connectivity. Detail is set by level (1, 2, etc.)'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-validate'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Validate NeuroML2 file(s) against the\\n'",
"'latest Schema'",
")",
")",
"mut_exc_opts",
".",
"add_argument",
"(",
"'-validatev1'",
",",
"action",
"=",
"'store_true'",
",",
"help",
"=",
"(",
"'(Via jNeuroML) Validate NeuroML file(s) against the\\n'",
"'v1.8.1 Schema'",
")",
")",
"return",
"parser",
".",
"parse_args",
"(",
")"
] |
Parse command line arguments
|
[
"Parse",
"command",
"line",
"arguments"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/pynml.py#L43-L245
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/pynml.py
|
quick_summary
|
def quick_summary(nml2_doc):
'''
Or better just use nml2_doc.summary(show_includes=False)
'''
info = 'Contents of NeuroML 2 document: %s\n'%nml2_doc.id
membs = inspect.getmembers(nml2_doc)
for memb in membs:
if isinstance(memb[1], list) and len(memb[1])>0 \
and not memb[0].endswith('_'):
info+=' %s:\n ['%memb[0]
for entry in memb[1]:
extra = '???'
extra = entry.name if hasattr(entry,'name') else extra
extra = entry.href if hasattr(entry,'href') else extra
extra = entry.id if hasattr(entry,'id') else extra
info+=" %s (%s),"%(entry, extra)
info+=']\n'
return info
|
python
|
def quick_summary(nml2_doc):
'''
Or better just use nml2_doc.summary(show_includes=False)
'''
info = 'Contents of NeuroML 2 document: %s\n'%nml2_doc.id
membs = inspect.getmembers(nml2_doc)
for memb in membs:
if isinstance(memb[1], list) and len(memb[1])>0 \
and not memb[0].endswith('_'):
info+=' %s:\n ['%memb[0]
for entry in memb[1]:
extra = '???'
extra = entry.name if hasattr(entry,'name') else extra
extra = entry.href if hasattr(entry,'href') else extra
extra = entry.id if hasattr(entry,'id') else extra
info+=" %s (%s),"%(entry, extra)
info+=']\n'
return info
|
[
"def",
"quick_summary",
"(",
"nml2_doc",
")",
":",
"info",
"=",
"'Contents of NeuroML 2 document: %s\\n'",
"%",
"nml2_doc",
".",
"id",
"membs",
"=",
"inspect",
".",
"getmembers",
"(",
"nml2_doc",
")",
"for",
"memb",
"in",
"membs",
":",
"if",
"isinstance",
"(",
"memb",
"[",
"1",
"]",
",",
"list",
")",
"and",
"len",
"(",
"memb",
"[",
"1",
"]",
")",
">",
"0",
"and",
"not",
"memb",
"[",
"0",
"]",
".",
"endswith",
"(",
"'_'",
")",
":",
"info",
"+=",
"' %s:\\n ['",
"%",
"memb",
"[",
"0",
"]",
"for",
"entry",
"in",
"memb",
"[",
"1",
"]",
":",
"extra",
"=",
"'???'",
"extra",
"=",
"entry",
".",
"name",
"if",
"hasattr",
"(",
"entry",
",",
"'name'",
")",
"else",
"extra",
"extra",
"=",
"entry",
".",
"href",
"if",
"hasattr",
"(",
"entry",
",",
"'href'",
")",
"else",
"extra",
"extra",
"=",
"entry",
".",
"id",
"if",
"hasattr",
"(",
"entry",
",",
"'id'",
")",
"else",
"extra",
"info",
"+=",
"\" %s (%s),\"",
"%",
"(",
"entry",
",",
"extra",
")",
"info",
"+=",
"']\\n'",
"return",
"info"
] |
Or better just use nml2_doc.summary(show_includes=False)
|
[
"Or",
"better",
"just",
"use",
"nml2_doc",
".",
"summary",
"(",
"show_includes",
"=",
"False",
")"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/pynml.py#L419-L441
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/pynml.py
|
execute_command_in_dir
|
def execute_command_in_dir(command, directory, verbose=DEFAULTS['v'],
prefix="Output: ", env=None):
"""Execute a command in specific working directory"""
if os.name == 'nt':
directory = os.path.normpath(directory)
print_comment("Executing: (%s) in directory: %s" % (command, directory),
verbose)
if env is not None:
print_comment("Extra env variables %s" % (env), verbose)
try:
if os.name == 'nt':
return_string = subprocess.check_output(command,
cwd=directory,
shell=True,
env=env,
close_fds=False)
else:
return_string = subprocess.check_output(command,
cwd=directory,
shell=True,
stderr=subprocess.STDOUT,
env=env,
close_fds=True)
return_string = return_string.decode("utf-8") # For Python 3
print_comment('Command completed. Output: \n %s%s' % \
(prefix,return_string.replace('\n','\n '+prefix)),
verbose)
return return_string
except AttributeError:
# For python 2.6...
print_comment_v('Assuming Python 2.6...')
return_string = subprocess.Popen(command,
cwd=directory,
shell=True,
stdout=subprocess.PIPE).communicate()[0]
return return_string
except subprocess.CalledProcessError as e:
print_comment_v('*** Problem running command: \n %s'%e)
print_comment_v('%s%s'%(prefix,e.output.decode().replace('\n','\n'+prefix)))
return None
except:
print_comment_v('*** Unknown problem running command: %s'%e)
return None
print_comment("Finished execution", verbose)
|
python
|
def execute_command_in_dir(command, directory, verbose=DEFAULTS['v'],
prefix="Output: ", env=None):
"""Execute a command in specific working directory"""
if os.name == 'nt':
directory = os.path.normpath(directory)
print_comment("Executing: (%s) in directory: %s" % (command, directory),
verbose)
if env is not None:
print_comment("Extra env variables %s" % (env), verbose)
try:
if os.name == 'nt':
return_string = subprocess.check_output(command,
cwd=directory,
shell=True,
env=env,
close_fds=False)
else:
return_string = subprocess.check_output(command,
cwd=directory,
shell=True,
stderr=subprocess.STDOUT,
env=env,
close_fds=True)
return_string = return_string.decode("utf-8") # For Python 3
print_comment('Command completed. Output: \n %s%s' % \
(prefix,return_string.replace('\n','\n '+prefix)),
verbose)
return return_string
except AttributeError:
# For python 2.6...
print_comment_v('Assuming Python 2.6...')
return_string = subprocess.Popen(command,
cwd=directory,
shell=True,
stdout=subprocess.PIPE).communicate()[0]
return return_string
except subprocess.CalledProcessError as e:
print_comment_v('*** Problem running command: \n %s'%e)
print_comment_v('%s%s'%(prefix,e.output.decode().replace('\n','\n'+prefix)))
return None
except:
print_comment_v('*** Unknown problem running command: %s'%e)
return None
print_comment("Finished execution", verbose)
|
[
"def",
"execute_command_in_dir",
"(",
"command",
",",
"directory",
",",
"verbose",
"=",
"DEFAULTS",
"[",
"'v'",
"]",
",",
"prefix",
"=",
"\"Output: \"",
",",
"env",
"=",
"None",
")",
":",
"if",
"os",
".",
"name",
"==",
"'nt'",
":",
"directory",
"=",
"os",
".",
"path",
".",
"normpath",
"(",
"directory",
")",
"print_comment",
"(",
"\"Executing: (%s) in directory: %s\"",
"%",
"(",
"command",
",",
"directory",
")",
",",
"verbose",
")",
"if",
"env",
"is",
"not",
"None",
":",
"print_comment",
"(",
"\"Extra env variables %s\"",
"%",
"(",
"env",
")",
",",
"verbose",
")",
"try",
":",
"if",
"os",
".",
"name",
"==",
"'nt'",
":",
"return_string",
"=",
"subprocess",
".",
"check_output",
"(",
"command",
",",
"cwd",
"=",
"directory",
",",
"shell",
"=",
"True",
",",
"env",
"=",
"env",
",",
"close_fds",
"=",
"False",
")",
"else",
":",
"return_string",
"=",
"subprocess",
".",
"check_output",
"(",
"command",
",",
"cwd",
"=",
"directory",
",",
"shell",
"=",
"True",
",",
"stderr",
"=",
"subprocess",
".",
"STDOUT",
",",
"env",
"=",
"env",
",",
"close_fds",
"=",
"True",
")",
"return_string",
"=",
"return_string",
".",
"decode",
"(",
"\"utf-8\"",
")",
"# For Python 3",
"print_comment",
"(",
"'Command completed. Output: \\n %s%s'",
"%",
"(",
"prefix",
",",
"return_string",
".",
"replace",
"(",
"'\\n'",
",",
"'\\n '",
"+",
"prefix",
")",
")",
",",
"verbose",
")",
"return",
"return_string",
"except",
"AttributeError",
":",
"# For python 2.6...",
"print_comment_v",
"(",
"'Assuming Python 2.6...'",
")",
"return_string",
"=",
"subprocess",
".",
"Popen",
"(",
"command",
",",
"cwd",
"=",
"directory",
",",
"shell",
"=",
"True",
",",
"stdout",
"=",
"subprocess",
".",
"PIPE",
")",
".",
"communicate",
"(",
")",
"[",
"0",
"]",
"return",
"return_string",
"except",
"subprocess",
".",
"CalledProcessError",
"as",
"e",
":",
"print_comment_v",
"(",
"'*** Problem running command: \\n %s'",
"%",
"e",
")",
"print_comment_v",
"(",
"'%s%s'",
"%",
"(",
"prefix",
",",
"e",
".",
"output",
".",
"decode",
"(",
")",
".",
"replace",
"(",
"'\\n'",
",",
"'\\n'",
"+",
"prefix",
")",
")",
")",
"return",
"None",
"except",
":",
"print_comment_v",
"(",
"'*** Unknown problem running command: %s'",
"%",
"e",
")",
"return",
"None",
"print_comment",
"(",
"\"Finished execution\"",
",",
"verbose",
")"
] |
Execute a command in specific working directory
|
[
"Execute",
"a",
"command",
"in",
"specific",
"working",
"directory"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/pynml.py#L1181-L1239
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/pynml.py
|
evaluate_component
|
def evaluate_component(comp_type, req_variables={}, parameter_values={}):
print_comment('Evaluating %s with req:%s; params:%s'%(comp_type.name,req_variables,parameter_values))
exec_str = ''
return_vals = {}
from math import exp
for p in parameter_values:
exec_str+='%s = %s\n'%(p, get_value_in_si(parameter_values[p]))
for r in req_variables:
exec_str+='%s = %s\n'%(r, get_value_in_si(req_variables[r]))
for c in comp_type.Constant:
exec_str+='%s = %s\n'%(c.name, get_value_in_si(c.value))
for d in comp_type.Dynamics:
for dv in d.DerivedVariable:
exec_str+='%s = %s\n'%(dv.name, dv.value)
exec_str+='return_vals["%s"] = %s\n'%(dv.name, dv.name)
for cdv in d.ConditionalDerivedVariable:
for case in cdv.Case:
if case.condition:
cond = case.condition.replace('.neq.','!=').replace('.eq.','==').replace('.gt.','<').replace('.lt.','<')
exec_str+='if ( %s ): %s = %s \n'%(cond, cdv.name, case.value)
else:
exec_str+='else: %s = %s \n'%(cdv.name, case.value)
exec_str+='\n'
exec_str+='return_vals["%s"] = %s\n'%(cdv.name, cdv.name)
'''print_comment_v(exec_str)'''
exec(exec_str)
return return_vals
|
python
|
def evaluate_component(comp_type, req_variables={}, parameter_values={}):
print_comment('Evaluating %s with req:%s; params:%s'%(comp_type.name,req_variables,parameter_values))
exec_str = ''
return_vals = {}
from math import exp
for p in parameter_values:
exec_str+='%s = %s\n'%(p, get_value_in_si(parameter_values[p]))
for r in req_variables:
exec_str+='%s = %s\n'%(r, get_value_in_si(req_variables[r]))
for c in comp_type.Constant:
exec_str+='%s = %s\n'%(c.name, get_value_in_si(c.value))
for d in comp_type.Dynamics:
for dv in d.DerivedVariable:
exec_str+='%s = %s\n'%(dv.name, dv.value)
exec_str+='return_vals["%s"] = %s\n'%(dv.name, dv.name)
for cdv in d.ConditionalDerivedVariable:
for case in cdv.Case:
if case.condition:
cond = case.condition.replace('.neq.','!=').replace('.eq.','==').replace('.gt.','<').replace('.lt.','<')
exec_str+='if ( %s ): %s = %s \n'%(cond, cdv.name, case.value)
else:
exec_str+='else: %s = %s \n'%(cdv.name, case.value)
exec_str+='\n'
exec_str+='return_vals["%s"] = %s\n'%(cdv.name, cdv.name)
'''print_comment_v(exec_str)'''
exec(exec_str)
return return_vals
|
[
"def",
"evaluate_component",
"(",
"comp_type",
",",
"req_variables",
"=",
"{",
"}",
",",
"parameter_values",
"=",
"{",
"}",
")",
":",
"print_comment",
"(",
"'Evaluating %s with req:%s; params:%s'",
"%",
"(",
"comp_type",
".",
"name",
",",
"req_variables",
",",
"parameter_values",
")",
")",
"exec_str",
"=",
"''",
"return_vals",
"=",
"{",
"}",
"from",
"math",
"import",
"exp",
"for",
"p",
"in",
"parameter_values",
":",
"exec_str",
"+=",
"'%s = %s\\n'",
"%",
"(",
"p",
",",
"get_value_in_si",
"(",
"parameter_values",
"[",
"p",
"]",
")",
")",
"for",
"r",
"in",
"req_variables",
":",
"exec_str",
"+=",
"'%s = %s\\n'",
"%",
"(",
"r",
",",
"get_value_in_si",
"(",
"req_variables",
"[",
"r",
"]",
")",
")",
"for",
"c",
"in",
"comp_type",
".",
"Constant",
":",
"exec_str",
"+=",
"'%s = %s\\n'",
"%",
"(",
"c",
".",
"name",
",",
"get_value_in_si",
"(",
"c",
".",
"value",
")",
")",
"for",
"d",
"in",
"comp_type",
".",
"Dynamics",
":",
"for",
"dv",
"in",
"d",
".",
"DerivedVariable",
":",
"exec_str",
"+=",
"'%s = %s\\n'",
"%",
"(",
"dv",
".",
"name",
",",
"dv",
".",
"value",
")",
"exec_str",
"+=",
"'return_vals[\"%s\"] = %s\\n'",
"%",
"(",
"dv",
".",
"name",
",",
"dv",
".",
"name",
")",
"for",
"cdv",
"in",
"d",
".",
"ConditionalDerivedVariable",
":",
"for",
"case",
"in",
"cdv",
".",
"Case",
":",
"if",
"case",
".",
"condition",
":",
"cond",
"=",
"case",
".",
"condition",
".",
"replace",
"(",
"'.neq.'",
",",
"'!='",
")",
".",
"replace",
"(",
"'.eq.'",
",",
"'=='",
")",
".",
"replace",
"(",
"'.gt.'",
",",
"'<'",
")",
".",
"replace",
"(",
"'.lt.'",
",",
"'<'",
")",
"exec_str",
"+=",
"'if ( %s ): %s = %s \\n'",
"%",
"(",
"cond",
",",
"cdv",
".",
"name",
",",
"case",
".",
"value",
")",
"else",
":",
"exec_str",
"+=",
"'else: %s = %s \\n'",
"%",
"(",
"cdv",
".",
"name",
",",
"case",
".",
"value",
")",
"exec_str",
"+=",
"'\\n'",
"exec_str",
"+=",
"'return_vals[\"%s\"] = %s\\n'",
"%",
"(",
"cdv",
".",
"name",
",",
"cdv",
".",
"name",
")",
"exec",
"(",
"exec_str",
")",
"return",
"return_vals"
] |
print_comment_v(exec_str)
|
[
"print_comment_v",
"(",
"exec_str",
")"
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/pynml.py#L1431-L1462
|
train
|
marrow/WebCore
|
web/ext/analytics.py
|
AnalyticsExtension.after
|
def after(self, context, exc=None):
"""Executed after dispatch has returned and the response populated, prior to anything being sent to the client."""
duration = context._duration = round((time.time() - context._start_time) * 1000) # Convert to ms.
delta = unicode(duration)
# Default response augmentation.
if self.header:
context.response.headers[self.header] = delta
if self.log:
self.log("Response generated in " + delta + " seconds.", extra=dict(
duration = duration,
request = id(context)
))
|
python
|
def after(self, context, exc=None):
"""Executed after dispatch has returned and the response populated, prior to anything being sent to the client."""
duration = context._duration = round((time.time() - context._start_time) * 1000) # Convert to ms.
delta = unicode(duration)
# Default response augmentation.
if self.header:
context.response.headers[self.header] = delta
if self.log:
self.log("Response generated in " + delta + " seconds.", extra=dict(
duration = duration,
request = id(context)
))
|
[
"def",
"after",
"(",
"self",
",",
"context",
",",
"exc",
"=",
"None",
")",
":",
"duration",
"=",
"context",
".",
"_duration",
"=",
"round",
"(",
"(",
"time",
".",
"time",
"(",
")",
"-",
"context",
".",
"_start_time",
")",
"*",
"1000",
")",
"# Convert to ms.",
"delta",
"=",
"unicode",
"(",
"duration",
")",
"# Default response augmentation.",
"if",
"self",
".",
"header",
":",
"context",
".",
"response",
".",
"headers",
"[",
"self",
".",
"header",
"]",
"=",
"delta",
"if",
"self",
".",
"log",
":",
"self",
".",
"log",
"(",
"\"Response generated in \"",
"+",
"delta",
"+",
"\" seconds.\"",
",",
"extra",
"=",
"dict",
"(",
"duration",
"=",
"duration",
",",
"request",
"=",
"id",
"(",
"context",
")",
")",
")"
] |
Executed after dispatch has returned and the response populated, prior to anything being sent to the client.
|
[
"Executed",
"after",
"dispatch",
"has",
"returned",
"and",
"the",
"response",
"populated",
"prior",
"to",
"anything",
"being",
"sent",
"to",
"the",
"client",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/analytics.py#L55-L69
|
train
|
marrow/WebCore
|
web/ext/args.py
|
ArgumentExtension._process_flat_kwargs
|
def _process_flat_kwargs(source, kwargs):
"""Apply a flat namespace transformation to recreate (in some respects) a rich structure.
This applies several transformations, which may be nested:
`foo` (singular): define a simple value named `foo`
`foo` (repeated): define a simple value for placement in an array named `foo`
`foo[]`: define a simple value for placement in an array, even if there is only one
`foo.<id>`: define a simple value to place in the `foo` array at the identified index
By nesting, you may define deeper, more complex structures:
`foo.bar`: define a value for the named element `bar` of the `foo` dictionary
`foo.<id>.bar`: define a `bar` dictionary element on the array element marked by that ID
References to `<id>` represent numeric "attributes", which makes the parent reference be treated as an array,
not a dictionary. Exact indexes might not be able to be preserved if there are voids; Python lists are not
sparse.
No validation of values is performed.
"""
ordered_arrays = []
# Process arguments one at a time and apply them to the kwargs passed in.
for name, value in source.items():
container = kwargs
if '.' in name:
parts = name.split('.')
name = name.rpartition('.')[2]
for target, following in zip(parts[:-1], parts[1:]):
if following.isnumeric(): # Prepare any use of numeric IDs.
container.setdefault(target, [{}])
if container[target] not in ordered_arrays:
ordered_arrays.append(container[target])
container = container[target][0]
continue
container = container.setdefault(target, {})
if name.endswith('[]'): # `foo[]` or `foo.bar[]` etc.
name = name[:-2]
container.setdefault(name, [])
container[name].append(value)
continue
if name.isnumeric() and container is not kwargs: # trailing identifiers, `foo.<id>`
container[int(name)] = value
continue
if name in container:
if not isinstance(container[name], list):
container[name] = [container[name]]
container[name].append(value)
continue
container[name] = value
for container in ordered_arrays:
elements = container[0]
del container[:]
container.extend(value for name, value in sorted(elements.items()))
|
python
|
def _process_flat_kwargs(source, kwargs):
"""Apply a flat namespace transformation to recreate (in some respects) a rich structure.
This applies several transformations, which may be nested:
`foo` (singular): define a simple value named `foo`
`foo` (repeated): define a simple value for placement in an array named `foo`
`foo[]`: define a simple value for placement in an array, even if there is only one
`foo.<id>`: define a simple value to place in the `foo` array at the identified index
By nesting, you may define deeper, more complex structures:
`foo.bar`: define a value for the named element `bar` of the `foo` dictionary
`foo.<id>.bar`: define a `bar` dictionary element on the array element marked by that ID
References to `<id>` represent numeric "attributes", which makes the parent reference be treated as an array,
not a dictionary. Exact indexes might not be able to be preserved if there are voids; Python lists are not
sparse.
No validation of values is performed.
"""
ordered_arrays = []
# Process arguments one at a time and apply them to the kwargs passed in.
for name, value in source.items():
container = kwargs
if '.' in name:
parts = name.split('.')
name = name.rpartition('.')[2]
for target, following in zip(parts[:-1], parts[1:]):
if following.isnumeric(): # Prepare any use of numeric IDs.
container.setdefault(target, [{}])
if container[target] not in ordered_arrays:
ordered_arrays.append(container[target])
container = container[target][0]
continue
container = container.setdefault(target, {})
if name.endswith('[]'): # `foo[]` or `foo.bar[]` etc.
name = name[:-2]
container.setdefault(name, [])
container[name].append(value)
continue
if name.isnumeric() and container is not kwargs: # trailing identifiers, `foo.<id>`
container[int(name)] = value
continue
if name in container:
if not isinstance(container[name], list):
container[name] = [container[name]]
container[name].append(value)
continue
container[name] = value
for container in ordered_arrays:
elements = container[0]
del container[:]
container.extend(value for name, value in sorted(elements.items()))
|
[
"def",
"_process_flat_kwargs",
"(",
"source",
",",
"kwargs",
")",
":",
"ordered_arrays",
"=",
"[",
"]",
"# Process arguments one at a time and apply them to the kwargs passed in.",
"for",
"name",
",",
"value",
"in",
"source",
".",
"items",
"(",
")",
":",
"container",
"=",
"kwargs",
"if",
"'.'",
"in",
"name",
":",
"parts",
"=",
"name",
".",
"split",
"(",
"'.'",
")",
"name",
"=",
"name",
".",
"rpartition",
"(",
"'.'",
")",
"[",
"2",
"]",
"for",
"target",
",",
"following",
"in",
"zip",
"(",
"parts",
"[",
":",
"-",
"1",
"]",
",",
"parts",
"[",
"1",
":",
"]",
")",
":",
"if",
"following",
".",
"isnumeric",
"(",
")",
":",
"# Prepare any use of numeric IDs.",
"container",
".",
"setdefault",
"(",
"target",
",",
"[",
"{",
"}",
"]",
")",
"if",
"container",
"[",
"target",
"]",
"not",
"in",
"ordered_arrays",
":",
"ordered_arrays",
".",
"append",
"(",
"container",
"[",
"target",
"]",
")",
"container",
"=",
"container",
"[",
"target",
"]",
"[",
"0",
"]",
"continue",
"container",
"=",
"container",
".",
"setdefault",
"(",
"target",
",",
"{",
"}",
")",
"if",
"name",
".",
"endswith",
"(",
"'[]'",
")",
":",
"# `foo[]` or `foo.bar[]` etc.",
"name",
"=",
"name",
"[",
":",
"-",
"2",
"]",
"container",
".",
"setdefault",
"(",
"name",
",",
"[",
"]",
")",
"container",
"[",
"name",
"]",
".",
"append",
"(",
"value",
")",
"continue",
"if",
"name",
".",
"isnumeric",
"(",
")",
"and",
"container",
"is",
"not",
"kwargs",
":",
"# trailing identifiers, `foo.<id>`",
"container",
"[",
"int",
"(",
"name",
")",
"]",
"=",
"value",
"continue",
"if",
"name",
"in",
"container",
":",
"if",
"not",
"isinstance",
"(",
"container",
"[",
"name",
"]",
",",
"list",
")",
":",
"container",
"[",
"name",
"]",
"=",
"[",
"container",
"[",
"name",
"]",
"]",
"container",
"[",
"name",
"]",
".",
"append",
"(",
"value",
")",
"continue",
"container",
"[",
"name",
"]",
"=",
"value",
"for",
"container",
"in",
"ordered_arrays",
":",
"elements",
"=",
"container",
"[",
"0",
"]",
"del",
"container",
"[",
":",
"]",
"container",
".",
"extend",
"(",
"value",
"for",
"name",
",",
"value",
"in",
"sorted",
"(",
"elements",
".",
"items",
"(",
")",
")",
")"
] |
Apply a flat namespace transformation to recreate (in some respects) a rich structure.
This applies several transformations, which may be nested:
`foo` (singular): define a simple value named `foo`
`foo` (repeated): define a simple value for placement in an array named `foo`
`foo[]`: define a simple value for placement in an array, even if there is only one
`foo.<id>`: define a simple value to place in the `foo` array at the identified index
By nesting, you may define deeper, more complex structures:
`foo.bar`: define a value for the named element `bar` of the `foo` dictionary
`foo.<id>.bar`: define a `bar` dictionary element on the array element marked by that ID
References to `<id>` represent numeric "attributes", which makes the parent reference be treated as an array,
not a dictionary. Exact indexes might not be able to be preserved if there are voids; Python lists are not
sparse.
No validation of values is performed.
|
[
"Apply",
"a",
"flat",
"namespace",
"transformation",
"to",
"recreate",
"(",
"in",
"some",
"respects",
")",
"a",
"rich",
"structure",
".",
"This",
"applies",
"several",
"transformations",
"which",
"may",
"be",
"nested",
":",
"foo",
"(",
"singular",
")",
":",
"define",
"a",
"simple",
"value",
"named",
"foo",
"foo",
"(",
"repeated",
")",
":",
"define",
"a",
"simple",
"value",
"for",
"placement",
"in",
"an",
"array",
"named",
"foo",
"foo",
"[]",
":",
"define",
"a",
"simple",
"value",
"for",
"placement",
"in",
"an",
"array",
"even",
"if",
"there",
"is",
"only",
"one",
"foo",
".",
"<id",
">",
":",
"define",
"a",
"simple",
"value",
"to",
"place",
"in",
"the",
"foo",
"array",
"at",
"the",
"identified",
"index",
"By",
"nesting",
"you",
"may",
"define",
"deeper",
"more",
"complex",
"structures",
":",
"foo",
".",
"bar",
":",
"define",
"a",
"value",
"for",
"the",
"named",
"element",
"bar",
"of",
"the",
"foo",
"dictionary",
"foo",
".",
"<id",
">",
".",
"bar",
":",
"define",
"a",
"bar",
"dictionary",
"element",
"on",
"the",
"array",
"element",
"marked",
"by",
"that",
"ID",
"References",
"to",
"<id",
">",
"represent",
"numeric",
"attributes",
"which",
"makes",
"the",
"parent",
"reference",
"be",
"treated",
"as",
"an",
"array",
"not",
"a",
"dictionary",
".",
"Exact",
"indexes",
"might",
"not",
"be",
"able",
"to",
"be",
"preserved",
"if",
"there",
"are",
"voids",
";",
"Python",
"lists",
"are",
"not",
"sparse",
".",
"No",
"validation",
"of",
"values",
"is",
"performed",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/ext/args.py#L22-L87
|
train
|
NeuroML/pyNeuroML
|
pyneuroml/povray/NeuroML2ToPOVRay.py
|
process_args
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(description="A file for converting NeuroML v2 files into POVRay files for 3D rendering")
parser.add_argument('neuroml_file', type=str, metavar='<NeuroML file>',
help='NeuroML (version 2 beta 3+) file to be converted to PovRay format (XML or HDF5 format)')
parser.add_argument('-split',
action='store_true',
default=False,
help="If this is specified, generate separate pov files for cells & network. Default is false")
parser.add_argument('-background',
type=str,
metavar='<background colour>',
default=_WHITE,
help='Colour of background, e.g. <0,0,0,0.55>')
parser.add_argument('-movie',
action='store_true',
default=False,
help="If this is specified, generate a ini file for generating a sequence of frames for a movie of the 3D structure")
parser.add_argument('-inputs',
action='store_true',
default=False,
help="If this is specified, show the locations of (synaptic, current clamp, etc.) inputs into the cells of the network")
parser.add_argument('-conns',
action='store_true',
default=False,
help="If this is specified, show the connections present in the network with lines")
parser.add_argument('-conn_points',
action='store_true',
default=False,
help="If this is specified, show the end points of the connections present in the network")
parser.add_argument('-v',
action='store_true',
default=False,
help="Verbose output")
parser.add_argument('-frames',
type=int,
metavar='<frames>',
default=36,
help='Number of frames in movie')
parser.add_argument('-posx',
type=float,
metavar='<position offset x>',
default=0,
help='Offset position in x dir (0 is centre, 1 is top)')
parser.add_argument('-posy',
type=float,
metavar='<position offset y>',
default=0,
help='Offset position in y dir (0 is centre, 1 is top)')
parser.add_argument('-posz',
type=float,
metavar='<position offset z>',
default=0,
help='Offset position in z dir (0 is centre, 1 is top)')
parser.add_argument('-viewx',
type=float,
metavar='<view offset x>',
default=0,
help='Offset viewing point in x dir (0 is centre, 1 is top)')
parser.add_argument('-viewy',
type=float,
metavar='<view offset y>',
default=0,
help='Offset viewing point in y dir (0 is centre, 1 is top)')
parser.add_argument('-viewz',
type=float,
metavar='<view offset z>',
default=0,
help='Offset viewing point in z dir (0 is centre, 1 is top)')
parser.add_argument('-scalex',
type=float,
metavar='<scale position x>',
default=1,
help='Scale position from network in x dir')
parser.add_argument('-scaley',
type=float,
metavar='<scale position y>',
default=1.5,
help='Scale position from network in y dir')
parser.add_argument('-scalez',
type=float,
metavar='<scale position z>',
default=1,
help='Scale position from network in z dir')
parser.add_argument('-mindiam',
type=float,
metavar='<minimum diameter dendrites/axons>',
default=0,
help='Minimum diameter for dendrites/axons (to improve visualisations)')
parser.add_argument('-plane',
action='store_true',
default=False,
help="If this is specified, add a 2D plane below cell/network")
parser.add_argument('-segids',
action='store_true',
default=False,
help="Show segment ids")
return parser.parse_args()
|
python
|
def process_args():
"""
Parse command-line arguments.
"""
parser = argparse.ArgumentParser(description="A file for converting NeuroML v2 files into POVRay files for 3D rendering")
parser.add_argument('neuroml_file', type=str, metavar='<NeuroML file>',
help='NeuroML (version 2 beta 3+) file to be converted to PovRay format (XML or HDF5 format)')
parser.add_argument('-split',
action='store_true',
default=False,
help="If this is specified, generate separate pov files for cells & network. Default is false")
parser.add_argument('-background',
type=str,
metavar='<background colour>',
default=_WHITE,
help='Colour of background, e.g. <0,0,0,0.55>')
parser.add_argument('-movie',
action='store_true',
default=False,
help="If this is specified, generate a ini file for generating a sequence of frames for a movie of the 3D structure")
parser.add_argument('-inputs',
action='store_true',
default=False,
help="If this is specified, show the locations of (synaptic, current clamp, etc.) inputs into the cells of the network")
parser.add_argument('-conns',
action='store_true',
default=False,
help="If this is specified, show the connections present in the network with lines")
parser.add_argument('-conn_points',
action='store_true',
default=False,
help="If this is specified, show the end points of the connections present in the network")
parser.add_argument('-v',
action='store_true',
default=False,
help="Verbose output")
parser.add_argument('-frames',
type=int,
metavar='<frames>',
default=36,
help='Number of frames in movie')
parser.add_argument('-posx',
type=float,
metavar='<position offset x>',
default=0,
help='Offset position in x dir (0 is centre, 1 is top)')
parser.add_argument('-posy',
type=float,
metavar='<position offset y>',
default=0,
help='Offset position in y dir (0 is centre, 1 is top)')
parser.add_argument('-posz',
type=float,
metavar='<position offset z>',
default=0,
help='Offset position in z dir (0 is centre, 1 is top)')
parser.add_argument('-viewx',
type=float,
metavar='<view offset x>',
default=0,
help='Offset viewing point in x dir (0 is centre, 1 is top)')
parser.add_argument('-viewy',
type=float,
metavar='<view offset y>',
default=0,
help='Offset viewing point in y dir (0 is centre, 1 is top)')
parser.add_argument('-viewz',
type=float,
metavar='<view offset z>',
default=0,
help='Offset viewing point in z dir (0 is centre, 1 is top)')
parser.add_argument('-scalex',
type=float,
metavar='<scale position x>',
default=1,
help='Scale position from network in x dir')
parser.add_argument('-scaley',
type=float,
metavar='<scale position y>',
default=1.5,
help='Scale position from network in y dir')
parser.add_argument('-scalez',
type=float,
metavar='<scale position z>',
default=1,
help='Scale position from network in z dir')
parser.add_argument('-mindiam',
type=float,
metavar='<minimum diameter dendrites/axons>',
default=0,
help='Minimum diameter for dendrites/axons (to improve visualisations)')
parser.add_argument('-plane',
action='store_true',
default=False,
help="If this is specified, add a 2D plane below cell/network")
parser.add_argument('-segids',
action='store_true',
default=False,
help="Show segment ids")
return parser.parse_args()
|
[
"def",
"process_args",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"\"A file for converting NeuroML v2 files into POVRay files for 3D rendering\"",
")",
"parser",
".",
"add_argument",
"(",
"'neuroml_file'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<NeuroML file>'",
",",
"help",
"=",
"'NeuroML (version 2 beta 3+) file to be converted to PovRay format (XML or HDF5 format)'",
")",
"parser",
".",
"add_argument",
"(",
"'-split'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, generate separate pov files for cells & network. Default is false\"",
")",
"parser",
".",
"add_argument",
"(",
"'-background'",
",",
"type",
"=",
"str",
",",
"metavar",
"=",
"'<background colour>'",
",",
"default",
"=",
"_WHITE",
",",
"help",
"=",
"'Colour of background, e.g. <0,0,0,0.55>'",
")",
"parser",
".",
"add_argument",
"(",
"'-movie'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, generate a ini file for generating a sequence of frames for a movie of the 3D structure\"",
")",
"parser",
".",
"add_argument",
"(",
"'-inputs'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, show the locations of (synaptic, current clamp, etc.) inputs into the cells of the network\"",
")",
"parser",
".",
"add_argument",
"(",
"'-conns'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, show the connections present in the network with lines\"",
")",
"parser",
".",
"add_argument",
"(",
"'-conn_points'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, show the end points of the connections present in the network\"",
")",
"parser",
".",
"add_argument",
"(",
"'-v'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"Verbose output\"",
")",
"parser",
".",
"add_argument",
"(",
"'-frames'",
",",
"type",
"=",
"int",
",",
"metavar",
"=",
"'<frames>'",
",",
"default",
"=",
"36",
",",
"help",
"=",
"'Number of frames in movie'",
")",
"parser",
".",
"add_argument",
"(",
"'-posx'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<position offset x>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Offset position in x dir (0 is centre, 1 is top)'",
")",
"parser",
".",
"add_argument",
"(",
"'-posy'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<position offset y>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Offset position in y dir (0 is centre, 1 is top)'",
")",
"parser",
".",
"add_argument",
"(",
"'-posz'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<position offset z>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Offset position in z dir (0 is centre, 1 is top)'",
")",
"parser",
".",
"add_argument",
"(",
"'-viewx'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<view offset x>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Offset viewing point in x dir (0 is centre, 1 is top)'",
")",
"parser",
".",
"add_argument",
"(",
"'-viewy'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<view offset y>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Offset viewing point in y dir (0 is centre, 1 is top)'",
")",
"parser",
".",
"add_argument",
"(",
"'-viewz'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<view offset z>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Offset viewing point in z dir (0 is centre, 1 is top)'",
")",
"parser",
".",
"add_argument",
"(",
"'-scalex'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<scale position x>'",
",",
"default",
"=",
"1",
",",
"help",
"=",
"'Scale position from network in x dir'",
")",
"parser",
".",
"add_argument",
"(",
"'-scaley'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<scale position y>'",
",",
"default",
"=",
"1.5",
",",
"help",
"=",
"'Scale position from network in y dir'",
")",
"parser",
".",
"add_argument",
"(",
"'-scalez'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<scale position z>'",
",",
"default",
"=",
"1",
",",
"help",
"=",
"'Scale position from network in z dir'",
")",
"parser",
".",
"add_argument",
"(",
"'-mindiam'",
",",
"type",
"=",
"float",
",",
"metavar",
"=",
"'<minimum diameter dendrites/axons>'",
",",
"default",
"=",
"0",
",",
"help",
"=",
"'Minimum diameter for dendrites/axons (to improve visualisations)'",
")",
"parser",
".",
"add_argument",
"(",
"'-plane'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"If this is specified, add a 2D plane below cell/network\"",
")",
"parser",
".",
"add_argument",
"(",
"'-segids'",
",",
"action",
"=",
"'store_true'",
",",
"default",
"=",
"False",
",",
"help",
"=",
"\"Show segment ids\"",
")",
"return",
"parser",
".",
"parse_args",
"(",
")"
] |
Parse command-line arguments.
|
[
"Parse",
"command",
"-",
"line",
"arguments",
"."
] |
aeba2e3040b360bb26556f643cccbfb3dac3b8fb
|
https://github.com/NeuroML/pyNeuroML/blob/aeba2e3040b360bb26556f643cccbfb3dac3b8fb/pyneuroml/povray/NeuroML2ToPOVRay.py#L27-L143
|
train
|
marrow/WebCore
|
web/core/application.py
|
Application._configure
|
def _configure(self, config):
"""Prepare the incoming configuration and ensure certain expected values are present.
For example, this ensures BaseExtension is included in the extension list, and populates the logging config.
"""
config = config or dict()
# We really need this to be there.
if 'extensions' not in config: config['extensions'] = list()
if not any(isinstance(ext, BaseExtension) for ext in config['extensions']):
# Always make sure the BaseExtension is present since request/response objects are handy.
config['extensions'].insert(0, BaseExtension())
if not any(isinstance(ext, arguments.ArgumentExtension) for ext in config['extensions']):
# Prepare a default set of argument mutators.
config['extensions'].extend([
arguments.ValidateArgumentsExtension(),
arguments.ContextArgsExtension(),
arguments.RemainderArgsExtension(),
arguments.QueryStringArgsExtension(),
arguments.FormEncodedKwargsExtension(),
arguments.JSONKwargsExtension(),
])
config['extensions'].append(self) # Allow the application object itself to register callbacks.
try:
addLoggingLevel('trace', logging.DEBUG - 5)
except AttributeError:
pass
# Tests are skipped on these as we have no particular need to test Python's own logging mechanism.
level = config.get('logging', {}).get('level', None)
if level: # pragma: no cover
logging.basicConfig(level=getattr(logging, level.upper()))
elif 'logging' in config: # pragma: no cover
logging.config.dictConfig(config['logging'])
return config
|
python
|
def _configure(self, config):
"""Prepare the incoming configuration and ensure certain expected values are present.
For example, this ensures BaseExtension is included in the extension list, and populates the logging config.
"""
config = config or dict()
# We really need this to be there.
if 'extensions' not in config: config['extensions'] = list()
if not any(isinstance(ext, BaseExtension) for ext in config['extensions']):
# Always make sure the BaseExtension is present since request/response objects are handy.
config['extensions'].insert(0, BaseExtension())
if not any(isinstance(ext, arguments.ArgumentExtension) for ext in config['extensions']):
# Prepare a default set of argument mutators.
config['extensions'].extend([
arguments.ValidateArgumentsExtension(),
arguments.ContextArgsExtension(),
arguments.RemainderArgsExtension(),
arguments.QueryStringArgsExtension(),
arguments.FormEncodedKwargsExtension(),
arguments.JSONKwargsExtension(),
])
config['extensions'].append(self) # Allow the application object itself to register callbacks.
try:
addLoggingLevel('trace', logging.DEBUG - 5)
except AttributeError:
pass
# Tests are skipped on these as we have no particular need to test Python's own logging mechanism.
level = config.get('logging', {}).get('level', None)
if level: # pragma: no cover
logging.basicConfig(level=getattr(logging, level.upper()))
elif 'logging' in config: # pragma: no cover
logging.config.dictConfig(config['logging'])
return config
|
[
"def",
"_configure",
"(",
"self",
",",
"config",
")",
":",
"config",
"=",
"config",
"or",
"dict",
"(",
")",
"# We really need this to be there.",
"if",
"'extensions'",
"not",
"in",
"config",
":",
"config",
"[",
"'extensions'",
"]",
"=",
"list",
"(",
")",
"if",
"not",
"any",
"(",
"isinstance",
"(",
"ext",
",",
"BaseExtension",
")",
"for",
"ext",
"in",
"config",
"[",
"'extensions'",
"]",
")",
":",
"# Always make sure the BaseExtension is present since request/response objects are handy.",
"config",
"[",
"'extensions'",
"]",
".",
"insert",
"(",
"0",
",",
"BaseExtension",
"(",
")",
")",
"if",
"not",
"any",
"(",
"isinstance",
"(",
"ext",
",",
"arguments",
".",
"ArgumentExtension",
")",
"for",
"ext",
"in",
"config",
"[",
"'extensions'",
"]",
")",
":",
"# Prepare a default set of argument mutators.",
"config",
"[",
"'extensions'",
"]",
".",
"extend",
"(",
"[",
"arguments",
".",
"ValidateArgumentsExtension",
"(",
")",
",",
"arguments",
".",
"ContextArgsExtension",
"(",
")",
",",
"arguments",
".",
"RemainderArgsExtension",
"(",
")",
",",
"arguments",
".",
"QueryStringArgsExtension",
"(",
")",
",",
"arguments",
".",
"FormEncodedKwargsExtension",
"(",
")",
",",
"arguments",
".",
"JSONKwargsExtension",
"(",
")",
",",
"]",
")",
"config",
"[",
"'extensions'",
"]",
".",
"append",
"(",
"self",
")",
"# Allow the application object itself to register callbacks.",
"try",
":",
"addLoggingLevel",
"(",
"'trace'",
",",
"logging",
".",
"DEBUG",
"-",
"5",
")",
"except",
"AttributeError",
":",
"pass",
"# Tests are skipped on these as we have no particular need to test Python's own logging mechanism.",
"level",
"=",
"config",
".",
"get",
"(",
"'logging'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'level'",
",",
"None",
")",
"if",
"level",
":",
"# pragma: no cover",
"logging",
".",
"basicConfig",
"(",
"level",
"=",
"getattr",
"(",
"logging",
",",
"level",
".",
"upper",
"(",
")",
")",
")",
"elif",
"'logging'",
"in",
"config",
":",
"# pragma: no cover",
"logging",
".",
"config",
".",
"dictConfig",
"(",
"config",
"[",
"'logging'",
"]",
")",
"return",
"config"
] |
Prepare the incoming configuration and ensure certain expected values are present.
For example, this ensures BaseExtension is included in the extension list, and populates the logging config.
|
[
"Prepare",
"the",
"incoming",
"configuration",
"and",
"ensure",
"certain",
"expected",
"values",
"are",
"present",
".",
"For",
"example",
"this",
"ensures",
"BaseExtension",
"is",
"included",
"in",
"the",
"extension",
"list",
"and",
"populates",
"the",
"logging",
"config",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/core/application.py#L108-L147
|
train
|
marrow/WebCore
|
web/core/application.py
|
Application.serve
|
def serve(self, service='auto', **options): # pragma: no cover
"""Initiate a web server service to serve this application.
You can always use the Application instance as a bare WSGI application, of course. This method is provided as
a convienence.
Pass in the name of the service you wish to use, and any additional configuration options appropriate for that
service. Almost all services accept `host` and `port` options, some also allow you to specify an on-disk
`socket`. By default all web servers will listen to `127.0.0.1` (loopback only) on port 8080.
"""
service = load(service, 'web.server') # We don't bother with a full registry for these one-time lookups.
try:
service(self, **options)
except KeyboardInterrupt: # We catch this as SIG_TERM or ^C are basically the only ways to stop most servers.
pass
# Notify extensions that the service has returned and we are exiting.
for ext in self.__context.extension.signal.stop: ext(self.__context)
|
python
|
def serve(self, service='auto', **options): # pragma: no cover
"""Initiate a web server service to serve this application.
You can always use the Application instance as a bare WSGI application, of course. This method is provided as
a convienence.
Pass in the name of the service you wish to use, and any additional configuration options appropriate for that
service. Almost all services accept `host` and `port` options, some also allow you to specify an on-disk
`socket`. By default all web servers will listen to `127.0.0.1` (loopback only) on port 8080.
"""
service = load(service, 'web.server') # We don't bother with a full registry for these one-time lookups.
try:
service(self, **options)
except KeyboardInterrupt: # We catch this as SIG_TERM or ^C are basically the only ways to stop most servers.
pass
# Notify extensions that the service has returned and we are exiting.
for ext in self.__context.extension.signal.stop: ext(self.__context)
|
[
"def",
"serve",
"(",
"self",
",",
"service",
"=",
"'auto'",
",",
"*",
"*",
"options",
")",
":",
"# pragma: no cover",
"service",
"=",
"load",
"(",
"service",
",",
"'web.server'",
")",
"# We don't bother with a full registry for these one-time lookups.",
"try",
":",
"service",
"(",
"self",
",",
"*",
"*",
"options",
")",
"except",
"KeyboardInterrupt",
":",
"# We catch this as SIG_TERM or ^C are basically the only ways to stop most servers.",
"pass",
"# Notify extensions that the service has returned and we are exiting.",
"for",
"ext",
"in",
"self",
".",
"__context",
".",
"extension",
".",
"signal",
".",
"stop",
":",
"ext",
"(",
"self",
".",
"__context",
")"
] |
Initiate a web server service to serve this application.
You can always use the Application instance as a bare WSGI application, of course. This method is provided as
a convienence.
Pass in the name of the service you wish to use, and any additional configuration options appropriate for that
service. Almost all services accept `host` and `port` options, some also allow you to specify an on-disk
`socket`. By default all web servers will listen to `127.0.0.1` (loopback only) on port 8080.
|
[
"Initiate",
"a",
"web",
"server",
"service",
"to",
"serve",
"this",
"application",
".",
"You",
"can",
"always",
"use",
"the",
"Application",
"instance",
"as",
"a",
"bare",
"WSGI",
"application",
"of",
"course",
".",
"This",
"method",
"is",
"provided",
"as",
"a",
"convienence",
".",
"Pass",
"in",
"the",
"name",
"of",
"the",
"service",
"you",
"wish",
"to",
"use",
"and",
"any",
"additional",
"configuration",
"options",
"appropriate",
"for",
"that",
"service",
".",
"Almost",
"all",
"services",
"accept",
"host",
"and",
"port",
"options",
"some",
"also",
"allow",
"you",
"to",
"specify",
"an",
"on",
"-",
"disk",
"socket",
".",
"By",
"default",
"all",
"web",
"servers",
"will",
"listen",
"to",
"127",
".",
"0",
".",
"0",
".",
"1",
"(",
"loopback",
"only",
")",
"on",
"port",
"8080",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/core/application.py#L151-L170
|
train
|
marrow/WebCore
|
web/core/application.py
|
Application.application
|
def application(self, environ, start_response):
"""Process a single WSGI request/response cycle.
This is the WSGI handler for WebCore. Depending on the presence of extensions providing WSGI middleware,
the `__call__` attribute of the Application instance will either become this, or become the outermost
middleware callable.
Most apps won't utilize middleware, the extension interface is preferred for most operations in WebCore.
They allow for code injection at various intermediary steps in the processing of a request and response.
"""
context = environ['wc.context'] = self.RequestContext(environ=environ)
signals = context.extension.signal
# Announce the start of a request cycle. This executes `prepare` and `before` callbacks in the correct order.
for ext in signals.pre: ext(context)
# Identify the endpoint for this request.
is_endpoint, handler = context.dispatch(context, context.root, context.environ['PATH_INFO'])
if is_endpoint:
try:
result = self._execute_endpoint(context, handler, signals) # Process the endpoint.
except Exception as e:
log.exception("Caught exception attempting to execute the endpoint.")
result = HTTPInternalServerError(str(e) if __debug__ else "Please see the logs.")
if 'debugger' in context.extension.feature:
context.response = result
for ext in signals.after: ext(context) # Allow signals to clean up early.
raise
else: # If no endpoint could be resolved, that's a 404.
result = HTTPNotFound("Dispatch failed." if __debug__ else None)
if __debug__:
log.debug("Result prepared, identifying view handler.", extra=dict(
request = id(context),
result = safe_name(type(result))
))
# Identify a view capable of handling this result.
for view in context.view(result):
if view(context, result): break
else:
# We've run off the bottom of the list of possible views.
raise TypeError("No view could be found to handle: " + repr(type(result)))
if __debug__:
log.debug("View identified, populating response.", extra=dict(
request = id(context),
view = repr(view),
))
for ext in signals.after: ext(context)
def capture_done(response):
for chunk in response:
yield chunk
for ext in signals.done: ext(context)
# This is really long due to the fact we don't want to capture the response too early.
# We need anything up to this point to be able to simply replace `context.response` if needed.
return capture_done(context.response.conditional_response_app(environ, start_response))
|
python
|
def application(self, environ, start_response):
"""Process a single WSGI request/response cycle.
This is the WSGI handler for WebCore. Depending on the presence of extensions providing WSGI middleware,
the `__call__` attribute of the Application instance will either become this, or become the outermost
middleware callable.
Most apps won't utilize middleware, the extension interface is preferred for most operations in WebCore.
They allow for code injection at various intermediary steps in the processing of a request and response.
"""
context = environ['wc.context'] = self.RequestContext(environ=environ)
signals = context.extension.signal
# Announce the start of a request cycle. This executes `prepare` and `before` callbacks in the correct order.
for ext in signals.pre: ext(context)
# Identify the endpoint for this request.
is_endpoint, handler = context.dispatch(context, context.root, context.environ['PATH_INFO'])
if is_endpoint:
try:
result = self._execute_endpoint(context, handler, signals) # Process the endpoint.
except Exception as e:
log.exception("Caught exception attempting to execute the endpoint.")
result = HTTPInternalServerError(str(e) if __debug__ else "Please see the logs.")
if 'debugger' in context.extension.feature:
context.response = result
for ext in signals.after: ext(context) # Allow signals to clean up early.
raise
else: # If no endpoint could be resolved, that's a 404.
result = HTTPNotFound("Dispatch failed." if __debug__ else None)
if __debug__:
log.debug("Result prepared, identifying view handler.", extra=dict(
request = id(context),
result = safe_name(type(result))
))
# Identify a view capable of handling this result.
for view in context.view(result):
if view(context, result): break
else:
# We've run off the bottom of the list of possible views.
raise TypeError("No view could be found to handle: " + repr(type(result)))
if __debug__:
log.debug("View identified, populating response.", extra=dict(
request = id(context),
view = repr(view),
))
for ext in signals.after: ext(context)
def capture_done(response):
for chunk in response:
yield chunk
for ext in signals.done: ext(context)
# This is really long due to the fact we don't want to capture the response too early.
# We need anything up to this point to be able to simply replace `context.response` if needed.
return capture_done(context.response.conditional_response_app(environ, start_response))
|
[
"def",
"application",
"(",
"self",
",",
"environ",
",",
"start_response",
")",
":",
"context",
"=",
"environ",
"[",
"'wc.context'",
"]",
"=",
"self",
".",
"RequestContext",
"(",
"environ",
"=",
"environ",
")",
"signals",
"=",
"context",
".",
"extension",
".",
"signal",
"# Announce the start of a request cycle. This executes `prepare` and `before` callbacks in the correct order.",
"for",
"ext",
"in",
"signals",
".",
"pre",
":",
"ext",
"(",
"context",
")",
"# Identify the endpoint for this request.",
"is_endpoint",
",",
"handler",
"=",
"context",
".",
"dispatch",
"(",
"context",
",",
"context",
".",
"root",
",",
"context",
".",
"environ",
"[",
"'PATH_INFO'",
"]",
")",
"if",
"is_endpoint",
":",
"try",
":",
"result",
"=",
"self",
".",
"_execute_endpoint",
"(",
"context",
",",
"handler",
",",
"signals",
")",
"# Process the endpoint.",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"exception",
"(",
"\"Caught exception attempting to execute the endpoint.\"",
")",
"result",
"=",
"HTTPInternalServerError",
"(",
"str",
"(",
"e",
")",
"if",
"__debug__",
"else",
"\"Please see the logs.\"",
")",
"if",
"'debugger'",
"in",
"context",
".",
"extension",
".",
"feature",
":",
"context",
".",
"response",
"=",
"result",
"for",
"ext",
"in",
"signals",
".",
"after",
":",
"ext",
"(",
"context",
")",
"# Allow signals to clean up early.",
"raise",
"else",
":",
"# If no endpoint could be resolved, that's a 404.",
"result",
"=",
"HTTPNotFound",
"(",
"\"Dispatch failed.\"",
"if",
"__debug__",
"else",
"None",
")",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"Result prepared, identifying view handler.\"",
",",
"extra",
"=",
"dict",
"(",
"request",
"=",
"id",
"(",
"context",
")",
",",
"result",
"=",
"safe_name",
"(",
"type",
"(",
"result",
")",
")",
")",
")",
"# Identify a view capable of handling this result.",
"for",
"view",
"in",
"context",
".",
"view",
"(",
"result",
")",
":",
"if",
"view",
"(",
"context",
",",
"result",
")",
":",
"break",
"else",
":",
"# We've run off the bottom of the list of possible views.",
"raise",
"TypeError",
"(",
"\"No view could be found to handle: \"",
"+",
"repr",
"(",
"type",
"(",
"result",
")",
")",
")",
"if",
"__debug__",
":",
"log",
".",
"debug",
"(",
"\"View identified, populating response.\"",
",",
"extra",
"=",
"dict",
"(",
"request",
"=",
"id",
"(",
"context",
")",
",",
"view",
"=",
"repr",
"(",
"view",
")",
",",
")",
")",
"for",
"ext",
"in",
"signals",
".",
"after",
":",
"ext",
"(",
"context",
")",
"def",
"capture_done",
"(",
"response",
")",
":",
"for",
"chunk",
"in",
"response",
":",
"yield",
"chunk",
"for",
"ext",
"in",
"signals",
".",
"done",
":",
"ext",
"(",
"context",
")",
"# This is really long due to the fact we don't want to capture the response too early.",
"# We need anything up to this point to be able to simply replace `context.response` if needed.",
"return",
"capture_done",
"(",
"context",
".",
"response",
".",
"conditional_response_app",
"(",
"environ",
",",
"start_response",
")",
")"
] |
Process a single WSGI request/response cycle.
This is the WSGI handler for WebCore. Depending on the presence of extensions providing WSGI middleware,
the `__call__` attribute of the Application instance will either become this, or become the outermost
middleware callable.
Most apps won't utilize middleware, the extension interface is preferred for most operations in WebCore.
They allow for code injection at various intermediary steps in the processing of a request and response.
|
[
"Process",
"a",
"single",
"WSGI",
"request",
"/",
"response",
"cycle",
".",
"This",
"is",
"the",
"WSGI",
"handler",
"for",
"WebCore",
".",
"Depending",
"on",
"the",
"presence",
"of",
"extensions",
"providing",
"WSGI",
"middleware",
"the",
"__call__",
"attribute",
"of",
"the",
"Application",
"instance",
"will",
"either",
"become",
"this",
"or",
"become",
"the",
"outermost",
"middleware",
"callable",
".",
"Most",
"apps",
"won",
"t",
"utilize",
"middleware",
"the",
"extension",
"interface",
"is",
"preferred",
"for",
"most",
"operations",
"in",
"WebCore",
".",
"They",
"allow",
"for",
"code",
"injection",
"at",
"various",
"intermediary",
"steps",
"in",
"the",
"processing",
"of",
"a",
"request",
"and",
"response",
"."
] |
38d50f8022ca62976a1e5ff23f7714bd647b6532
|
https://github.com/marrow/WebCore/blob/38d50f8022ca62976a1e5ff23f7714bd647b6532/web/core/application.py#L217-L280
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment._swap
|
def _swap(self):
'''Swaps the alignment so that the reference becomes the query and vice-versa. Swaps their names, coordinates etc. The frame is not changed'''
self.ref_start, self.qry_start = self.qry_start, self.ref_start
self.ref_end, self.qry_end = self.qry_end, self.ref_end
self.hit_length_ref, self.hit_length_qry = self.hit_length_qry, self.hit_length_ref
self.ref_length, self.qry_length = self.qry_length, self.ref_length
self.ref_name, self.qry_name = self.qry_name, self.ref_name
|
python
|
def _swap(self):
'''Swaps the alignment so that the reference becomes the query and vice-versa. Swaps their names, coordinates etc. The frame is not changed'''
self.ref_start, self.qry_start = self.qry_start, self.ref_start
self.ref_end, self.qry_end = self.qry_end, self.ref_end
self.hit_length_ref, self.hit_length_qry = self.hit_length_qry, self.hit_length_ref
self.ref_length, self.qry_length = self.qry_length, self.ref_length
self.ref_name, self.qry_name = self.qry_name, self.ref_name
|
[
"def",
"_swap",
"(",
"self",
")",
":",
"self",
".",
"ref_start",
",",
"self",
".",
"qry_start",
"=",
"self",
".",
"qry_start",
",",
"self",
".",
"ref_start",
"self",
".",
"ref_end",
",",
"self",
".",
"qry_end",
"=",
"self",
".",
"qry_end",
",",
"self",
".",
"ref_end",
"self",
".",
"hit_length_ref",
",",
"self",
".",
"hit_length_qry",
"=",
"self",
".",
"hit_length_qry",
",",
"self",
".",
"hit_length_ref",
"self",
".",
"ref_length",
",",
"self",
".",
"qry_length",
"=",
"self",
".",
"qry_length",
",",
"self",
".",
"ref_length",
"self",
".",
"ref_name",
",",
"self",
".",
"qry_name",
"=",
"self",
".",
"qry_name",
",",
"self",
".",
"ref_name"
] |
Swaps the alignment so that the reference becomes the query and vice-versa. Swaps their names, coordinates etc. The frame is not changed
|
[
"Swaps",
"the",
"alignment",
"so",
"that",
"the",
"reference",
"becomes",
"the",
"query",
"and",
"vice",
"-",
"versa",
".",
"Swaps",
"their",
"names",
"coordinates",
"etc",
".",
"The",
"frame",
"is",
"not",
"changed"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L52-L58
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.qry_coords
|
def qry_coords(self):
'''Returns a pyfastaq.intervals.Interval object of the start and end coordinates in the query sequence'''
return pyfastaq.intervals.Interval(min(self.qry_start, self.qry_end), max(self.qry_start, self.qry_end))
|
python
|
def qry_coords(self):
'''Returns a pyfastaq.intervals.Interval object of the start and end coordinates in the query sequence'''
return pyfastaq.intervals.Interval(min(self.qry_start, self.qry_end), max(self.qry_start, self.qry_end))
|
[
"def",
"qry_coords",
"(",
"self",
")",
":",
"return",
"pyfastaq",
".",
"intervals",
".",
"Interval",
"(",
"min",
"(",
"self",
".",
"qry_start",
",",
"self",
".",
"qry_end",
")",
",",
"max",
"(",
"self",
".",
"qry_start",
",",
"self",
".",
"qry_end",
")",
")"
] |
Returns a pyfastaq.intervals.Interval object of the start and end coordinates in the query sequence
|
[
"Returns",
"a",
"pyfastaq",
".",
"intervals",
".",
"Interval",
"object",
"of",
"the",
"start",
"and",
"end",
"coordinates",
"in",
"the",
"query",
"sequence"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L61-L63
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.ref_coords
|
def ref_coords(self):
'''Returns a pyfastaq.intervals.Interval object of the start and end coordinates in the reference sequence'''
return pyfastaq.intervals.Interval(min(self.ref_start, self.ref_end), max(self.ref_start, self.ref_end))
|
python
|
def ref_coords(self):
'''Returns a pyfastaq.intervals.Interval object of the start and end coordinates in the reference sequence'''
return pyfastaq.intervals.Interval(min(self.ref_start, self.ref_end), max(self.ref_start, self.ref_end))
|
[
"def",
"ref_coords",
"(",
"self",
")",
":",
"return",
"pyfastaq",
".",
"intervals",
".",
"Interval",
"(",
"min",
"(",
"self",
".",
"ref_start",
",",
"self",
".",
"ref_end",
")",
",",
"max",
"(",
"self",
".",
"ref_start",
",",
"self",
".",
"ref_end",
")",
")"
] |
Returns a pyfastaq.intervals.Interval object of the start and end coordinates in the reference sequence
|
[
"Returns",
"a",
"pyfastaq",
".",
"intervals",
".",
"Interval",
"object",
"of",
"the",
"start",
"and",
"end",
"coordinates",
"in",
"the",
"reference",
"sequence"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L66-L68
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.on_same_strand
|
def on_same_strand(self):
'''Returns true iff the direction of the alignment is the same in the reference and the query'''
return (self.ref_start < self.ref_end) == (self.qry_start < self.qry_end)
|
python
|
def on_same_strand(self):
'''Returns true iff the direction of the alignment is the same in the reference and the query'''
return (self.ref_start < self.ref_end) == (self.qry_start < self.qry_end)
|
[
"def",
"on_same_strand",
"(",
"self",
")",
":",
"return",
"(",
"self",
".",
"ref_start",
"<",
"self",
".",
"ref_end",
")",
"==",
"(",
"self",
".",
"qry_start",
"<",
"self",
".",
"qry_end",
")"
] |
Returns true iff the direction of the alignment is the same in the reference and the query
|
[
"Returns",
"true",
"iff",
"the",
"direction",
"of",
"the",
"alignment",
"is",
"the",
"same",
"in",
"the",
"reference",
"and",
"the",
"query"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L71-L73
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.is_self_hit
|
def is_self_hit(self):
'''Returns true iff the alignment is of a sequence to itself: names and all coordinates are the same and 100 percent identity'''
return self.ref_name == self.qry_name \
and self.ref_start == self.qry_start \
and self.ref_end == self.qry_end \
and self.percent_identity == 100
|
python
|
def is_self_hit(self):
'''Returns true iff the alignment is of a sequence to itself: names and all coordinates are the same and 100 percent identity'''
return self.ref_name == self.qry_name \
and self.ref_start == self.qry_start \
and self.ref_end == self.qry_end \
and self.percent_identity == 100
|
[
"def",
"is_self_hit",
"(",
"self",
")",
":",
"return",
"self",
".",
"ref_name",
"==",
"self",
".",
"qry_name",
"and",
"self",
".",
"ref_start",
"==",
"self",
".",
"qry_start",
"and",
"self",
".",
"ref_end",
"==",
"self",
".",
"qry_end",
"and",
"self",
".",
"percent_identity",
"==",
"100"
] |
Returns true iff the alignment is of a sequence to itself: names and all coordinates are the same and 100 percent identity
|
[
"Returns",
"true",
"iff",
"the",
"alignment",
"is",
"of",
"a",
"sequence",
"to",
"itself",
":",
"names",
"and",
"all",
"coordinates",
"are",
"the",
"same",
"and",
"100",
"percent",
"identity"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L76-L81
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.reverse_query
|
def reverse_query(self):
'''Changes the coordinates as if the query sequence has been reverse complemented'''
self.qry_start = self.qry_length - self.qry_start - 1
self.qry_end = self.qry_length - self.qry_end - 1
|
python
|
def reverse_query(self):
'''Changes the coordinates as if the query sequence has been reverse complemented'''
self.qry_start = self.qry_length - self.qry_start - 1
self.qry_end = self.qry_length - self.qry_end - 1
|
[
"def",
"reverse_query",
"(",
"self",
")",
":",
"self",
".",
"qry_start",
"=",
"self",
".",
"qry_length",
"-",
"self",
".",
"qry_start",
"-",
"1",
"self",
".",
"qry_end",
"=",
"self",
".",
"qry_length",
"-",
"self",
".",
"qry_end",
"-",
"1"
] |
Changes the coordinates as if the query sequence has been reverse complemented
|
[
"Changes",
"the",
"coordinates",
"as",
"if",
"the",
"query",
"sequence",
"has",
"been",
"reverse",
"complemented"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L84-L87
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.reverse_reference
|
def reverse_reference(self):
'''Changes the coordinates as if the reference sequence has been reverse complemented'''
self.ref_start = self.ref_length - self.ref_start - 1
self.ref_end = self.ref_length - self.ref_end - 1
|
python
|
def reverse_reference(self):
'''Changes the coordinates as if the reference sequence has been reverse complemented'''
self.ref_start = self.ref_length - self.ref_start - 1
self.ref_end = self.ref_length - self.ref_end - 1
|
[
"def",
"reverse_reference",
"(",
"self",
")",
":",
"self",
".",
"ref_start",
"=",
"self",
".",
"ref_length",
"-",
"self",
".",
"ref_start",
"-",
"1",
"self",
".",
"ref_end",
"=",
"self",
".",
"ref_length",
"-",
"self",
".",
"ref_end",
"-",
"1"
] |
Changes the coordinates as if the reference sequence has been reverse complemented
|
[
"Changes",
"the",
"coordinates",
"as",
"if",
"the",
"reference",
"sequence",
"has",
"been",
"reverse",
"complemented"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L90-L93
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.to_msp_crunch
|
def to_msp_crunch(self):
'''Returns the alignment as a line in MSPcrunch format. The columns are space-separated and are:
1. score
2. percent identity
3. match start in the query sequence
4. match end in the query sequence
5. query sequence name
6. subject sequence start
7. subject sequence end
8. subject sequence name'''
# we don't know the alignment score. Estimate it. This approximates 1 for a match.
aln_score = int(self.percent_identity * 0.005 * (self.hit_length_ref + self.hit_length_qry))
return ' '.join(str(x) for x in [
aln_score,
'{0:.2f}'.format(self.percent_identity),
self.qry_start + 1,
self.qry_end + 1,
self.qry_name,
self.ref_start + 1,
self.ref_end + 1,
self.ref_name
])
|
python
|
def to_msp_crunch(self):
'''Returns the alignment as a line in MSPcrunch format. The columns are space-separated and are:
1. score
2. percent identity
3. match start in the query sequence
4. match end in the query sequence
5. query sequence name
6. subject sequence start
7. subject sequence end
8. subject sequence name'''
# we don't know the alignment score. Estimate it. This approximates 1 for a match.
aln_score = int(self.percent_identity * 0.005 * (self.hit_length_ref + self.hit_length_qry))
return ' '.join(str(x) for x in [
aln_score,
'{0:.2f}'.format(self.percent_identity),
self.qry_start + 1,
self.qry_end + 1,
self.qry_name,
self.ref_start + 1,
self.ref_end + 1,
self.ref_name
])
|
[
"def",
"to_msp_crunch",
"(",
"self",
")",
":",
"# we don't know the alignment score. Estimate it. This approximates 1 for a match.",
"aln_score",
"=",
"int",
"(",
"self",
".",
"percent_identity",
"*",
"0.005",
"*",
"(",
"self",
".",
"hit_length_ref",
"+",
"self",
".",
"hit_length_qry",
")",
")",
"return",
"' '",
".",
"join",
"(",
"str",
"(",
"x",
")",
"for",
"x",
"in",
"[",
"aln_score",
",",
"'{0:.2f}'",
".",
"format",
"(",
"self",
".",
"percent_identity",
")",
",",
"self",
".",
"qry_start",
"+",
"1",
",",
"self",
".",
"qry_end",
"+",
"1",
",",
"self",
".",
"qry_name",
",",
"self",
".",
"ref_start",
"+",
"1",
",",
"self",
".",
"ref_end",
"+",
"1",
",",
"self",
".",
"ref_name",
"]",
")"
] |
Returns the alignment as a line in MSPcrunch format. The columns are space-separated and are:
1. score
2. percent identity
3. match start in the query sequence
4. match end in the query sequence
5. query sequence name
6. subject sequence start
7. subject sequence end
8. subject sequence name
|
[
"Returns",
"the",
"alignment",
"as",
"a",
"line",
"in",
"MSPcrunch",
"format",
".",
"The",
"columns",
"are",
"space",
"-",
"separated",
"and",
"are",
":",
"1",
".",
"score",
"2",
".",
"percent",
"identity",
"3",
".",
"match",
"start",
"in",
"the",
"query",
"sequence",
"4",
".",
"match",
"end",
"in",
"the",
"query",
"sequence",
"5",
".",
"query",
"sequence",
"name",
"6",
".",
"subject",
"sequence",
"start",
"7",
".",
"subject",
"sequence",
"end",
"8",
".",
"subject",
"sequence",
"name"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L113-L136
|
train
|
sanger-pathogens/pymummer
|
pymummer/alignment.py
|
Alignment.qry_coords_from_ref_coord
|
def qry_coords_from_ref_coord(self, ref_coord, variant_list):
'''Given a reference position and a list of variants ([variant.Variant]),
works out the position in the query sequence, accounting for indels.
Returns a tuple: (position, True|False), where second element is whether
or not the ref_coord lies in an indel. If it is, then
returns the corresponding start position
of the indel in the query'''
if self.ref_coords().distance_to_point(ref_coord) > 0:
raise Error('Cannot get query coord in qry_coords_from_ref_coord because given ref_coord ' + str(ref_coord) + ' does not lie in nucmer alignment:\n' + str(self))
indel_variant_indexes = []
for i in range(len(variant_list)):
if variant_list[i].var_type not in {variant.INS, variant.DEL}:
continue
if not self.intersects_variant(variant_list[i]):
continue
if variant_list[i].ref_start <= ref_coord <= variant_list[i].ref_end:
return variant_list[i].qry_start, True
elif variant_list[i].ref_start < ref_coord:
indel_variant_indexes.append(i)
distance = ref_coord - min(self.ref_start, self.ref_end)
for i in indel_variant_indexes:
if variant_list[i].var_type == variant.INS:
distance += len(variant_list[i].qry_base)
else:
assert variant_list[i].var_type == variant.DEL
distance -= len(variant_list[i].ref_base)
if self.on_same_strand():
return min(self.qry_start, self.qry_end) + distance, False
else:
return max(self.qry_start, self.qry_end) - distance, False
|
python
|
def qry_coords_from_ref_coord(self, ref_coord, variant_list):
'''Given a reference position and a list of variants ([variant.Variant]),
works out the position in the query sequence, accounting for indels.
Returns a tuple: (position, True|False), where second element is whether
or not the ref_coord lies in an indel. If it is, then
returns the corresponding start position
of the indel in the query'''
if self.ref_coords().distance_to_point(ref_coord) > 0:
raise Error('Cannot get query coord in qry_coords_from_ref_coord because given ref_coord ' + str(ref_coord) + ' does not lie in nucmer alignment:\n' + str(self))
indel_variant_indexes = []
for i in range(len(variant_list)):
if variant_list[i].var_type not in {variant.INS, variant.DEL}:
continue
if not self.intersects_variant(variant_list[i]):
continue
if variant_list[i].ref_start <= ref_coord <= variant_list[i].ref_end:
return variant_list[i].qry_start, True
elif variant_list[i].ref_start < ref_coord:
indel_variant_indexes.append(i)
distance = ref_coord - min(self.ref_start, self.ref_end)
for i in indel_variant_indexes:
if variant_list[i].var_type == variant.INS:
distance += len(variant_list[i].qry_base)
else:
assert variant_list[i].var_type == variant.DEL
distance -= len(variant_list[i].ref_base)
if self.on_same_strand():
return min(self.qry_start, self.qry_end) + distance, False
else:
return max(self.qry_start, self.qry_end) - distance, False
|
[
"def",
"qry_coords_from_ref_coord",
"(",
"self",
",",
"ref_coord",
",",
"variant_list",
")",
":",
"if",
"self",
".",
"ref_coords",
"(",
")",
".",
"distance_to_point",
"(",
"ref_coord",
")",
">",
"0",
":",
"raise",
"Error",
"(",
"'Cannot get query coord in qry_coords_from_ref_coord because given ref_coord '",
"+",
"str",
"(",
"ref_coord",
")",
"+",
"' does not lie in nucmer alignment:\\n'",
"+",
"str",
"(",
"self",
")",
")",
"indel_variant_indexes",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"variant_list",
")",
")",
":",
"if",
"variant_list",
"[",
"i",
"]",
".",
"var_type",
"not",
"in",
"{",
"variant",
".",
"INS",
",",
"variant",
".",
"DEL",
"}",
":",
"continue",
"if",
"not",
"self",
".",
"intersects_variant",
"(",
"variant_list",
"[",
"i",
"]",
")",
":",
"continue",
"if",
"variant_list",
"[",
"i",
"]",
".",
"ref_start",
"<=",
"ref_coord",
"<=",
"variant_list",
"[",
"i",
"]",
".",
"ref_end",
":",
"return",
"variant_list",
"[",
"i",
"]",
".",
"qry_start",
",",
"True",
"elif",
"variant_list",
"[",
"i",
"]",
".",
"ref_start",
"<",
"ref_coord",
":",
"indel_variant_indexes",
".",
"append",
"(",
"i",
")",
"distance",
"=",
"ref_coord",
"-",
"min",
"(",
"self",
".",
"ref_start",
",",
"self",
".",
"ref_end",
")",
"for",
"i",
"in",
"indel_variant_indexes",
":",
"if",
"variant_list",
"[",
"i",
"]",
".",
"var_type",
"==",
"variant",
".",
"INS",
":",
"distance",
"+=",
"len",
"(",
"variant_list",
"[",
"i",
"]",
".",
"qry_base",
")",
"else",
":",
"assert",
"variant_list",
"[",
"i",
"]",
".",
"var_type",
"==",
"variant",
".",
"DEL",
"distance",
"-=",
"len",
"(",
"variant_list",
"[",
"i",
"]",
".",
"ref_base",
")",
"if",
"self",
".",
"on_same_strand",
"(",
")",
":",
"return",
"min",
"(",
"self",
".",
"qry_start",
",",
"self",
".",
"qry_end",
")",
"+",
"distance",
",",
"False",
"else",
":",
"return",
"max",
"(",
"self",
".",
"qry_start",
",",
"self",
".",
"qry_end",
")",
"-",
"distance",
",",
"False"
] |
Given a reference position and a list of variants ([variant.Variant]),
works out the position in the query sequence, accounting for indels.
Returns a tuple: (position, True|False), where second element is whether
or not the ref_coord lies in an indel. If it is, then
returns the corresponding start position
of the indel in the query
|
[
"Given",
"a",
"reference",
"position",
"and",
"a",
"list",
"of",
"variants",
"(",
"[",
"variant",
".",
"Variant",
"]",
")",
"works",
"out",
"the",
"position",
"in",
"the",
"query",
"sequence",
"accounting",
"for",
"indels",
".",
"Returns",
"a",
"tuple",
":",
"(",
"position",
"True|False",
")",
"where",
"second",
"element",
"is",
"whether",
"or",
"not",
"the",
"ref_coord",
"lies",
"in",
"an",
"indel",
".",
"If",
"it",
"is",
"then",
"returns",
"the",
"corresponding",
"start",
"position",
"of",
"the",
"indel",
"in",
"the",
"query"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/alignment.py#L147-L181
|
train
|
sanger-pathogens/pymummer
|
pymummer/nucmer.py
|
Runner._nucmer_command
|
def _nucmer_command(self, ref, qry, outprefix):
'''Construct the nucmer command'''
if self.use_promer:
command = 'promer'
else:
command = 'nucmer'
command += ' -p ' + outprefix
if self.breaklen is not None:
command += ' -b ' + str(self.breaklen)
if self.diagdiff is not None and not self.use_promer:
command += ' -D ' + str(self.diagdiff)
if self.diagfactor:
command += ' -d ' + str(self.diagfactor)
if self.maxgap:
command += ' -g ' + str(self.maxgap)
if self.maxmatch:
command += ' --maxmatch'
if self.mincluster is not None:
command += ' -c ' + str(self.mincluster)
if not self.simplify and not self.use_promer:
command += ' --nosimplify'
return command + ' ' + ref + ' ' + qry
|
python
|
def _nucmer_command(self, ref, qry, outprefix):
'''Construct the nucmer command'''
if self.use_promer:
command = 'promer'
else:
command = 'nucmer'
command += ' -p ' + outprefix
if self.breaklen is not None:
command += ' -b ' + str(self.breaklen)
if self.diagdiff is not None and not self.use_promer:
command += ' -D ' + str(self.diagdiff)
if self.diagfactor:
command += ' -d ' + str(self.diagfactor)
if self.maxgap:
command += ' -g ' + str(self.maxgap)
if self.maxmatch:
command += ' --maxmatch'
if self.mincluster is not None:
command += ' -c ' + str(self.mincluster)
if not self.simplify and not self.use_promer:
command += ' --nosimplify'
return command + ' ' + ref + ' ' + qry
|
[
"def",
"_nucmer_command",
"(",
"self",
",",
"ref",
",",
"qry",
",",
"outprefix",
")",
":",
"if",
"self",
".",
"use_promer",
":",
"command",
"=",
"'promer'",
"else",
":",
"command",
"=",
"'nucmer'",
"command",
"+=",
"' -p '",
"+",
"outprefix",
"if",
"self",
".",
"breaklen",
"is",
"not",
"None",
":",
"command",
"+=",
"' -b '",
"+",
"str",
"(",
"self",
".",
"breaklen",
")",
"if",
"self",
".",
"diagdiff",
"is",
"not",
"None",
"and",
"not",
"self",
".",
"use_promer",
":",
"command",
"+=",
"' -D '",
"+",
"str",
"(",
"self",
".",
"diagdiff",
")",
"if",
"self",
".",
"diagfactor",
":",
"command",
"+=",
"' -d '",
"+",
"str",
"(",
"self",
".",
"diagfactor",
")",
"if",
"self",
".",
"maxgap",
":",
"command",
"+=",
"' -g '",
"+",
"str",
"(",
"self",
".",
"maxgap",
")",
"if",
"self",
".",
"maxmatch",
":",
"command",
"+=",
"' --maxmatch'",
"if",
"self",
".",
"mincluster",
"is",
"not",
"None",
":",
"command",
"+=",
"' -c '",
"+",
"str",
"(",
"self",
".",
"mincluster",
")",
"if",
"not",
"self",
".",
"simplify",
"and",
"not",
"self",
".",
"use_promer",
":",
"command",
"+=",
"' --nosimplify'",
"return",
"command",
"+",
"' '",
"+",
"ref",
"+",
"' '",
"+",
"qry"
] |
Construct the nucmer command
|
[
"Construct",
"the",
"nucmer",
"command"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/nucmer.py#L53-L83
|
train
|
sanger-pathogens/pymummer
|
pymummer/nucmer.py
|
Runner._delta_filter_command
|
def _delta_filter_command(self, infile, outfile):
'''Construct delta-filter command'''
command = 'delta-filter'
if self.min_id is not None:
command += ' -i ' + str(self.min_id)
if self.min_length is not None:
command += ' -l ' + str(self.min_length)
return command + ' ' + infile + ' > ' + outfile
|
python
|
def _delta_filter_command(self, infile, outfile):
'''Construct delta-filter command'''
command = 'delta-filter'
if self.min_id is not None:
command += ' -i ' + str(self.min_id)
if self.min_length is not None:
command += ' -l ' + str(self.min_length)
return command + ' ' + infile + ' > ' + outfile
|
[
"def",
"_delta_filter_command",
"(",
"self",
",",
"infile",
",",
"outfile",
")",
":",
"command",
"=",
"'delta-filter'",
"if",
"self",
".",
"min_id",
"is",
"not",
"None",
":",
"command",
"+=",
"' -i '",
"+",
"str",
"(",
"self",
".",
"min_id",
")",
"if",
"self",
".",
"min_length",
"is",
"not",
"None",
":",
"command",
"+=",
"' -l '",
"+",
"str",
"(",
"self",
".",
"min_length",
")",
"return",
"command",
"+",
"' '",
"+",
"infile",
"+",
"' > '",
"+",
"outfile"
] |
Construct delta-filter command
|
[
"Construct",
"delta",
"-",
"filter",
"command"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/nucmer.py#L86-L96
|
train
|
sanger-pathogens/pymummer
|
pymummer/nucmer.py
|
Runner._show_coords_command
|
def _show_coords_command(self, infile, outfile):
'''Construct show-coords command'''
command = 'show-coords -dTlro'
if not self.coords_header:
command += ' -H'
return command + ' ' + infile + ' > ' + outfile
|
python
|
def _show_coords_command(self, infile, outfile):
'''Construct show-coords command'''
command = 'show-coords -dTlro'
if not self.coords_header:
command += ' -H'
return command + ' ' + infile + ' > ' + outfile
|
[
"def",
"_show_coords_command",
"(",
"self",
",",
"infile",
",",
"outfile",
")",
":",
"command",
"=",
"'show-coords -dTlro'",
"if",
"not",
"self",
".",
"coords_header",
":",
"command",
"+=",
"' -H'",
"return",
"command",
"+",
"' '",
"+",
"infile",
"+",
"' > '",
"+",
"outfile"
] |
Construct show-coords command
|
[
"Construct",
"show",
"-",
"coords",
"command"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/nucmer.py#L99-L106
|
train
|
sanger-pathogens/pymummer
|
pymummer/nucmer.py
|
Runner._write_script
|
def _write_script(self, script_name, ref, qry, outfile):
'''Write commands into a bash script'''
f = pyfastaq.utils.open_file_write(script_name)
print(self._nucmer_command(ref, qry, 'p'), file=f)
print(self._delta_filter_command('p.delta', 'p.delta.filter'), file=f)
print(self._show_coords_command('p.delta.filter', outfile), file=f)
if self.show_snps:
print(self._show_snps_command('p.delta.filter', outfile + '.snps'), file=f)
pyfastaq.utils.close(f)
|
python
|
def _write_script(self, script_name, ref, qry, outfile):
'''Write commands into a bash script'''
f = pyfastaq.utils.open_file_write(script_name)
print(self._nucmer_command(ref, qry, 'p'), file=f)
print(self._delta_filter_command('p.delta', 'p.delta.filter'), file=f)
print(self._show_coords_command('p.delta.filter', outfile), file=f)
if self.show_snps:
print(self._show_snps_command('p.delta.filter', outfile + '.snps'), file=f)
pyfastaq.utils.close(f)
|
[
"def",
"_write_script",
"(",
"self",
",",
"script_name",
",",
"ref",
",",
"qry",
",",
"outfile",
")",
":",
"f",
"=",
"pyfastaq",
".",
"utils",
".",
"open_file_write",
"(",
"script_name",
")",
"print",
"(",
"self",
".",
"_nucmer_command",
"(",
"ref",
",",
"qry",
",",
"'p'",
")",
",",
"file",
"=",
"f",
")",
"print",
"(",
"self",
".",
"_delta_filter_command",
"(",
"'p.delta'",
",",
"'p.delta.filter'",
")",
",",
"file",
"=",
"f",
")",
"print",
"(",
"self",
".",
"_show_coords_command",
"(",
"'p.delta.filter'",
",",
"outfile",
")",
",",
"file",
"=",
"f",
")",
"if",
"self",
".",
"show_snps",
":",
"print",
"(",
"self",
".",
"_show_snps_command",
"(",
"'p.delta.filter'",
",",
"outfile",
"+",
"'.snps'",
")",
",",
"file",
"=",
"f",
")",
"pyfastaq",
".",
"utils",
".",
"close",
"(",
"f",
")"
] |
Write commands into a bash script
|
[
"Write",
"commands",
"into",
"a",
"bash",
"script"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/nucmer.py#L118-L126
|
train
|
sanger-pathogens/pymummer
|
pymummer/nucmer.py
|
Runner.run
|
def run(self):
'''
Change to a temp directory
Run bash script containing commands
Place results in specified output file
Clean up temp directory
'''
qry = os.path.abspath(self.qry)
ref = os.path.abspath(self.ref)
outfile = os.path.abspath(self.outfile)
tmpdir = tempfile.mkdtemp(prefix='tmp.run_nucmer.', dir=os.getcwd())
original_dir = os.getcwd()
os.chdir(tmpdir)
script = 'run_nucmer.sh'
self._write_script(script, ref, qry, outfile)
syscall.run('bash ' + script, verbose=self.verbose)
os.chdir(original_dir)
shutil.rmtree(tmpdir)
|
python
|
def run(self):
'''
Change to a temp directory
Run bash script containing commands
Place results in specified output file
Clean up temp directory
'''
qry = os.path.abspath(self.qry)
ref = os.path.abspath(self.ref)
outfile = os.path.abspath(self.outfile)
tmpdir = tempfile.mkdtemp(prefix='tmp.run_nucmer.', dir=os.getcwd())
original_dir = os.getcwd()
os.chdir(tmpdir)
script = 'run_nucmer.sh'
self._write_script(script, ref, qry, outfile)
syscall.run('bash ' + script, verbose=self.verbose)
os.chdir(original_dir)
shutil.rmtree(tmpdir)
|
[
"def",
"run",
"(",
"self",
")",
":",
"qry",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"self",
".",
"qry",
")",
"ref",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"self",
".",
"ref",
")",
"outfile",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"self",
".",
"outfile",
")",
"tmpdir",
"=",
"tempfile",
".",
"mkdtemp",
"(",
"prefix",
"=",
"'tmp.run_nucmer.'",
",",
"dir",
"=",
"os",
".",
"getcwd",
"(",
")",
")",
"original_dir",
"=",
"os",
".",
"getcwd",
"(",
")",
"os",
".",
"chdir",
"(",
"tmpdir",
")",
"script",
"=",
"'run_nucmer.sh'",
"self",
".",
"_write_script",
"(",
"script",
",",
"ref",
",",
"qry",
",",
"outfile",
")",
"syscall",
".",
"run",
"(",
"'bash '",
"+",
"script",
",",
"verbose",
"=",
"self",
".",
"verbose",
")",
"os",
".",
"chdir",
"(",
"original_dir",
")",
"shutil",
".",
"rmtree",
"(",
"tmpdir",
")"
] |
Change to a temp directory
Run bash script containing commands
Place results in specified output file
Clean up temp directory
|
[
"Change",
"to",
"a",
"temp",
"directory",
"Run",
"bash",
"script",
"containing",
"commands",
"Place",
"results",
"in",
"specified",
"output",
"file",
"Clean",
"up",
"temp",
"directory"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/nucmer.py#L129-L146
|
train
|
sanger-pathogens/pymummer
|
pymummer/variant.py
|
Variant.update_indel
|
def update_indel(self, nucmer_snp):
'''Indels are reported over multiple lines, 1 base insertion or deletion per line. This method extends the current variant by 1 base if it's an indel and adjacent to the new SNP and returns True. If the current variant is a SNP, does nothing and returns False'''
new_variant = Variant(nucmer_snp)
if self.var_type not in [INS, DEL] \
or self.var_type != new_variant.var_type \
or self.qry_name != new_variant.qry_name \
or self.ref_name != new_variant.ref_name \
or self.reverse != new_variant.reverse:
return False
if self.var_type == INS \
and self.ref_start == new_variant.ref_start \
and self.qry_end + 1 == new_variant.qry_start:
self.qry_base += new_variant.qry_base
self.qry_end += 1
return True
if self.var_type == DEL \
and self.qry_start == new_variant.qry_start \
and self.ref_end + 1 == new_variant.ref_start:
self.ref_base += new_variant.ref_base
self.ref_end += 1
return True
return False
|
python
|
def update_indel(self, nucmer_snp):
'''Indels are reported over multiple lines, 1 base insertion or deletion per line. This method extends the current variant by 1 base if it's an indel and adjacent to the new SNP and returns True. If the current variant is a SNP, does nothing and returns False'''
new_variant = Variant(nucmer_snp)
if self.var_type not in [INS, DEL] \
or self.var_type != new_variant.var_type \
or self.qry_name != new_variant.qry_name \
or self.ref_name != new_variant.ref_name \
or self.reverse != new_variant.reverse:
return False
if self.var_type == INS \
and self.ref_start == new_variant.ref_start \
and self.qry_end + 1 == new_variant.qry_start:
self.qry_base += new_variant.qry_base
self.qry_end += 1
return True
if self.var_type == DEL \
and self.qry_start == new_variant.qry_start \
and self.ref_end + 1 == new_variant.ref_start:
self.ref_base += new_variant.ref_base
self.ref_end += 1
return True
return False
|
[
"def",
"update_indel",
"(",
"self",
",",
"nucmer_snp",
")",
":",
"new_variant",
"=",
"Variant",
"(",
"nucmer_snp",
")",
"if",
"self",
".",
"var_type",
"not",
"in",
"[",
"INS",
",",
"DEL",
"]",
"or",
"self",
".",
"var_type",
"!=",
"new_variant",
".",
"var_type",
"or",
"self",
".",
"qry_name",
"!=",
"new_variant",
".",
"qry_name",
"or",
"self",
".",
"ref_name",
"!=",
"new_variant",
".",
"ref_name",
"or",
"self",
".",
"reverse",
"!=",
"new_variant",
".",
"reverse",
":",
"return",
"False",
"if",
"self",
".",
"var_type",
"==",
"INS",
"and",
"self",
".",
"ref_start",
"==",
"new_variant",
".",
"ref_start",
"and",
"self",
".",
"qry_end",
"+",
"1",
"==",
"new_variant",
".",
"qry_start",
":",
"self",
".",
"qry_base",
"+=",
"new_variant",
".",
"qry_base",
"self",
".",
"qry_end",
"+=",
"1",
"return",
"True",
"if",
"self",
".",
"var_type",
"==",
"DEL",
"and",
"self",
".",
"qry_start",
"==",
"new_variant",
".",
"qry_start",
"and",
"self",
".",
"ref_end",
"+",
"1",
"==",
"new_variant",
".",
"ref_start",
":",
"self",
".",
"ref_base",
"+=",
"new_variant",
".",
"ref_base",
"self",
".",
"ref_end",
"+=",
"1",
"return",
"True",
"return",
"False"
] |
Indels are reported over multiple lines, 1 base insertion or deletion per line. This method extends the current variant by 1 base if it's an indel and adjacent to the new SNP and returns True. If the current variant is a SNP, does nothing and returns False
|
[
"Indels",
"are",
"reported",
"over",
"multiple",
"lines",
"1",
"base",
"insertion",
"or",
"deletion",
"per",
"line",
".",
"This",
"method",
"extends",
"the",
"current",
"variant",
"by",
"1",
"base",
"if",
"it",
"s",
"an",
"indel",
"and",
"adjacent",
"to",
"the",
"new",
"SNP",
"and",
"returns",
"True",
".",
"If",
"the",
"current",
"variant",
"is",
"a",
"SNP",
"does",
"nothing",
"and",
"returns",
"False"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/variant.py#L62-L84
|
train
|
sanger-pathogens/pymummer
|
pymummer/coords_file.py
|
reader
|
def reader(fname):
'''Helper function to open the results file (coords file) and create alignment objects with the values in it'''
f = pyfastaq.utils.open_file_read(fname)
for line in f:
if line.startswith('[') or (not '\t' in line):
continue
yield alignment.Alignment(line)
pyfastaq.utils.close(f)
|
python
|
def reader(fname):
'''Helper function to open the results file (coords file) and create alignment objects with the values in it'''
f = pyfastaq.utils.open_file_read(fname)
for line in f:
if line.startswith('[') or (not '\t' in line):
continue
yield alignment.Alignment(line)
pyfastaq.utils.close(f)
|
[
"def",
"reader",
"(",
"fname",
")",
":",
"f",
"=",
"pyfastaq",
".",
"utils",
".",
"open_file_read",
"(",
"fname",
")",
"for",
"line",
"in",
"f",
":",
"if",
"line",
".",
"startswith",
"(",
"'['",
")",
"or",
"(",
"not",
"'\\t'",
"in",
"line",
")",
":",
"continue",
"yield",
"alignment",
".",
"Alignment",
"(",
"line",
")",
"pyfastaq",
".",
"utils",
".",
"close",
"(",
"f",
")"
] |
Helper function to open the results file (coords file) and create alignment objects with the values in it
|
[
"Helper",
"function",
"to",
"open",
"the",
"results",
"file",
"(",
"coords",
"file",
")",
"and",
"create",
"alignment",
"objects",
"with",
"the",
"values",
"in",
"it"
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/coords_file.py#L6-L16
|
train
|
sanger-pathogens/pymummer
|
pymummer/coords_file.py
|
convert_to_msp_crunch
|
def convert_to_msp_crunch(infile, outfile, ref_fai=None, qry_fai=None):
'''Converts a coords file to a file in MSPcrunch format (for use with ACT, most likely).
ACT ignores sequence names in the crunch file, and just looks at the numbers.
To make a compatible file, the coords all must be shifted appropriately, which
can be done by providing both the ref_fai and qry_fai options.
Both or neither of these must be used, otherwise an error will be thrown.'''
fai_files = {ref_fai, qry_fai}
if None in fai_files and len(fai_files) != 1:
print(fai_files)
raise Error('Error in convert_to_msp_crunch. Must use both of ref_fai and qry_fai, or neither of them')
if ref_fai is not None:
assert qry_fai is not None
ref_offsets = pyfastaq.tasks.length_offsets_from_fai(ref_fai)
qry_offsets = pyfastaq.tasks.length_offsets_from_fai(qry_fai)
file_reader = reader(infile)
f_out = pyfastaq.utils.open_file_write(outfile)
for aln in file_reader:
if ref_fai is not None:
aln.ref_start += ref_offsets[aln.ref_name]
aln.ref_end += ref_offsets[aln.ref_name]
aln.qry_start += qry_offsets[aln.qry_name]
aln.qry_end += qry_offsets[aln.qry_name]
print(aln.to_msp_crunch(), file=f_out)
pyfastaq.utils.close(f_out)
|
python
|
def convert_to_msp_crunch(infile, outfile, ref_fai=None, qry_fai=None):
'''Converts a coords file to a file in MSPcrunch format (for use with ACT, most likely).
ACT ignores sequence names in the crunch file, and just looks at the numbers.
To make a compatible file, the coords all must be shifted appropriately, which
can be done by providing both the ref_fai and qry_fai options.
Both or neither of these must be used, otherwise an error will be thrown.'''
fai_files = {ref_fai, qry_fai}
if None in fai_files and len(fai_files) != 1:
print(fai_files)
raise Error('Error in convert_to_msp_crunch. Must use both of ref_fai and qry_fai, or neither of them')
if ref_fai is not None:
assert qry_fai is not None
ref_offsets = pyfastaq.tasks.length_offsets_from_fai(ref_fai)
qry_offsets = pyfastaq.tasks.length_offsets_from_fai(qry_fai)
file_reader = reader(infile)
f_out = pyfastaq.utils.open_file_write(outfile)
for aln in file_reader:
if ref_fai is not None:
aln.ref_start += ref_offsets[aln.ref_name]
aln.ref_end += ref_offsets[aln.ref_name]
aln.qry_start += qry_offsets[aln.qry_name]
aln.qry_end += qry_offsets[aln.qry_name]
print(aln.to_msp_crunch(), file=f_out)
pyfastaq.utils.close(f_out)
|
[
"def",
"convert_to_msp_crunch",
"(",
"infile",
",",
"outfile",
",",
"ref_fai",
"=",
"None",
",",
"qry_fai",
"=",
"None",
")",
":",
"fai_files",
"=",
"{",
"ref_fai",
",",
"qry_fai",
"}",
"if",
"None",
"in",
"fai_files",
"and",
"len",
"(",
"fai_files",
")",
"!=",
"1",
":",
"print",
"(",
"fai_files",
")",
"raise",
"Error",
"(",
"'Error in convert_to_msp_crunch. Must use both of ref_fai and qry_fai, or neither of them'",
")",
"if",
"ref_fai",
"is",
"not",
"None",
":",
"assert",
"qry_fai",
"is",
"not",
"None",
"ref_offsets",
"=",
"pyfastaq",
".",
"tasks",
".",
"length_offsets_from_fai",
"(",
"ref_fai",
")",
"qry_offsets",
"=",
"pyfastaq",
".",
"tasks",
".",
"length_offsets_from_fai",
"(",
"qry_fai",
")",
"file_reader",
"=",
"reader",
"(",
"infile",
")",
"f_out",
"=",
"pyfastaq",
".",
"utils",
".",
"open_file_write",
"(",
"outfile",
")",
"for",
"aln",
"in",
"file_reader",
":",
"if",
"ref_fai",
"is",
"not",
"None",
":",
"aln",
".",
"ref_start",
"+=",
"ref_offsets",
"[",
"aln",
".",
"ref_name",
"]",
"aln",
".",
"ref_end",
"+=",
"ref_offsets",
"[",
"aln",
".",
"ref_name",
"]",
"aln",
".",
"qry_start",
"+=",
"qry_offsets",
"[",
"aln",
".",
"qry_name",
"]",
"aln",
".",
"qry_end",
"+=",
"qry_offsets",
"[",
"aln",
".",
"qry_name",
"]",
"print",
"(",
"aln",
".",
"to_msp_crunch",
"(",
")",
",",
"file",
"=",
"f_out",
")",
"pyfastaq",
".",
"utils",
".",
"close",
"(",
"f_out",
")"
] |
Converts a coords file to a file in MSPcrunch format (for use with ACT, most likely).
ACT ignores sequence names in the crunch file, and just looks at the numbers.
To make a compatible file, the coords all must be shifted appropriately, which
can be done by providing both the ref_fai and qry_fai options.
Both or neither of these must be used, otherwise an error will be thrown.
|
[
"Converts",
"a",
"coords",
"file",
"to",
"a",
"file",
"in",
"MSPcrunch",
"format",
"(",
"for",
"use",
"with",
"ACT",
"most",
"likely",
")",
".",
"ACT",
"ignores",
"sequence",
"names",
"in",
"the",
"crunch",
"file",
"and",
"just",
"looks",
"at",
"the",
"numbers",
".",
"To",
"make",
"a",
"compatible",
"file",
"the",
"coords",
"all",
"must",
"be",
"shifted",
"appropriately",
"which",
"can",
"be",
"done",
"by",
"providing",
"both",
"the",
"ref_fai",
"and",
"qry_fai",
"options",
".",
"Both",
"or",
"neither",
"of",
"these",
"must",
"be",
"used",
"otherwise",
"an",
"error",
"will",
"be",
"thrown",
"."
] |
fd97bccfbae62719a7247473d73dd6733d4fa903
|
https://github.com/sanger-pathogens/pymummer/blob/fd97bccfbae62719a7247473d73dd6733d4fa903/pymummer/coords_file.py#L19-L47
|
train
|
anandtiwarics/python-burp-rest-api
|
PyBurprestapi/burpscanner.py
|
BurpApi._request
|
def _request(self, method, url, params=None, headers=None, data=None):
"""Common handler for all the HTTP requests."""
if not params:
params = {}
# set default headers
if not headers:
headers = {
'accept': '*/*'
}
if method == 'POST' or method == 'PUT':
headers.update({'Content-Type': 'application/json'})
try:
response = requests.request(method=method, url=self.host + self.key + url, params=params,
headers=headers, data=data)
try:
response.raise_for_status()
response_code = response.status_code
success = True if response_code // 100 == 2 else False
if response.text:
try:
data = response.json()
except ValueError:
data = response.content
else:
data = ''
response_headers = response.headers
return BurpResponse(success=success, response_code=response_code, data=data,
response_headers=response_headers)
except ValueError as e:
return BurpResponse(success=False, message="JSON response could not be decoded {}.".format(e))
except requests.exceptions.HTTPError as e:
if response.status_code == 400:
return BurpResponse(success=False, response_code=400, message='Bad Request')
else:
return BurpResponse(
message='There was an error while handling the request. {}'.format(response.content),
success=False)
except Exception as e:
return BurpResponse(success=False, message='Eerror is %s' % e)
|
python
|
def _request(self, method, url, params=None, headers=None, data=None):
"""Common handler for all the HTTP requests."""
if not params:
params = {}
# set default headers
if not headers:
headers = {
'accept': '*/*'
}
if method == 'POST' or method == 'PUT':
headers.update({'Content-Type': 'application/json'})
try:
response = requests.request(method=method, url=self.host + self.key + url, params=params,
headers=headers, data=data)
try:
response.raise_for_status()
response_code = response.status_code
success = True if response_code // 100 == 2 else False
if response.text:
try:
data = response.json()
except ValueError:
data = response.content
else:
data = ''
response_headers = response.headers
return BurpResponse(success=success, response_code=response_code, data=data,
response_headers=response_headers)
except ValueError as e:
return BurpResponse(success=False, message="JSON response could not be decoded {}.".format(e))
except requests.exceptions.HTTPError as e:
if response.status_code == 400:
return BurpResponse(success=False, response_code=400, message='Bad Request')
else:
return BurpResponse(
message='There was an error while handling the request. {}'.format(response.content),
success=False)
except Exception as e:
return BurpResponse(success=False, message='Eerror is %s' % e)
|
[
"def",
"_request",
"(",
"self",
",",
"method",
",",
"url",
",",
"params",
"=",
"None",
",",
"headers",
"=",
"None",
",",
"data",
"=",
"None",
")",
":",
"if",
"not",
"params",
":",
"params",
"=",
"{",
"}",
"# set default headers",
"if",
"not",
"headers",
":",
"headers",
"=",
"{",
"'accept'",
":",
"'*/*'",
"}",
"if",
"method",
"==",
"'POST'",
"or",
"method",
"==",
"'PUT'",
":",
"headers",
".",
"update",
"(",
"{",
"'Content-Type'",
":",
"'application/json'",
"}",
")",
"try",
":",
"response",
"=",
"requests",
".",
"request",
"(",
"method",
"=",
"method",
",",
"url",
"=",
"self",
".",
"host",
"+",
"self",
".",
"key",
"+",
"url",
",",
"params",
"=",
"params",
",",
"headers",
"=",
"headers",
",",
"data",
"=",
"data",
")",
"try",
":",
"response",
".",
"raise_for_status",
"(",
")",
"response_code",
"=",
"response",
".",
"status_code",
"success",
"=",
"True",
"if",
"response_code",
"//",
"100",
"==",
"2",
"else",
"False",
"if",
"response",
".",
"text",
":",
"try",
":",
"data",
"=",
"response",
".",
"json",
"(",
")",
"except",
"ValueError",
":",
"data",
"=",
"response",
".",
"content",
"else",
":",
"data",
"=",
"''",
"response_headers",
"=",
"response",
".",
"headers",
"return",
"BurpResponse",
"(",
"success",
"=",
"success",
",",
"response_code",
"=",
"response_code",
",",
"data",
"=",
"data",
",",
"response_headers",
"=",
"response_headers",
")",
"except",
"ValueError",
"as",
"e",
":",
"return",
"BurpResponse",
"(",
"success",
"=",
"False",
",",
"message",
"=",
"\"JSON response could not be decoded {}.\"",
".",
"format",
"(",
"e",
")",
")",
"except",
"requests",
".",
"exceptions",
".",
"HTTPError",
"as",
"e",
":",
"if",
"response",
".",
"status_code",
"==",
"400",
":",
"return",
"BurpResponse",
"(",
"success",
"=",
"False",
",",
"response_code",
"=",
"400",
",",
"message",
"=",
"'Bad Request'",
")",
"else",
":",
"return",
"BurpResponse",
"(",
"message",
"=",
"'There was an error while handling the request. {}'",
".",
"format",
"(",
"response",
".",
"content",
")",
",",
"success",
"=",
"False",
")",
"except",
"Exception",
"as",
"e",
":",
"return",
"BurpResponse",
"(",
"success",
"=",
"False",
",",
"message",
"=",
"'Eerror is %s'",
"%",
"e",
")"
] |
Common handler for all the HTTP requests.
|
[
"Common",
"handler",
"for",
"all",
"the",
"HTTP",
"requests",
"."
] |
1b3fb263ef1d006e181b838ea017ac2abbd68a30
|
https://github.com/anandtiwarics/python-burp-rest-api/blob/1b3fb263ef1d006e181b838ea017ac2abbd68a30/PyBurprestapi/burpscanner.py#L55-L98
|
train
|
sloria/sphinx-issues
|
sphinx_issues.py
|
user_role
|
def user_role(name, rawtext, text, lineno, inliner, options=None, content=None):
"""Sphinx role for linking to a user profile. Defaults to linking to
Github profiles, but the profile URIS can be configured via the
``issues_user_uri`` config value.
Examples: ::
:user:`sloria`
Anchor text also works: ::
:user:`Steven Loria <sloria>`
"""
options = options or {}
content = content or []
has_explicit_title, title, target = split_explicit_title(text)
target = utils.unescape(target).strip()
title = utils.unescape(title).strip()
config = inliner.document.settings.env.app.config
if config.issues_user_uri:
ref = config.issues_user_uri.format(user=target)
else:
ref = "https://github.com/{0}".format(target)
if has_explicit_title:
text = title
else:
text = "@{0}".format(target)
link = nodes.reference(text=text, refuri=ref, **options)
return [link], []
|
python
|
def user_role(name, rawtext, text, lineno, inliner, options=None, content=None):
"""Sphinx role for linking to a user profile. Defaults to linking to
Github profiles, but the profile URIS can be configured via the
``issues_user_uri`` config value.
Examples: ::
:user:`sloria`
Anchor text also works: ::
:user:`Steven Loria <sloria>`
"""
options = options or {}
content = content or []
has_explicit_title, title, target = split_explicit_title(text)
target = utils.unescape(target).strip()
title = utils.unescape(title).strip()
config = inliner.document.settings.env.app.config
if config.issues_user_uri:
ref = config.issues_user_uri.format(user=target)
else:
ref = "https://github.com/{0}".format(target)
if has_explicit_title:
text = title
else:
text = "@{0}".format(target)
link = nodes.reference(text=text, refuri=ref, **options)
return [link], []
|
[
"def",
"user_role",
"(",
"name",
",",
"rawtext",
",",
"text",
",",
"lineno",
",",
"inliner",
",",
"options",
"=",
"None",
",",
"content",
"=",
"None",
")",
":",
"options",
"=",
"options",
"or",
"{",
"}",
"content",
"=",
"content",
"or",
"[",
"]",
"has_explicit_title",
",",
"title",
",",
"target",
"=",
"split_explicit_title",
"(",
"text",
")",
"target",
"=",
"utils",
".",
"unescape",
"(",
"target",
")",
".",
"strip",
"(",
")",
"title",
"=",
"utils",
".",
"unescape",
"(",
"title",
")",
".",
"strip",
"(",
")",
"config",
"=",
"inliner",
".",
"document",
".",
"settings",
".",
"env",
".",
"app",
".",
"config",
"if",
"config",
".",
"issues_user_uri",
":",
"ref",
"=",
"config",
".",
"issues_user_uri",
".",
"format",
"(",
"user",
"=",
"target",
")",
"else",
":",
"ref",
"=",
"\"https://github.com/{0}\"",
".",
"format",
"(",
"target",
")",
"if",
"has_explicit_title",
":",
"text",
"=",
"title",
"else",
":",
"text",
"=",
"\"@{0}\"",
".",
"format",
"(",
"target",
")",
"link",
"=",
"nodes",
".",
"reference",
"(",
"text",
"=",
"text",
",",
"refuri",
"=",
"ref",
",",
"*",
"*",
"options",
")",
"return",
"[",
"link",
"]",
",",
"[",
"]"
] |
Sphinx role for linking to a user profile. Defaults to linking to
Github profiles, but the profile URIS can be configured via the
``issues_user_uri`` config value.
Examples: ::
:user:`sloria`
Anchor text also works: ::
:user:`Steven Loria <sloria>`
|
[
"Sphinx",
"role",
"for",
"linking",
"to",
"a",
"user",
"profile",
".",
"Defaults",
"to",
"linking",
"to",
"Github",
"profiles",
"but",
"the",
"profile",
"URIS",
"can",
"be",
"configured",
"via",
"the",
"issues_user_uri",
"config",
"value",
"."
] |
0a9597472645dc728c2aef12e0653aabfdb68ab2
|
https://github.com/sloria/sphinx-issues/blob/0a9597472645dc728c2aef12e0653aabfdb68ab2/sphinx_issues.py#L13-L43
|
train
|
sloria/sphinx-issues
|
sphinx_issues.py
|
cve_role
|
def cve_role(name, rawtext, text, lineno, inliner, options=None, content=None):
"""Sphinx role for linking to a CVE on https://cve.mitre.org.
Examples: ::
:cve:`CVE-2018-17175`
"""
options = options or {}
content = content or []
has_explicit_title, title, target = split_explicit_title(text)
target = utils.unescape(target).strip()
title = utils.unescape(title).strip()
ref = "https://cve.mitre.org/cgi-bin/cvename.cgi?name={0}".format(target)
text = title if has_explicit_title else target
link = nodes.reference(text=text, refuri=ref, **options)
return [link], []
|
python
|
def cve_role(name, rawtext, text, lineno, inliner, options=None, content=None):
"""Sphinx role for linking to a CVE on https://cve.mitre.org.
Examples: ::
:cve:`CVE-2018-17175`
"""
options = options or {}
content = content or []
has_explicit_title, title, target = split_explicit_title(text)
target = utils.unescape(target).strip()
title = utils.unescape(title).strip()
ref = "https://cve.mitre.org/cgi-bin/cvename.cgi?name={0}".format(target)
text = title if has_explicit_title else target
link = nodes.reference(text=text, refuri=ref, **options)
return [link], []
|
[
"def",
"cve_role",
"(",
"name",
",",
"rawtext",
",",
"text",
",",
"lineno",
",",
"inliner",
",",
"options",
"=",
"None",
",",
"content",
"=",
"None",
")",
":",
"options",
"=",
"options",
"or",
"{",
"}",
"content",
"=",
"content",
"or",
"[",
"]",
"has_explicit_title",
",",
"title",
",",
"target",
"=",
"split_explicit_title",
"(",
"text",
")",
"target",
"=",
"utils",
".",
"unescape",
"(",
"target",
")",
".",
"strip",
"(",
")",
"title",
"=",
"utils",
".",
"unescape",
"(",
"title",
")",
".",
"strip",
"(",
")",
"ref",
"=",
"\"https://cve.mitre.org/cgi-bin/cvename.cgi?name={0}\"",
".",
"format",
"(",
"target",
")",
"text",
"=",
"title",
"if",
"has_explicit_title",
"else",
"target",
"link",
"=",
"nodes",
".",
"reference",
"(",
"text",
"=",
"text",
",",
"refuri",
"=",
"ref",
",",
"*",
"*",
"options",
")",
"return",
"[",
"link",
"]",
",",
"[",
"]"
] |
Sphinx role for linking to a CVE on https://cve.mitre.org.
Examples: ::
:cve:`CVE-2018-17175`
|
[
"Sphinx",
"role",
"for",
"linking",
"to",
"a",
"CVE",
"on",
"https",
":",
"//",
"cve",
".",
"mitre",
".",
"org",
"."
] |
0a9597472645dc728c2aef12e0653aabfdb68ab2
|
https://github.com/sloria/sphinx-issues/blob/0a9597472645dc728c2aef12e0653aabfdb68ab2/sphinx_issues.py#L46-L63
|
train
|
adamnovak/tsv
|
tsv.py
|
TsvWriter.list_line
|
def list_line(self, line):
"""
Write the given iterable of values (line) to the file as items on the
same line. Any argument that stringifies to a string legal as a TSV data
item can be written.
Does not copy the line or build a big string in memory.
"""
if len(line) == 0:
return
self.stream.write(str(line[0]))
for item in line[1:]:
self.stream.write("\t")
self.stream.write(str(item))
self.stream.write("\n")
|
python
|
def list_line(self, line):
"""
Write the given iterable of values (line) to the file as items on the
same line. Any argument that stringifies to a string legal as a TSV data
item can be written.
Does not copy the line or build a big string in memory.
"""
if len(line) == 0:
return
self.stream.write(str(line[0]))
for item in line[1:]:
self.stream.write("\t")
self.stream.write(str(item))
self.stream.write("\n")
|
[
"def",
"list_line",
"(",
"self",
",",
"line",
")",
":",
"if",
"len",
"(",
"line",
")",
"==",
"0",
":",
"return",
"self",
".",
"stream",
".",
"write",
"(",
"str",
"(",
"line",
"[",
"0",
"]",
")",
")",
"for",
"item",
"in",
"line",
"[",
"1",
":",
"]",
":",
"self",
".",
"stream",
".",
"write",
"(",
"\"\\t\"",
")",
"self",
".",
"stream",
".",
"write",
"(",
"str",
"(",
"item",
")",
")",
"self",
".",
"stream",
".",
"write",
"(",
"\"\\n\"",
")"
] |
Write the given iterable of values (line) to the file as items on the
same line. Any argument that stringifies to a string legal as a TSV data
item can be written.
Does not copy the line or build a big string in memory.
|
[
"Write",
"the",
"given",
"iterable",
"of",
"values",
"(",
"line",
")",
"to",
"the",
"file",
"as",
"items",
"on",
"the",
"same",
"line",
".",
"Any",
"argument",
"that",
"stringifies",
"to",
"a",
"string",
"legal",
"as",
"a",
"TSV",
"data",
"item",
"can",
"be",
"written",
".",
"Does",
"not",
"copy",
"the",
"line",
"or",
"build",
"a",
"big",
"string",
"in",
"memory",
"."
] |
379189e9da4c1b65d0587bb32f3b51e6a7c936c8
|
https://github.com/adamnovak/tsv/blob/379189e9da4c1b65d0587bb32f3b51e6a7c936c8/tsv.py#L42-L60
|
train
|
michaelstepner/pandoc-mustache
|
pandoc_mustache/pandoc_mustache.py
|
prepare
|
def prepare(doc):
""" Parse metadata to obtain list of mustache templates,
then load those templates.
"""
doc.mustache_files = doc.get_metadata('mustache')
if isinstance(doc.mustache_files, basestring): # process single YAML value stored as string
if not doc.mustache_files:
doc.mustache_files = None # switch empty string back to None
else:
doc.mustache_files = [ doc.mustache_files ] # put non-empty string in list
# with open('debug.txt', 'a') as the_file:
# the_file.write(str(doc.mustache_files))
# the_file.write('\n')
if doc.mustache_files is not None:
doc.mustache_hashes = [yaml.load(open(file, 'r').read()) for file in doc.mustache_files]
doc.mhash = { k: v for mdict in doc.mustache_hashes for k, v in mdict.items() } # combine list of dicts into a single dict
doc.mrenderer = pystache.Renderer(escape=lambda u: u, missing_tags='strict')
else:
doc.mhash = None
|
python
|
def prepare(doc):
""" Parse metadata to obtain list of mustache templates,
then load those templates.
"""
doc.mustache_files = doc.get_metadata('mustache')
if isinstance(doc.mustache_files, basestring): # process single YAML value stored as string
if not doc.mustache_files:
doc.mustache_files = None # switch empty string back to None
else:
doc.mustache_files = [ doc.mustache_files ] # put non-empty string in list
# with open('debug.txt', 'a') as the_file:
# the_file.write(str(doc.mustache_files))
# the_file.write('\n')
if doc.mustache_files is not None:
doc.mustache_hashes = [yaml.load(open(file, 'r').read()) for file in doc.mustache_files]
doc.mhash = { k: v for mdict in doc.mustache_hashes for k, v in mdict.items() } # combine list of dicts into a single dict
doc.mrenderer = pystache.Renderer(escape=lambda u: u, missing_tags='strict')
else:
doc.mhash = None
|
[
"def",
"prepare",
"(",
"doc",
")",
":",
"doc",
".",
"mustache_files",
"=",
"doc",
".",
"get_metadata",
"(",
"'mustache'",
")",
"if",
"isinstance",
"(",
"doc",
".",
"mustache_files",
",",
"basestring",
")",
":",
"# process single YAML value stored as string",
"if",
"not",
"doc",
".",
"mustache_files",
":",
"doc",
".",
"mustache_files",
"=",
"None",
"# switch empty string back to None",
"else",
":",
"doc",
".",
"mustache_files",
"=",
"[",
"doc",
".",
"mustache_files",
"]",
"# put non-empty string in list",
"# with open('debug.txt', 'a') as the_file:",
"# the_file.write(str(doc.mustache_files))",
"# the_file.write('\\n')",
"if",
"doc",
".",
"mustache_files",
"is",
"not",
"None",
":",
"doc",
".",
"mustache_hashes",
"=",
"[",
"yaml",
".",
"load",
"(",
"open",
"(",
"file",
",",
"'r'",
")",
".",
"read",
"(",
")",
")",
"for",
"file",
"in",
"doc",
".",
"mustache_files",
"]",
"doc",
".",
"mhash",
"=",
"{",
"k",
":",
"v",
"for",
"mdict",
"in",
"doc",
".",
"mustache_hashes",
"for",
"k",
",",
"v",
"in",
"mdict",
".",
"items",
"(",
")",
"}",
"# combine list of dicts into a single dict",
"doc",
".",
"mrenderer",
"=",
"pystache",
".",
"Renderer",
"(",
"escape",
"=",
"lambda",
"u",
":",
"u",
",",
"missing_tags",
"=",
"'strict'",
")",
"else",
":",
"doc",
".",
"mhash",
"=",
"None"
] |
Parse metadata to obtain list of mustache templates,
then load those templates.
|
[
"Parse",
"metadata",
"to",
"obtain",
"list",
"of",
"mustache",
"templates",
"then",
"load",
"those",
"templates",
"."
] |
52d71190d134964596e7f8e109564452e916b3fc
|
https://github.com/michaelstepner/pandoc-mustache/blob/52d71190d134964596e7f8e109564452e916b3fc/pandoc_mustache/pandoc_mustache.py#L8-L26
|
train
|
michaelstepner/pandoc-mustache
|
pandoc_mustache/pandoc_mustache.py
|
action
|
def action(elem, doc):
""" Apply combined mustache template to all strings in document.
"""
if type(elem) == Str and doc.mhash is not None:
elem.text = doc.mrenderer.render(elem.text, doc.mhash)
return elem
|
python
|
def action(elem, doc):
""" Apply combined mustache template to all strings in document.
"""
if type(elem) == Str and doc.mhash is not None:
elem.text = doc.mrenderer.render(elem.text, doc.mhash)
return elem
|
[
"def",
"action",
"(",
"elem",
",",
"doc",
")",
":",
"if",
"type",
"(",
"elem",
")",
"==",
"Str",
"and",
"doc",
".",
"mhash",
"is",
"not",
"None",
":",
"elem",
".",
"text",
"=",
"doc",
".",
"mrenderer",
".",
"render",
"(",
"elem",
".",
"text",
",",
"doc",
".",
"mhash",
")",
"return",
"elem"
] |
Apply combined mustache template to all strings in document.
|
[
"Apply",
"combined",
"mustache",
"template",
"to",
"all",
"strings",
"in",
"document",
"."
] |
52d71190d134964596e7f8e109564452e916b3fc
|
https://github.com/michaelstepner/pandoc-mustache/blob/52d71190d134964596e7f8e109564452e916b3fc/pandoc_mustache/pandoc_mustache.py#L28-L33
|
train
|
jpadilla/django-rest-framework-jsonp
|
rest_framework_jsonp/renderers.py
|
JSONPRenderer.get_callback
|
def get_callback(self, renderer_context):
"""
Determine the name of the callback to wrap around the json output.
"""
request = renderer_context.get('request', None)
params = request and get_query_params(request) or {}
return params.get(self.callback_parameter, self.default_callback)
|
python
|
def get_callback(self, renderer_context):
"""
Determine the name of the callback to wrap around the json output.
"""
request = renderer_context.get('request', None)
params = request and get_query_params(request) or {}
return params.get(self.callback_parameter, self.default_callback)
|
[
"def",
"get_callback",
"(",
"self",
",",
"renderer_context",
")",
":",
"request",
"=",
"renderer_context",
".",
"get",
"(",
"'request'",
",",
"None",
")",
"params",
"=",
"request",
"and",
"get_query_params",
"(",
"request",
")",
"or",
"{",
"}",
"return",
"params",
".",
"get",
"(",
"self",
".",
"callback_parameter",
",",
"self",
".",
"default_callback",
")"
] |
Determine the name of the callback to wrap around the json output.
|
[
"Determine",
"the",
"name",
"of",
"the",
"callback",
"to",
"wrap",
"around",
"the",
"json",
"output",
"."
] |
64e91d451206a815b98d4b4a4e27e705104ad06e
|
https://github.com/jpadilla/django-rest-framework-jsonp/blob/64e91d451206a815b98d4b4a4e27e705104ad06e/rest_framework_jsonp/renderers.py#L23-L29
|
train
|
jpadilla/django-rest-framework-jsonp
|
rest_framework_jsonp/renderers.py
|
JSONPRenderer.render
|
def render(self, data, accepted_media_type=None, renderer_context=None):
"""
Renders into jsonp, wrapping the json output in a callback function.
Clients may set the callback function name using a query parameter
on the URL, for example: ?callback=exampleCallbackName
"""
renderer_context = renderer_context or {}
callback = self.get_callback(renderer_context)
json = super(JSONPRenderer, self).render(data, accepted_media_type,
renderer_context)
return callback.encode(self.charset) + b'(' + json + b');'
|
python
|
def render(self, data, accepted_media_type=None, renderer_context=None):
"""
Renders into jsonp, wrapping the json output in a callback function.
Clients may set the callback function name using a query parameter
on the URL, for example: ?callback=exampleCallbackName
"""
renderer_context = renderer_context or {}
callback = self.get_callback(renderer_context)
json = super(JSONPRenderer, self).render(data, accepted_media_type,
renderer_context)
return callback.encode(self.charset) + b'(' + json + b');'
|
[
"def",
"render",
"(",
"self",
",",
"data",
",",
"accepted_media_type",
"=",
"None",
",",
"renderer_context",
"=",
"None",
")",
":",
"renderer_context",
"=",
"renderer_context",
"or",
"{",
"}",
"callback",
"=",
"self",
".",
"get_callback",
"(",
"renderer_context",
")",
"json",
"=",
"super",
"(",
"JSONPRenderer",
",",
"self",
")",
".",
"render",
"(",
"data",
",",
"accepted_media_type",
",",
"renderer_context",
")",
"return",
"callback",
".",
"encode",
"(",
"self",
".",
"charset",
")",
"+",
"b'('",
"+",
"json",
"+",
"b');'"
] |
Renders into jsonp, wrapping the json output in a callback function.
Clients may set the callback function name using a query parameter
on the URL, for example: ?callback=exampleCallbackName
|
[
"Renders",
"into",
"jsonp",
"wrapping",
"the",
"json",
"output",
"in",
"a",
"callback",
"function",
"."
] |
64e91d451206a815b98d4b4a4e27e705104ad06e
|
https://github.com/jpadilla/django-rest-framework-jsonp/blob/64e91d451206a815b98d4b4a4e27e705104ad06e/rest_framework_jsonp/renderers.py#L31-L42
|
train
|
3ll3d00d/vibe
|
backend/src/analyser/resources/timeseries.py
|
TimeSeries.get
|
def get(self, measurementId):
"""
Analyses the measurement with the given parameters
:param measurementId:
:return:
"""
logger.info('Loading raw data for ' + measurementId)
measurement = self._measurementController.getMeasurement(measurementId, MeasurementStatus.COMPLETE)
if measurement is not None:
if measurement.inflate():
data = {
name: {
'raw': {
'x': self._jsonify(data.raw('x')),
'y': self._jsonify(data.raw('y')),
'z': self._jsonify(data.raw('z'))
},
'vibration': {
'x': self._jsonify(data.vibration('x')),
'y': self._jsonify(data.vibration('y')),
'z': self._jsonify(data.vibration('z'))
},
'tilt': {
'x': self._jsonify(data.tilt('x')),
'y': self._jsonify(data.tilt('y')),
'z': self._jsonify(data.tilt('z'))
}
}
for name, data in measurement.data.items()
}
return data, 200
else:
return None, 404
else:
return None, 404
|
python
|
def get(self, measurementId):
"""
Analyses the measurement with the given parameters
:param measurementId:
:return:
"""
logger.info('Loading raw data for ' + measurementId)
measurement = self._measurementController.getMeasurement(measurementId, MeasurementStatus.COMPLETE)
if measurement is not None:
if measurement.inflate():
data = {
name: {
'raw': {
'x': self._jsonify(data.raw('x')),
'y': self._jsonify(data.raw('y')),
'z': self._jsonify(data.raw('z'))
},
'vibration': {
'x': self._jsonify(data.vibration('x')),
'y': self._jsonify(data.vibration('y')),
'z': self._jsonify(data.vibration('z'))
},
'tilt': {
'x': self._jsonify(data.tilt('x')),
'y': self._jsonify(data.tilt('y')),
'z': self._jsonify(data.tilt('z'))
}
}
for name, data in measurement.data.items()
}
return data, 200
else:
return None, 404
else:
return None, 404
|
[
"def",
"get",
"(",
"self",
",",
"measurementId",
")",
":",
"logger",
".",
"info",
"(",
"'Loading raw data for '",
"+",
"measurementId",
")",
"measurement",
"=",
"self",
".",
"_measurementController",
".",
"getMeasurement",
"(",
"measurementId",
",",
"MeasurementStatus",
".",
"COMPLETE",
")",
"if",
"measurement",
"is",
"not",
"None",
":",
"if",
"measurement",
".",
"inflate",
"(",
")",
":",
"data",
"=",
"{",
"name",
":",
"{",
"'raw'",
":",
"{",
"'x'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"raw",
"(",
"'x'",
")",
")",
",",
"'y'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"raw",
"(",
"'y'",
")",
")",
",",
"'z'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"raw",
"(",
"'z'",
")",
")",
"}",
",",
"'vibration'",
":",
"{",
"'x'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"vibration",
"(",
"'x'",
")",
")",
",",
"'y'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"vibration",
"(",
"'y'",
")",
")",
",",
"'z'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"vibration",
"(",
"'z'",
")",
")",
"}",
",",
"'tilt'",
":",
"{",
"'x'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"tilt",
"(",
"'x'",
")",
")",
",",
"'y'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"tilt",
"(",
"'y'",
")",
")",
",",
"'z'",
":",
"self",
".",
"_jsonify",
"(",
"data",
".",
"tilt",
"(",
"'z'",
")",
")",
"}",
"}",
"for",
"name",
",",
"data",
"in",
"measurement",
".",
"data",
".",
"items",
"(",
")",
"}",
"return",
"data",
",",
"200",
"else",
":",
"return",
"None",
",",
"404",
"else",
":",
"return",
"None",
",",
"404"
] |
Analyses the measurement with the given parameters
:param measurementId:
:return:
|
[
"Analyses",
"the",
"measurement",
"with",
"the",
"given",
"parameters",
":",
"param",
"measurementId",
":",
":",
"return",
":"
] |
124b029f13ac746723e92cb47e9cb56edd2e54b5
|
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/resources/timeseries.py#L13-L47
|
train
|
fitnr/convertdate
|
convertdate/iso.py
|
to_jd
|
def to_jd(year, week, day):
'''Return Julian day count of given ISO year, week, and day'''
return day + n_weeks(SUN, gregorian.to_jd(year - 1, 12, 28), week)
|
python
|
def to_jd(year, week, day):
'''Return Julian day count of given ISO year, week, and day'''
return day + n_weeks(SUN, gregorian.to_jd(year - 1, 12, 28), week)
|
[
"def",
"to_jd",
"(",
"year",
",",
"week",
",",
"day",
")",
":",
"return",
"day",
"+",
"n_weeks",
"(",
"SUN",
",",
"gregorian",
".",
"to_jd",
"(",
"year",
"-",
"1",
",",
"12",
",",
"28",
")",
",",
"week",
")"
] |
Return Julian day count of given ISO year, week, and day
|
[
"Return",
"Julian",
"day",
"count",
"of",
"given",
"ISO",
"year",
"week",
"and",
"day"
] |
e920f168a87f99183b0aa7290d6c3af222582d43
|
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/iso.py#L24-L26
|
train
|
fitnr/convertdate
|
convertdate/iso.py
|
from_jd
|
def from_jd(jd):
'''Return tuple of ISO (year, week, day) for Julian day'''
year = gregorian.from_jd(jd)[0]
day = jwday(jd) + 1
dayofyear = ordinal.from_jd(jd)[1]
week = trunc((dayofyear - day + 10) / 7)
# Reset year
if week < 1:
week = weeks_per_year(year - 1)
year = year - 1
# Check that year actually has 53 weeks
elif week == 53 and weeks_per_year(year) != 53:
week = 1
year = year + 1
return year, week, day
|
python
|
def from_jd(jd):
'''Return tuple of ISO (year, week, day) for Julian day'''
year = gregorian.from_jd(jd)[0]
day = jwday(jd) + 1
dayofyear = ordinal.from_jd(jd)[1]
week = trunc((dayofyear - day + 10) / 7)
# Reset year
if week < 1:
week = weeks_per_year(year - 1)
year = year - 1
# Check that year actually has 53 weeks
elif week == 53 and weeks_per_year(year) != 53:
week = 1
year = year + 1
return year, week, day
|
[
"def",
"from_jd",
"(",
"jd",
")",
":",
"year",
"=",
"gregorian",
".",
"from_jd",
"(",
"jd",
")",
"[",
"0",
"]",
"day",
"=",
"jwday",
"(",
"jd",
")",
"+",
"1",
"dayofyear",
"=",
"ordinal",
".",
"from_jd",
"(",
"jd",
")",
"[",
"1",
"]",
"week",
"=",
"trunc",
"(",
"(",
"dayofyear",
"-",
"day",
"+",
"10",
")",
"/",
"7",
")",
"# Reset year",
"if",
"week",
"<",
"1",
":",
"week",
"=",
"weeks_per_year",
"(",
"year",
"-",
"1",
")",
"year",
"=",
"year",
"-",
"1",
"# Check that year actually has 53 weeks",
"elif",
"week",
"==",
"53",
"and",
"weeks_per_year",
"(",
"year",
")",
"!=",
"53",
":",
"week",
"=",
"1",
"year",
"=",
"year",
"+",
"1",
"return",
"year",
",",
"week",
",",
"day"
] |
Return tuple of ISO (year, week, day) for Julian day
|
[
"Return",
"tuple",
"of",
"ISO",
"(",
"year",
"week",
"day",
")",
"for",
"Julian",
"day"
] |
e920f168a87f99183b0aa7290d6c3af222582d43
|
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/iso.py#L29-L47
|
train
|
fitnr/convertdate
|
convertdate/iso.py
|
weeks_per_year
|
def weeks_per_year(year):
'''Number of ISO weeks in a year'''
# 53 weeks: any year starting on Thursday and any leap year starting on Wednesday
jan1 = jwday(gregorian.to_jd(year, 1, 1))
if jan1 == THU or (jan1 == WED and isleap(year)):
return 53
else:
return 52
|
python
|
def weeks_per_year(year):
'''Number of ISO weeks in a year'''
# 53 weeks: any year starting on Thursday and any leap year starting on Wednesday
jan1 = jwday(gregorian.to_jd(year, 1, 1))
if jan1 == THU or (jan1 == WED and isleap(year)):
return 53
else:
return 52
|
[
"def",
"weeks_per_year",
"(",
"year",
")",
":",
"# 53 weeks: any year starting on Thursday and any leap year starting on Wednesday",
"jan1",
"=",
"jwday",
"(",
"gregorian",
".",
"to_jd",
"(",
"year",
",",
"1",
",",
"1",
")",
")",
"if",
"jan1",
"==",
"THU",
"or",
"(",
"jan1",
"==",
"WED",
"and",
"isleap",
"(",
"year",
")",
")",
":",
"return",
"53",
"else",
":",
"return",
"52"
] |
Number of ISO weeks in a year
|
[
"Number",
"of",
"ISO",
"weeks",
"in",
"a",
"year"
] |
e920f168a87f99183b0aa7290d6c3af222582d43
|
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/iso.py#L50-L58
|
train
|
spacetelescope/stsci.tools
|
lib/stsci/tools/readgeis.py
|
stsci
|
def stsci(hdulist):
"""For STScI GEIS files, need to do extra steps."""
instrument = hdulist[0].header.get('INSTRUME', '')
# Update extension header keywords
if instrument in ("WFPC2", "FOC"):
rootname = hdulist[0].header.get('ROOTNAME', '')
filetype = hdulist[0].header.get('FILETYPE', '')
for i in range(1, len(hdulist)):
# Add name and extver attributes to match PyFITS data structure
hdulist[i].name = filetype
hdulist[i]._extver = i
# Add extension keywords for this chip to extension
hdulist[i].header['EXPNAME'] = (rootname, "9 character exposure identifier")
hdulist[i].header['EXTVER']= (i, "extension version number")
hdulist[i].header['EXTNAME'] = (filetype, "extension name")
hdulist[i].header['INHERIT'] = (True, "inherit the primary header")
hdulist[i].header['ROOTNAME'] = (rootname, "rootname of the observation set")
|
python
|
def stsci(hdulist):
"""For STScI GEIS files, need to do extra steps."""
instrument = hdulist[0].header.get('INSTRUME', '')
# Update extension header keywords
if instrument in ("WFPC2", "FOC"):
rootname = hdulist[0].header.get('ROOTNAME', '')
filetype = hdulist[0].header.get('FILETYPE', '')
for i in range(1, len(hdulist)):
# Add name and extver attributes to match PyFITS data structure
hdulist[i].name = filetype
hdulist[i]._extver = i
# Add extension keywords for this chip to extension
hdulist[i].header['EXPNAME'] = (rootname, "9 character exposure identifier")
hdulist[i].header['EXTVER']= (i, "extension version number")
hdulist[i].header['EXTNAME'] = (filetype, "extension name")
hdulist[i].header['INHERIT'] = (True, "inherit the primary header")
hdulist[i].header['ROOTNAME'] = (rootname, "rootname of the observation set")
|
[
"def",
"stsci",
"(",
"hdulist",
")",
":",
"instrument",
"=",
"hdulist",
"[",
"0",
"]",
".",
"header",
".",
"get",
"(",
"'INSTRUME'",
",",
"''",
")",
"# Update extension header keywords",
"if",
"instrument",
"in",
"(",
"\"WFPC2\"",
",",
"\"FOC\"",
")",
":",
"rootname",
"=",
"hdulist",
"[",
"0",
"]",
".",
"header",
".",
"get",
"(",
"'ROOTNAME'",
",",
"''",
")",
"filetype",
"=",
"hdulist",
"[",
"0",
"]",
".",
"header",
".",
"get",
"(",
"'FILETYPE'",
",",
"''",
")",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"len",
"(",
"hdulist",
")",
")",
":",
"# Add name and extver attributes to match PyFITS data structure",
"hdulist",
"[",
"i",
"]",
".",
"name",
"=",
"filetype",
"hdulist",
"[",
"i",
"]",
".",
"_extver",
"=",
"i",
"# Add extension keywords for this chip to extension",
"hdulist",
"[",
"i",
"]",
".",
"header",
"[",
"'EXPNAME'",
"]",
"=",
"(",
"rootname",
",",
"\"9 character exposure identifier\"",
")",
"hdulist",
"[",
"i",
"]",
".",
"header",
"[",
"'EXTVER'",
"]",
"=",
"(",
"i",
",",
"\"extension version number\"",
")",
"hdulist",
"[",
"i",
"]",
".",
"header",
"[",
"'EXTNAME'",
"]",
"=",
"(",
"filetype",
",",
"\"extension name\"",
")",
"hdulist",
"[",
"i",
"]",
".",
"header",
"[",
"'INHERIT'",
"]",
"=",
"(",
"True",
",",
"\"inherit the primary header\"",
")",
"hdulist",
"[",
"i",
"]",
".",
"header",
"[",
"'ROOTNAME'",
"]",
"=",
"(",
"rootname",
",",
"\"rootname of the observation set\"",
")"
] |
For STScI GEIS files, need to do extra steps.
|
[
"For",
"STScI",
"GEIS",
"files",
"need",
"to",
"do",
"extra",
"steps",
"."
] |
9a022503ad24ca54ce83331482dfa3ff6de9f403
|
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/readgeis.py#L79-L97
|
train
|
spacetelescope/stsci.tools
|
lib/stsci/tools/readgeis.py
|
stsci2
|
def stsci2(hdulist, filename):
"""For STScI GEIS files, need to do extra steps."""
# Write output file name to the primary header
instrument = hdulist[0].header.get('INSTRUME', '')
if instrument in ("WFPC2", "FOC"):
hdulist[0].header['FILENAME'] = filename
|
python
|
def stsci2(hdulist, filename):
"""For STScI GEIS files, need to do extra steps."""
# Write output file name to the primary header
instrument = hdulist[0].header.get('INSTRUME', '')
if instrument in ("WFPC2", "FOC"):
hdulist[0].header['FILENAME'] = filename
|
[
"def",
"stsci2",
"(",
"hdulist",
",",
"filename",
")",
":",
"# Write output file name to the primary header",
"instrument",
"=",
"hdulist",
"[",
"0",
"]",
".",
"header",
".",
"get",
"(",
"'INSTRUME'",
",",
"''",
")",
"if",
"instrument",
"in",
"(",
"\"WFPC2\"",
",",
"\"FOC\"",
")",
":",
"hdulist",
"[",
"0",
"]",
".",
"header",
"[",
"'FILENAME'",
"]",
"=",
"filename"
] |
For STScI GEIS files, need to do extra steps.
|
[
"For",
"STScI",
"GEIS",
"files",
"need",
"to",
"do",
"extra",
"steps",
"."
] |
9a022503ad24ca54ce83331482dfa3ff6de9f403
|
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/readgeis.py#L100-L106
|
train
|
spacetelescope/stsci.tools
|
lib/stsci/tools/readgeis.py
|
readgeis
|
def readgeis(input):
"""Input GEIS files "input" will be read and a HDUList object will
be returned.
The user can use the writeto method to write the HDUList object to
a FITS file.
"""
global dat
cardLen = fits.Card.length
# input file(s) must be of the form *.??h and *.??d
if input[-1] != 'h' or input[-4] != '.':
raise "Illegal input GEIS file name %s" % input
data_file = input[:-1]+'d'
_os = sys.platform
if _os[:5] == 'linux' or _os[:5] == 'win32' or _os[:5] == 'sunos' or _os[:3] == 'osf' or _os[:6] == 'darwin':
bytes_per_line = cardLen+1
else:
raise "Platform %s is not supported (yet)." % _os
geis_fmt = {'REAL':'f', 'INTEGER':'i', 'LOGICAL':'i','CHARACTER':'S'}
end_card = 'END'+' '* (cardLen-3)
# open input file
im = open(input)
# Generate the primary HDU
cards = []
while 1:
line = im.read(bytes_per_line)[:cardLen]
line = line[:8].upper() + line[8:]
if line == end_card:
break
cards.append(fits.Card.fromstring(line))
phdr = fits.Header(cards)
im.close()
_naxis0 = phdr.get('NAXIS', 0)
_naxis = [phdr['NAXIS'+str(j)] for j in range(1, _naxis0+1)]
_naxis.insert(0, _naxis0)
_bitpix = phdr['BITPIX']
_psize = phdr['PSIZE']
if phdr['DATATYPE'][:4] == 'REAL':
_bitpix = -_bitpix
if _naxis0 > 0:
size = reduce(lambda x,y:x*y, _naxis[1:])
data_size = abs(_bitpix) * size // 8
else:
data_size = 0
group_size = data_size + _psize // 8
# decode the group parameter definitions,
# group parameters will become extension header
groups = phdr['GROUPS']
gcount = phdr['GCOUNT']
pcount = phdr['PCOUNT']
formats = []
bools = []
floats = []
_range = range(1, pcount+1)
key = [phdr['PTYPE'+str(j)] for j in _range]
comm = [phdr.cards['PTYPE'+str(j)].comment for j in _range]
# delete group parameter definition header keywords
_list = ['PTYPE'+str(j) for j in _range] + \
['PDTYPE'+str(j) for j in _range] + \
['PSIZE'+str(j) for j in _range] + \
['DATATYPE', 'PSIZE', 'GCOUNT', 'PCOUNT', 'BSCALE', 'BZERO']
# Construct record array formats for the group parameters
# as interpreted from the Primary header file
for i in range(1, pcount+1):
ptype = key[i-1]
pdtype = phdr['PDTYPE'+str(i)]
star = pdtype.find('*')
_type = pdtype[:star]
_bytes = pdtype[star+1:]
# collect boolean keywords since they need special attention later
if _type == 'LOGICAL':
bools.append(i)
if pdtype == 'REAL*4':
floats.append(i)
fmt = geis_fmt[_type] + _bytes
formats.append((ptype,fmt))
_shape = _naxis[1:]
_shape.reverse()
_code = fits.BITPIX2DTYPE[_bitpix]
_bscale = phdr.get('BSCALE', 1)
_bzero = phdr.get('BZERO', 0)
if phdr['DATATYPE'][:10] == 'UNSIGNED*2':
_uint16 = 1
_bzero = 32768
else:
_uint16 = 0
# delete from the end, so it will not conflict with previous delete
for i in range(len(phdr)-1, -1, -1):
if phdr.cards[i].keyword in _list:
del phdr[i]
# clean up other primary header keywords
phdr['SIMPLE'] = True
phdr['BITPIX'] = 16
phdr['GROUPS'] = False
_after = 'NAXIS'
if _naxis0 > 0:
_after += str(_naxis0)
phdr.set('EXTEND', value=True, comment="FITS dataset may contain extensions", after=_after)
phdr.set('NEXTEND', value=gcount, comment="Number of standard extensions")
hdulist = fits.HDUList([fits.PrimaryHDU(header=phdr, data=None)])
# Use copy-on-write for all data types since byteswap may be needed
# in some platforms.
f1 = open(data_file, mode='rb')
dat = f1.read()
# dat = memmap(data_file, mode='c')
hdulist.mmobject = dat
errormsg = ""
loc = 0
for k in range(gcount):
ext_dat = numpy.fromstring(dat[loc:loc+data_size], dtype=_code)
ext_dat = ext_dat.reshape(_shape)
if _uint16:
ext_dat += _bzero
# Check to see whether there are any NaN's or infs which might indicate
# a byte-swapping problem, such as being written out on little-endian
# and being read in on big-endian or vice-versa.
if _code.find('float') >= 0 and \
(numpy.any(numpy.isnan(ext_dat)) or numpy.any(numpy.isinf(ext_dat))):
errormsg += "===================================\n"
errormsg += "= WARNING: =\n"
errormsg += "= Input image: =\n"
errormsg += input+"[%d]\n"%(k+1)
errormsg += "= had floating point data values =\n"
errormsg += "= of NaN and/or Inf. =\n"
errormsg += "===================================\n"
elif _code.find('int') >= 0:
# Check INT data for max values
ext_dat_frac,ext_dat_exp = numpy.frexp(ext_dat)
if ext_dat_exp.max() == int(_bitpix) - 1:
# Potential problems with byteswapping
errormsg += "===================================\n"
errormsg += "= WARNING: =\n"
errormsg += "= Input image: =\n"
errormsg += input+"[%d]\n"%(k+1)
errormsg += "= had integer data values =\n"
errormsg += "= with maximum bitvalues. =\n"
errormsg += "===================================\n"
ext_hdu = fits.ImageHDU(data=ext_dat)
rec = numpy.fromstring(dat[loc+data_size:loc+group_size], dtype=formats)
loc += group_size
# Create separate PyFITS Card objects for each entry in 'rec'
for i in range(1, pcount+1):
#val = rec.field(i-1)[0]
val = rec[0][i-1]
if val.dtype.kind == 'S':
val = val.decode('ascii')
if i in bools:
if val:
val = True
else:
val = False
if i in floats:
# use fromstring, format in Card is deprecated in pyfits 0.9
_str = '%-8s= %20.7G / %s' % (key[i-1], val, comm[i-1])
_card = fits.Card.fromstring(_str)
else:
_card = fits.Card(keyword=key[i-1], value=val, comment=comm[i-1])
ext_hdu.header.append(_card)
# deal with bscale/bzero
if (_bscale != 1 or _bzero != 0):
ext_hdu.header['BSCALE'] = _bscale
ext_hdu.header['BZERO'] = _bzero
hdulist.append(ext_hdu)
if errormsg != "":
errormsg += "===================================\n"
errormsg += "= This file may have been =\n"
errormsg += "= written out on a platform =\n"
errormsg += "= with a different byte-order. =\n"
errormsg += "= =\n"
errormsg += "= Please verify that the values =\n"
errormsg += "= are correct or apply the =\n"
errormsg += "= '.byteswap()' method. =\n"
errormsg += "===================================\n"
print(errormsg)
f1.close()
stsci(hdulist)
return hdulist
|
python
|
def readgeis(input):
"""Input GEIS files "input" will be read and a HDUList object will
be returned.
The user can use the writeto method to write the HDUList object to
a FITS file.
"""
global dat
cardLen = fits.Card.length
# input file(s) must be of the form *.??h and *.??d
if input[-1] != 'h' or input[-4] != '.':
raise "Illegal input GEIS file name %s" % input
data_file = input[:-1]+'d'
_os = sys.platform
if _os[:5] == 'linux' or _os[:5] == 'win32' or _os[:5] == 'sunos' or _os[:3] == 'osf' or _os[:6] == 'darwin':
bytes_per_line = cardLen+1
else:
raise "Platform %s is not supported (yet)." % _os
geis_fmt = {'REAL':'f', 'INTEGER':'i', 'LOGICAL':'i','CHARACTER':'S'}
end_card = 'END'+' '* (cardLen-3)
# open input file
im = open(input)
# Generate the primary HDU
cards = []
while 1:
line = im.read(bytes_per_line)[:cardLen]
line = line[:8].upper() + line[8:]
if line == end_card:
break
cards.append(fits.Card.fromstring(line))
phdr = fits.Header(cards)
im.close()
_naxis0 = phdr.get('NAXIS', 0)
_naxis = [phdr['NAXIS'+str(j)] for j in range(1, _naxis0+1)]
_naxis.insert(0, _naxis0)
_bitpix = phdr['BITPIX']
_psize = phdr['PSIZE']
if phdr['DATATYPE'][:4] == 'REAL':
_bitpix = -_bitpix
if _naxis0 > 0:
size = reduce(lambda x,y:x*y, _naxis[1:])
data_size = abs(_bitpix) * size // 8
else:
data_size = 0
group_size = data_size + _psize // 8
# decode the group parameter definitions,
# group parameters will become extension header
groups = phdr['GROUPS']
gcount = phdr['GCOUNT']
pcount = phdr['PCOUNT']
formats = []
bools = []
floats = []
_range = range(1, pcount+1)
key = [phdr['PTYPE'+str(j)] for j in _range]
comm = [phdr.cards['PTYPE'+str(j)].comment for j in _range]
# delete group parameter definition header keywords
_list = ['PTYPE'+str(j) for j in _range] + \
['PDTYPE'+str(j) for j in _range] + \
['PSIZE'+str(j) for j in _range] + \
['DATATYPE', 'PSIZE', 'GCOUNT', 'PCOUNT', 'BSCALE', 'BZERO']
# Construct record array formats for the group parameters
# as interpreted from the Primary header file
for i in range(1, pcount+1):
ptype = key[i-1]
pdtype = phdr['PDTYPE'+str(i)]
star = pdtype.find('*')
_type = pdtype[:star]
_bytes = pdtype[star+1:]
# collect boolean keywords since they need special attention later
if _type == 'LOGICAL':
bools.append(i)
if pdtype == 'REAL*4':
floats.append(i)
fmt = geis_fmt[_type] + _bytes
formats.append((ptype,fmt))
_shape = _naxis[1:]
_shape.reverse()
_code = fits.BITPIX2DTYPE[_bitpix]
_bscale = phdr.get('BSCALE', 1)
_bzero = phdr.get('BZERO', 0)
if phdr['DATATYPE'][:10] == 'UNSIGNED*2':
_uint16 = 1
_bzero = 32768
else:
_uint16 = 0
# delete from the end, so it will not conflict with previous delete
for i in range(len(phdr)-1, -1, -1):
if phdr.cards[i].keyword in _list:
del phdr[i]
# clean up other primary header keywords
phdr['SIMPLE'] = True
phdr['BITPIX'] = 16
phdr['GROUPS'] = False
_after = 'NAXIS'
if _naxis0 > 0:
_after += str(_naxis0)
phdr.set('EXTEND', value=True, comment="FITS dataset may contain extensions", after=_after)
phdr.set('NEXTEND', value=gcount, comment="Number of standard extensions")
hdulist = fits.HDUList([fits.PrimaryHDU(header=phdr, data=None)])
# Use copy-on-write for all data types since byteswap may be needed
# in some platforms.
f1 = open(data_file, mode='rb')
dat = f1.read()
# dat = memmap(data_file, mode='c')
hdulist.mmobject = dat
errormsg = ""
loc = 0
for k in range(gcount):
ext_dat = numpy.fromstring(dat[loc:loc+data_size], dtype=_code)
ext_dat = ext_dat.reshape(_shape)
if _uint16:
ext_dat += _bzero
# Check to see whether there are any NaN's or infs which might indicate
# a byte-swapping problem, such as being written out on little-endian
# and being read in on big-endian or vice-versa.
if _code.find('float') >= 0 and \
(numpy.any(numpy.isnan(ext_dat)) or numpy.any(numpy.isinf(ext_dat))):
errormsg += "===================================\n"
errormsg += "= WARNING: =\n"
errormsg += "= Input image: =\n"
errormsg += input+"[%d]\n"%(k+1)
errormsg += "= had floating point data values =\n"
errormsg += "= of NaN and/or Inf. =\n"
errormsg += "===================================\n"
elif _code.find('int') >= 0:
# Check INT data for max values
ext_dat_frac,ext_dat_exp = numpy.frexp(ext_dat)
if ext_dat_exp.max() == int(_bitpix) - 1:
# Potential problems with byteswapping
errormsg += "===================================\n"
errormsg += "= WARNING: =\n"
errormsg += "= Input image: =\n"
errormsg += input+"[%d]\n"%(k+1)
errormsg += "= had integer data values =\n"
errormsg += "= with maximum bitvalues. =\n"
errormsg += "===================================\n"
ext_hdu = fits.ImageHDU(data=ext_dat)
rec = numpy.fromstring(dat[loc+data_size:loc+group_size], dtype=formats)
loc += group_size
# Create separate PyFITS Card objects for each entry in 'rec'
for i in range(1, pcount+1):
#val = rec.field(i-1)[0]
val = rec[0][i-1]
if val.dtype.kind == 'S':
val = val.decode('ascii')
if i in bools:
if val:
val = True
else:
val = False
if i in floats:
# use fromstring, format in Card is deprecated in pyfits 0.9
_str = '%-8s= %20.7G / %s' % (key[i-1], val, comm[i-1])
_card = fits.Card.fromstring(_str)
else:
_card = fits.Card(keyword=key[i-1], value=val, comment=comm[i-1])
ext_hdu.header.append(_card)
# deal with bscale/bzero
if (_bscale != 1 or _bzero != 0):
ext_hdu.header['BSCALE'] = _bscale
ext_hdu.header['BZERO'] = _bzero
hdulist.append(ext_hdu)
if errormsg != "":
errormsg += "===================================\n"
errormsg += "= This file may have been =\n"
errormsg += "= written out on a platform =\n"
errormsg += "= with a different byte-order. =\n"
errormsg += "= =\n"
errormsg += "= Please verify that the values =\n"
errormsg += "= are correct or apply the =\n"
errormsg += "= '.byteswap()' method. =\n"
errormsg += "===================================\n"
print(errormsg)
f1.close()
stsci(hdulist)
return hdulist
|
[
"def",
"readgeis",
"(",
"input",
")",
":",
"global",
"dat",
"cardLen",
"=",
"fits",
".",
"Card",
".",
"length",
"# input file(s) must be of the form *.??h and *.??d",
"if",
"input",
"[",
"-",
"1",
"]",
"!=",
"'h'",
"or",
"input",
"[",
"-",
"4",
"]",
"!=",
"'.'",
":",
"raise",
"\"Illegal input GEIS file name %s\"",
"%",
"input",
"data_file",
"=",
"input",
"[",
":",
"-",
"1",
"]",
"+",
"'d'",
"_os",
"=",
"sys",
".",
"platform",
"if",
"_os",
"[",
":",
"5",
"]",
"==",
"'linux'",
"or",
"_os",
"[",
":",
"5",
"]",
"==",
"'win32'",
"or",
"_os",
"[",
":",
"5",
"]",
"==",
"'sunos'",
"or",
"_os",
"[",
":",
"3",
"]",
"==",
"'osf'",
"or",
"_os",
"[",
":",
"6",
"]",
"==",
"'darwin'",
":",
"bytes_per_line",
"=",
"cardLen",
"+",
"1",
"else",
":",
"raise",
"\"Platform %s is not supported (yet).\"",
"%",
"_os",
"geis_fmt",
"=",
"{",
"'REAL'",
":",
"'f'",
",",
"'INTEGER'",
":",
"'i'",
",",
"'LOGICAL'",
":",
"'i'",
",",
"'CHARACTER'",
":",
"'S'",
"}",
"end_card",
"=",
"'END'",
"+",
"' '",
"*",
"(",
"cardLen",
"-",
"3",
")",
"# open input file",
"im",
"=",
"open",
"(",
"input",
")",
"# Generate the primary HDU",
"cards",
"=",
"[",
"]",
"while",
"1",
":",
"line",
"=",
"im",
".",
"read",
"(",
"bytes_per_line",
")",
"[",
":",
"cardLen",
"]",
"line",
"=",
"line",
"[",
":",
"8",
"]",
".",
"upper",
"(",
")",
"+",
"line",
"[",
"8",
":",
"]",
"if",
"line",
"==",
"end_card",
":",
"break",
"cards",
".",
"append",
"(",
"fits",
".",
"Card",
".",
"fromstring",
"(",
"line",
")",
")",
"phdr",
"=",
"fits",
".",
"Header",
"(",
"cards",
")",
"im",
".",
"close",
"(",
")",
"_naxis0",
"=",
"phdr",
".",
"get",
"(",
"'NAXIS'",
",",
"0",
")",
"_naxis",
"=",
"[",
"phdr",
"[",
"'NAXIS'",
"+",
"str",
"(",
"j",
")",
"]",
"for",
"j",
"in",
"range",
"(",
"1",
",",
"_naxis0",
"+",
"1",
")",
"]",
"_naxis",
".",
"insert",
"(",
"0",
",",
"_naxis0",
")",
"_bitpix",
"=",
"phdr",
"[",
"'BITPIX'",
"]",
"_psize",
"=",
"phdr",
"[",
"'PSIZE'",
"]",
"if",
"phdr",
"[",
"'DATATYPE'",
"]",
"[",
":",
"4",
"]",
"==",
"'REAL'",
":",
"_bitpix",
"=",
"-",
"_bitpix",
"if",
"_naxis0",
">",
"0",
":",
"size",
"=",
"reduce",
"(",
"lambda",
"x",
",",
"y",
":",
"x",
"*",
"y",
",",
"_naxis",
"[",
"1",
":",
"]",
")",
"data_size",
"=",
"abs",
"(",
"_bitpix",
")",
"*",
"size",
"//",
"8",
"else",
":",
"data_size",
"=",
"0",
"group_size",
"=",
"data_size",
"+",
"_psize",
"//",
"8",
"# decode the group parameter definitions,",
"# group parameters will become extension header",
"groups",
"=",
"phdr",
"[",
"'GROUPS'",
"]",
"gcount",
"=",
"phdr",
"[",
"'GCOUNT'",
"]",
"pcount",
"=",
"phdr",
"[",
"'PCOUNT'",
"]",
"formats",
"=",
"[",
"]",
"bools",
"=",
"[",
"]",
"floats",
"=",
"[",
"]",
"_range",
"=",
"range",
"(",
"1",
",",
"pcount",
"+",
"1",
")",
"key",
"=",
"[",
"phdr",
"[",
"'PTYPE'",
"+",
"str",
"(",
"j",
")",
"]",
"for",
"j",
"in",
"_range",
"]",
"comm",
"=",
"[",
"phdr",
".",
"cards",
"[",
"'PTYPE'",
"+",
"str",
"(",
"j",
")",
"]",
".",
"comment",
"for",
"j",
"in",
"_range",
"]",
"# delete group parameter definition header keywords",
"_list",
"=",
"[",
"'PTYPE'",
"+",
"str",
"(",
"j",
")",
"for",
"j",
"in",
"_range",
"]",
"+",
"[",
"'PDTYPE'",
"+",
"str",
"(",
"j",
")",
"for",
"j",
"in",
"_range",
"]",
"+",
"[",
"'PSIZE'",
"+",
"str",
"(",
"j",
")",
"for",
"j",
"in",
"_range",
"]",
"+",
"[",
"'DATATYPE'",
",",
"'PSIZE'",
",",
"'GCOUNT'",
",",
"'PCOUNT'",
",",
"'BSCALE'",
",",
"'BZERO'",
"]",
"# Construct record array formats for the group parameters",
"# as interpreted from the Primary header file",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"pcount",
"+",
"1",
")",
":",
"ptype",
"=",
"key",
"[",
"i",
"-",
"1",
"]",
"pdtype",
"=",
"phdr",
"[",
"'PDTYPE'",
"+",
"str",
"(",
"i",
")",
"]",
"star",
"=",
"pdtype",
".",
"find",
"(",
"'*'",
")",
"_type",
"=",
"pdtype",
"[",
":",
"star",
"]",
"_bytes",
"=",
"pdtype",
"[",
"star",
"+",
"1",
":",
"]",
"# collect boolean keywords since they need special attention later",
"if",
"_type",
"==",
"'LOGICAL'",
":",
"bools",
".",
"append",
"(",
"i",
")",
"if",
"pdtype",
"==",
"'REAL*4'",
":",
"floats",
".",
"append",
"(",
"i",
")",
"fmt",
"=",
"geis_fmt",
"[",
"_type",
"]",
"+",
"_bytes",
"formats",
".",
"append",
"(",
"(",
"ptype",
",",
"fmt",
")",
")",
"_shape",
"=",
"_naxis",
"[",
"1",
":",
"]",
"_shape",
".",
"reverse",
"(",
")",
"_code",
"=",
"fits",
".",
"BITPIX2DTYPE",
"[",
"_bitpix",
"]",
"_bscale",
"=",
"phdr",
".",
"get",
"(",
"'BSCALE'",
",",
"1",
")",
"_bzero",
"=",
"phdr",
".",
"get",
"(",
"'BZERO'",
",",
"0",
")",
"if",
"phdr",
"[",
"'DATATYPE'",
"]",
"[",
":",
"10",
"]",
"==",
"'UNSIGNED*2'",
":",
"_uint16",
"=",
"1",
"_bzero",
"=",
"32768",
"else",
":",
"_uint16",
"=",
"0",
"# delete from the end, so it will not conflict with previous delete",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"phdr",
")",
"-",
"1",
",",
"-",
"1",
",",
"-",
"1",
")",
":",
"if",
"phdr",
".",
"cards",
"[",
"i",
"]",
".",
"keyword",
"in",
"_list",
":",
"del",
"phdr",
"[",
"i",
"]",
"# clean up other primary header keywords",
"phdr",
"[",
"'SIMPLE'",
"]",
"=",
"True",
"phdr",
"[",
"'BITPIX'",
"]",
"=",
"16",
"phdr",
"[",
"'GROUPS'",
"]",
"=",
"False",
"_after",
"=",
"'NAXIS'",
"if",
"_naxis0",
">",
"0",
":",
"_after",
"+=",
"str",
"(",
"_naxis0",
")",
"phdr",
".",
"set",
"(",
"'EXTEND'",
",",
"value",
"=",
"True",
",",
"comment",
"=",
"\"FITS dataset may contain extensions\"",
",",
"after",
"=",
"_after",
")",
"phdr",
".",
"set",
"(",
"'NEXTEND'",
",",
"value",
"=",
"gcount",
",",
"comment",
"=",
"\"Number of standard extensions\"",
")",
"hdulist",
"=",
"fits",
".",
"HDUList",
"(",
"[",
"fits",
".",
"PrimaryHDU",
"(",
"header",
"=",
"phdr",
",",
"data",
"=",
"None",
")",
"]",
")",
"# Use copy-on-write for all data types since byteswap may be needed",
"# in some platforms.",
"f1",
"=",
"open",
"(",
"data_file",
",",
"mode",
"=",
"'rb'",
")",
"dat",
"=",
"f1",
".",
"read",
"(",
")",
"# dat = memmap(data_file, mode='c')",
"hdulist",
".",
"mmobject",
"=",
"dat",
"errormsg",
"=",
"\"\"",
"loc",
"=",
"0",
"for",
"k",
"in",
"range",
"(",
"gcount",
")",
":",
"ext_dat",
"=",
"numpy",
".",
"fromstring",
"(",
"dat",
"[",
"loc",
":",
"loc",
"+",
"data_size",
"]",
",",
"dtype",
"=",
"_code",
")",
"ext_dat",
"=",
"ext_dat",
".",
"reshape",
"(",
"_shape",
")",
"if",
"_uint16",
":",
"ext_dat",
"+=",
"_bzero",
"# Check to see whether there are any NaN's or infs which might indicate",
"# a byte-swapping problem, such as being written out on little-endian",
"# and being read in on big-endian or vice-versa.",
"if",
"_code",
".",
"find",
"(",
"'float'",
")",
">=",
"0",
"and",
"(",
"numpy",
".",
"any",
"(",
"numpy",
".",
"isnan",
"(",
"ext_dat",
")",
")",
"or",
"numpy",
".",
"any",
"(",
"numpy",
".",
"isinf",
"(",
"ext_dat",
")",
")",
")",
":",
"errormsg",
"+=",
"\"===================================\\n\"",
"errormsg",
"+=",
"\"= WARNING: =\\n\"",
"errormsg",
"+=",
"\"= Input image: =\\n\"",
"errormsg",
"+=",
"input",
"+",
"\"[%d]\\n\"",
"%",
"(",
"k",
"+",
"1",
")",
"errormsg",
"+=",
"\"= had floating point data values =\\n\"",
"errormsg",
"+=",
"\"= of NaN and/or Inf. =\\n\"",
"errormsg",
"+=",
"\"===================================\\n\"",
"elif",
"_code",
".",
"find",
"(",
"'int'",
")",
">=",
"0",
":",
"# Check INT data for max values",
"ext_dat_frac",
",",
"ext_dat_exp",
"=",
"numpy",
".",
"frexp",
"(",
"ext_dat",
")",
"if",
"ext_dat_exp",
".",
"max",
"(",
")",
"==",
"int",
"(",
"_bitpix",
")",
"-",
"1",
":",
"# Potential problems with byteswapping",
"errormsg",
"+=",
"\"===================================\\n\"",
"errormsg",
"+=",
"\"= WARNING: =\\n\"",
"errormsg",
"+=",
"\"= Input image: =\\n\"",
"errormsg",
"+=",
"input",
"+",
"\"[%d]\\n\"",
"%",
"(",
"k",
"+",
"1",
")",
"errormsg",
"+=",
"\"= had integer data values =\\n\"",
"errormsg",
"+=",
"\"= with maximum bitvalues. =\\n\"",
"errormsg",
"+=",
"\"===================================\\n\"",
"ext_hdu",
"=",
"fits",
".",
"ImageHDU",
"(",
"data",
"=",
"ext_dat",
")",
"rec",
"=",
"numpy",
".",
"fromstring",
"(",
"dat",
"[",
"loc",
"+",
"data_size",
":",
"loc",
"+",
"group_size",
"]",
",",
"dtype",
"=",
"formats",
")",
"loc",
"+=",
"group_size",
"# Create separate PyFITS Card objects for each entry in 'rec'",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"pcount",
"+",
"1",
")",
":",
"#val = rec.field(i-1)[0]",
"val",
"=",
"rec",
"[",
"0",
"]",
"[",
"i",
"-",
"1",
"]",
"if",
"val",
".",
"dtype",
".",
"kind",
"==",
"'S'",
":",
"val",
"=",
"val",
".",
"decode",
"(",
"'ascii'",
")",
"if",
"i",
"in",
"bools",
":",
"if",
"val",
":",
"val",
"=",
"True",
"else",
":",
"val",
"=",
"False",
"if",
"i",
"in",
"floats",
":",
"# use fromstring, format in Card is deprecated in pyfits 0.9",
"_str",
"=",
"'%-8s= %20.7G / %s'",
"%",
"(",
"key",
"[",
"i",
"-",
"1",
"]",
",",
"val",
",",
"comm",
"[",
"i",
"-",
"1",
"]",
")",
"_card",
"=",
"fits",
".",
"Card",
".",
"fromstring",
"(",
"_str",
")",
"else",
":",
"_card",
"=",
"fits",
".",
"Card",
"(",
"keyword",
"=",
"key",
"[",
"i",
"-",
"1",
"]",
",",
"value",
"=",
"val",
",",
"comment",
"=",
"comm",
"[",
"i",
"-",
"1",
"]",
")",
"ext_hdu",
".",
"header",
".",
"append",
"(",
"_card",
")",
"# deal with bscale/bzero",
"if",
"(",
"_bscale",
"!=",
"1",
"or",
"_bzero",
"!=",
"0",
")",
":",
"ext_hdu",
".",
"header",
"[",
"'BSCALE'",
"]",
"=",
"_bscale",
"ext_hdu",
".",
"header",
"[",
"'BZERO'",
"]",
"=",
"_bzero",
"hdulist",
".",
"append",
"(",
"ext_hdu",
")",
"if",
"errormsg",
"!=",
"\"\"",
":",
"errormsg",
"+=",
"\"===================================\\n\"",
"errormsg",
"+=",
"\"= This file may have been =\\n\"",
"errormsg",
"+=",
"\"= written out on a platform =\\n\"",
"errormsg",
"+=",
"\"= with a different byte-order. =\\n\"",
"errormsg",
"+=",
"\"= =\\n\"",
"errormsg",
"+=",
"\"= Please verify that the values =\\n\"",
"errormsg",
"+=",
"\"= are correct or apply the =\\n\"",
"errormsg",
"+=",
"\"= '.byteswap()' method. =\\n\"",
"errormsg",
"+=",
"\"===================================\\n\"",
"print",
"(",
"errormsg",
")",
"f1",
".",
"close",
"(",
")",
"stsci",
"(",
"hdulist",
")",
"return",
"hdulist"
] |
Input GEIS files "input" will be read and a HDUList object will
be returned.
The user can use the writeto method to write the HDUList object to
a FITS file.
|
[
"Input",
"GEIS",
"files",
"input",
"will",
"be",
"read",
"and",
"a",
"HDUList",
"object",
"will",
"be",
"returned",
"."
] |
9a022503ad24ca54ce83331482dfa3ff6de9f403
|
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/readgeis.py#L109-L320
|
train
|
spacetelescope/stsci.tools
|
lib/stsci/tools/readgeis.py
|
parse_path
|
def parse_path(f1, f2):
"""Parse two input arguments and return two lists of file names"""
import glob
# if second argument is missing or is a wild card, point it
# to the current directory
f2 = f2.strip()
if f2 == '' or f2 == '*':
f2 = './'
# if the first argument is a directory, use all GEIS files
if os.path.isdir(f1):
f1 = os.path.join(f1, '*.??h')
list1 = glob.glob(f1)
list1 = [name for name in list1 if name[-1] == 'h' and name[-4] == '.']
# if the second argument is a directory, use file names in the
# first argument to construct file names, i.e.
# abc.xyh will be converted to abc_xyf.fits
if os.path.isdir(f2):
list2 = []
for file in list1:
name = os.path.split(file)[-1]
fitsname = name[:-4] + '_' + name[-3:-1] + 'f.fits'
list2.append(os.path.join(f2, fitsname))
else:
list2 = [s.strip() for s in f2.split(",")]
if list1 == [] or list2 == []:
err_msg = ""
if list1 == []:
err_msg += "Input files `{:s}` not usable/available. ".format(f1)
if list2 == []:
err_msg += "Input files `{:s}` not usable/available. ".format(f2)
raise IOError(err_msg)
else:
return list1, list2
|
python
|
def parse_path(f1, f2):
"""Parse two input arguments and return two lists of file names"""
import glob
# if second argument is missing or is a wild card, point it
# to the current directory
f2 = f2.strip()
if f2 == '' or f2 == '*':
f2 = './'
# if the first argument is a directory, use all GEIS files
if os.path.isdir(f1):
f1 = os.path.join(f1, '*.??h')
list1 = glob.glob(f1)
list1 = [name for name in list1 if name[-1] == 'h' and name[-4] == '.']
# if the second argument is a directory, use file names in the
# first argument to construct file names, i.e.
# abc.xyh will be converted to abc_xyf.fits
if os.path.isdir(f2):
list2 = []
for file in list1:
name = os.path.split(file)[-1]
fitsname = name[:-4] + '_' + name[-3:-1] + 'f.fits'
list2.append(os.path.join(f2, fitsname))
else:
list2 = [s.strip() for s in f2.split(",")]
if list1 == [] or list2 == []:
err_msg = ""
if list1 == []:
err_msg += "Input files `{:s}` not usable/available. ".format(f1)
if list2 == []:
err_msg += "Input files `{:s}` not usable/available. ".format(f2)
raise IOError(err_msg)
else:
return list1, list2
|
[
"def",
"parse_path",
"(",
"f1",
",",
"f2",
")",
":",
"import",
"glob",
"# if second argument is missing or is a wild card, point it",
"# to the current directory",
"f2",
"=",
"f2",
".",
"strip",
"(",
")",
"if",
"f2",
"==",
"''",
"or",
"f2",
"==",
"'*'",
":",
"f2",
"=",
"'./'",
"# if the first argument is a directory, use all GEIS files",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"f1",
")",
":",
"f1",
"=",
"os",
".",
"path",
".",
"join",
"(",
"f1",
",",
"'*.??h'",
")",
"list1",
"=",
"glob",
".",
"glob",
"(",
"f1",
")",
"list1",
"=",
"[",
"name",
"for",
"name",
"in",
"list1",
"if",
"name",
"[",
"-",
"1",
"]",
"==",
"'h'",
"and",
"name",
"[",
"-",
"4",
"]",
"==",
"'.'",
"]",
"# if the second argument is a directory, use file names in the",
"# first argument to construct file names, i.e.",
"# abc.xyh will be converted to abc_xyf.fits",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"f2",
")",
":",
"list2",
"=",
"[",
"]",
"for",
"file",
"in",
"list1",
":",
"name",
"=",
"os",
".",
"path",
".",
"split",
"(",
"file",
")",
"[",
"-",
"1",
"]",
"fitsname",
"=",
"name",
"[",
":",
"-",
"4",
"]",
"+",
"'_'",
"+",
"name",
"[",
"-",
"3",
":",
"-",
"1",
"]",
"+",
"'f.fits'",
"list2",
".",
"append",
"(",
"os",
".",
"path",
".",
"join",
"(",
"f2",
",",
"fitsname",
")",
")",
"else",
":",
"list2",
"=",
"[",
"s",
".",
"strip",
"(",
")",
"for",
"s",
"in",
"f2",
".",
"split",
"(",
"\",\"",
")",
"]",
"if",
"list1",
"==",
"[",
"]",
"or",
"list2",
"==",
"[",
"]",
":",
"err_msg",
"=",
"\"\"",
"if",
"list1",
"==",
"[",
"]",
":",
"err_msg",
"+=",
"\"Input files `{:s}` not usable/available. \"",
".",
"format",
"(",
"f1",
")",
"if",
"list2",
"==",
"[",
"]",
":",
"err_msg",
"+=",
"\"Input files `{:s}` not usable/available. \"",
".",
"format",
"(",
"f2",
")",
"raise",
"IOError",
"(",
"err_msg",
")",
"else",
":",
"return",
"list1",
",",
"list2"
] |
Parse two input arguments and return two lists of file names
|
[
"Parse",
"two",
"input",
"arguments",
"and",
"return",
"two",
"lists",
"of",
"file",
"names"
] |
9a022503ad24ca54ce83331482dfa3ff6de9f403
|
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/readgeis.py#L322-L363
|
train
|
spacetelescope/stsci.tools
|
lib/stsci/tools/parseinput.py
|
parseinput
|
def parseinput(inputlist,outputname=None, atfile=None):
"""
Recursively parse user input based upon the irafglob
program and construct a list of files that need to be processed.
This program addresses the following deficiencies of the irafglob program::
parseinput can extract filenames from association tables
Returns
-------
This program will return a list of input files that will need to
be processed in addition to the name of any outfiles specified in
an association table.
Parameters
----------
inputlist - string
specification of input files using either wild-cards, @-file or
comma-separated list of filenames
outputname - string
desired name for output product to be created from the input files
atfile - object
function to use in interpreting the @-file columns that gets passed to irafglob
Returns
-------
files - list of strings
names of output files to be processed
newoutputname - string
name of output file to be created.
See Also
--------
stsci.tools.irafglob
"""
# Initalize some variables
files = [] # list used to store names of input files
newoutputname = outputname # Outputname returned to calling program.
# The value of outputname is only changed
# if it had a value of 'None' on input.
# We can use irafglob to parse the input. If the input wasn't
# an association table, it needs to be either a wildcard, '@' file,
# or comma seperated list.
files = irafglob(inputlist, atfile=atfile)
# Now that we have expanded the inputlist into a python list
# containing the list of input files, it is necessary to examine
# each of the files to make sure none of them are association tables.
#
# If an association table is found, the entries should be read
# Determine if the input is an association table
for file in files:
if (checkASN(file) == True):
# Create a list to store the files extracted from the
# association tiable
assoclist = []
# The input is an association table
try:
# Open the association table
assocdict = readASNTable(file, None, prodonly=False)
except:
errorstr = "###################################\n"
errorstr += "# #\n"
errorstr += "# UNABLE TO READ ASSOCIATION FILE,#\n"
errorstr += str(file)+'\n'
errorstr += "# DURING FILE PARSING. #\n"
errorstr += "# #\n"
errorstr += "# Please determine if the file is #\n"
errorstr += "# in the current directory and #\n"
errorstr += "# that it has been properly #\n"
errorstr += "# formatted. #\n"
errorstr += "# #\n"
errorstr += "# This error message is being #\n"
errorstr += "# generated from within the #\n"
errorstr += "# parseinput.py module. #\n"
errorstr += "# #\n"
errorstr += "###################################\n"
raise ValueError(errorstr)
# Extract the output name from the association table if None
# was provided on input.
if outputname is None:
newoutputname = assocdict['output']
# Loop over the association dictionary to extract the input
# file names.
for f in assocdict['order']:
assoclist.append(fileutil.buildRootname(f))
# Remove the name of the association table from the list of files
files.remove(file)
# Append the list of filenames generated from the association table
# to the master list of input files.
files.extend(assoclist)
# Return the list of the input files and the output name if provided in an association.
return files, newoutputname
|
python
|
def parseinput(inputlist,outputname=None, atfile=None):
"""
Recursively parse user input based upon the irafglob
program and construct a list of files that need to be processed.
This program addresses the following deficiencies of the irafglob program::
parseinput can extract filenames from association tables
Returns
-------
This program will return a list of input files that will need to
be processed in addition to the name of any outfiles specified in
an association table.
Parameters
----------
inputlist - string
specification of input files using either wild-cards, @-file or
comma-separated list of filenames
outputname - string
desired name for output product to be created from the input files
atfile - object
function to use in interpreting the @-file columns that gets passed to irafglob
Returns
-------
files - list of strings
names of output files to be processed
newoutputname - string
name of output file to be created.
See Also
--------
stsci.tools.irafglob
"""
# Initalize some variables
files = [] # list used to store names of input files
newoutputname = outputname # Outputname returned to calling program.
# The value of outputname is only changed
# if it had a value of 'None' on input.
# We can use irafglob to parse the input. If the input wasn't
# an association table, it needs to be either a wildcard, '@' file,
# or comma seperated list.
files = irafglob(inputlist, atfile=atfile)
# Now that we have expanded the inputlist into a python list
# containing the list of input files, it is necessary to examine
# each of the files to make sure none of them are association tables.
#
# If an association table is found, the entries should be read
# Determine if the input is an association table
for file in files:
if (checkASN(file) == True):
# Create a list to store the files extracted from the
# association tiable
assoclist = []
# The input is an association table
try:
# Open the association table
assocdict = readASNTable(file, None, prodonly=False)
except:
errorstr = "###################################\n"
errorstr += "# #\n"
errorstr += "# UNABLE TO READ ASSOCIATION FILE,#\n"
errorstr += str(file)+'\n'
errorstr += "# DURING FILE PARSING. #\n"
errorstr += "# #\n"
errorstr += "# Please determine if the file is #\n"
errorstr += "# in the current directory and #\n"
errorstr += "# that it has been properly #\n"
errorstr += "# formatted. #\n"
errorstr += "# #\n"
errorstr += "# This error message is being #\n"
errorstr += "# generated from within the #\n"
errorstr += "# parseinput.py module. #\n"
errorstr += "# #\n"
errorstr += "###################################\n"
raise ValueError(errorstr)
# Extract the output name from the association table if None
# was provided on input.
if outputname is None:
newoutputname = assocdict['output']
# Loop over the association dictionary to extract the input
# file names.
for f in assocdict['order']:
assoclist.append(fileutil.buildRootname(f))
# Remove the name of the association table from the list of files
files.remove(file)
# Append the list of filenames generated from the association table
# to the master list of input files.
files.extend(assoclist)
# Return the list of the input files and the output name if provided in an association.
return files, newoutputname
|
[
"def",
"parseinput",
"(",
"inputlist",
",",
"outputname",
"=",
"None",
",",
"atfile",
"=",
"None",
")",
":",
"# Initalize some variables",
"files",
"=",
"[",
"]",
"# list used to store names of input files",
"newoutputname",
"=",
"outputname",
"# Outputname returned to calling program.",
"# The value of outputname is only changed",
"# if it had a value of 'None' on input.",
"# We can use irafglob to parse the input. If the input wasn't",
"# an association table, it needs to be either a wildcard, '@' file,",
"# or comma seperated list.",
"files",
"=",
"irafglob",
"(",
"inputlist",
",",
"atfile",
"=",
"atfile",
")",
"# Now that we have expanded the inputlist into a python list",
"# containing the list of input files, it is necessary to examine",
"# each of the files to make sure none of them are association tables.",
"#",
"# If an association table is found, the entries should be read",
"# Determine if the input is an association table",
"for",
"file",
"in",
"files",
":",
"if",
"(",
"checkASN",
"(",
"file",
")",
"==",
"True",
")",
":",
"# Create a list to store the files extracted from the",
"# association tiable",
"assoclist",
"=",
"[",
"]",
"# The input is an association table",
"try",
":",
"# Open the association table",
"assocdict",
"=",
"readASNTable",
"(",
"file",
",",
"None",
",",
"prodonly",
"=",
"False",
")",
"except",
":",
"errorstr",
"=",
"\"###################################\\n\"",
"errorstr",
"+=",
"\"# #\\n\"",
"errorstr",
"+=",
"\"# UNABLE TO READ ASSOCIATION FILE,#\\n\"",
"errorstr",
"+=",
"str",
"(",
"file",
")",
"+",
"'\\n'",
"errorstr",
"+=",
"\"# DURING FILE PARSING. #\\n\"",
"errorstr",
"+=",
"\"# #\\n\"",
"errorstr",
"+=",
"\"# Please determine if the file is #\\n\"",
"errorstr",
"+=",
"\"# in the current directory and #\\n\"",
"errorstr",
"+=",
"\"# that it has been properly #\\n\"",
"errorstr",
"+=",
"\"# formatted. #\\n\"",
"errorstr",
"+=",
"\"# #\\n\"",
"errorstr",
"+=",
"\"# This error message is being #\\n\"",
"errorstr",
"+=",
"\"# generated from within the #\\n\"",
"errorstr",
"+=",
"\"# parseinput.py module. #\\n\"",
"errorstr",
"+=",
"\"# #\\n\"",
"errorstr",
"+=",
"\"###################################\\n\"",
"raise",
"ValueError",
"(",
"errorstr",
")",
"# Extract the output name from the association table if None",
"# was provided on input.",
"if",
"outputname",
"is",
"None",
":",
"newoutputname",
"=",
"assocdict",
"[",
"'output'",
"]",
"# Loop over the association dictionary to extract the input",
"# file names.",
"for",
"f",
"in",
"assocdict",
"[",
"'order'",
"]",
":",
"assoclist",
".",
"append",
"(",
"fileutil",
".",
"buildRootname",
"(",
"f",
")",
")",
"# Remove the name of the association table from the list of files",
"files",
".",
"remove",
"(",
"file",
")",
"# Append the list of filenames generated from the association table",
"# to the master list of input files.",
"files",
".",
"extend",
"(",
"assoclist",
")",
"# Return the list of the input files and the output name if provided in an association.",
"return",
"files",
",",
"newoutputname"
] |
Recursively parse user input based upon the irafglob
program and construct a list of files that need to be processed.
This program addresses the following deficiencies of the irafglob program::
parseinput can extract filenames from association tables
Returns
-------
This program will return a list of input files that will need to
be processed in addition to the name of any outfiles specified in
an association table.
Parameters
----------
inputlist - string
specification of input files using either wild-cards, @-file or
comma-separated list of filenames
outputname - string
desired name for output product to be created from the input files
atfile - object
function to use in interpreting the @-file columns that gets passed to irafglob
Returns
-------
files - list of strings
names of output files to be processed
newoutputname - string
name of output file to be created.
See Also
--------
stsci.tools.irafglob
|
[
"Recursively",
"parse",
"user",
"input",
"based",
"upon",
"the",
"irafglob",
"program",
"and",
"construct",
"a",
"list",
"of",
"files",
"that",
"need",
"to",
"be",
"processed",
".",
"This",
"program",
"addresses",
"the",
"following",
"deficiencies",
"of",
"the",
"irafglob",
"program",
"::"
] |
9a022503ad24ca54ce83331482dfa3ff6de9f403
|
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/parseinput.py#L31-L134
|
train
|
spacetelescope/stsci.tools
|
lib/stsci/tools/parseinput.py
|
checkASN
|
def checkASN(filename):
"""
Determine if the filename provided to the function belongs to
an association.
Parameters
----------
filename: string
Returns
-------
validASN : boolean value
"""
# Extract the file extn type:
extnType = filename[filename.rfind('_')+1:filename.rfind('.')]
# Determine if this extn name is valid for an assocation file
if isValidAssocExtn(extnType):
return True
else:
return False
|
python
|
def checkASN(filename):
"""
Determine if the filename provided to the function belongs to
an association.
Parameters
----------
filename: string
Returns
-------
validASN : boolean value
"""
# Extract the file extn type:
extnType = filename[filename.rfind('_')+1:filename.rfind('.')]
# Determine if this extn name is valid for an assocation file
if isValidAssocExtn(extnType):
return True
else:
return False
|
[
"def",
"checkASN",
"(",
"filename",
")",
":",
"# Extract the file extn type:",
"extnType",
"=",
"filename",
"[",
"filename",
".",
"rfind",
"(",
"'_'",
")",
"+",
"1",
":",
"filename",
".",
"rfind",
"(",
"'.'",
")",
"]",
"# Determine if this extn name is valid for an assocation file",
"if",
"isValidAssocExtn",
"(",
"extnType",
")",
":",
"return",
"True",
"else",
":",
"return",
"False"
] |
Determine if the filename provided to the function belongs to
an association.
Parameters
----------
filename: string
Returns
-------
validASN : boolean value
|
[
"Determine",
"if",
"the",
"filename",
"provided",
"to",
"the",
"function",
"belongs",
"to",
"an",
"association",
"."
] |
9a022503ad24ca54ce83331482dfa3ff6de9f403
|
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/parseinput.py#L137-L158
|
train
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.