repo
stringlengths 7
55
| path
stringlengths 4
223
| func_name
stringlengths 1
134
| original_string
stringlengths 75
104k
| language
stringclasses 1
value | code
stringlengths 75
104k
| code_tokens
listlengths 19
28.4k
| docstring
stringlengths 1
46.9k
| docstring_tokens
listlengths 1
1.97k
| sha
stringlengths 40
40
| url
stringlengths 87
315
| partition
stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
xingjiepan/cylinder_fitting
|
cylinder_fitting/visualize.py
|
show_fit
|
def show_fit(w_fit, C_fit, r_fit, Xs):
'''Plot the fitting given the fitted axis direction, the fitted
center, the fitted radius and the data points.
'''
fig = plt.figure()
ax = fig.gca(projection='3d')
# Plot the data points
ax.scatter([X[0] for X in Xs], [X[1] for X in Xs], [X[2] for X in Xs])
# Get the transformation matrix
theta = np.arccos(np.dot(w_fit, np.array([0, 0, 1])))
phi = np.arctan2(w_fit[1], w_fit[0])
M = np.dot(rotation_matrix_from_axis_and_angle(np.array([0, 0, 1]), phi),
rotation_matrix_from_axis_and_angle(np.array([0, 1, 0]), theta))
# Plot the cylinder surface
delta = np.linspace(-np.pi, np.pi, 20)
z = np.linspace(-10, 10, 20)
Delta, Z = np.meshgrid(delta, z)
X = r_fit * np.cos(Delta)
Y = r_fit * np.sin(Delta)
for i in range(len(X)):
for j in range(len(X[i])):
p = np.dot(M, np.array([X[i][j], Y[i][j], Z[i][j]])) + C_fit
X[i][j] = p[0]
Y[i][j] = p[1]
Z[i][j] = p[2]
ax.plot_surface(X, Y, Z, alpha=0.2)
# Plot the center and direction
ax.quiver(C_fit[0], C_fit[1], C_fit[2],
r_fit * w_fit[0], r_fit * w_fit[1], r_fit * w_fit[2], color='red')
plt.show()
|
python
|
def show_fit(w_fit, C_fit, r_fit, Xs):
'''Plot the fitting given the fitted axis direction, the fitted
center, the fitted radius and the data points.
'''
fig = plt.figure()
ax = fig.gca(projection='3d')
# Plot the data points
ax.scatter([X[0] for X in Xs], [X[1] for X in Xs], [X[2] for X in Xs])
# Get the transformation matrix
theta = np.arccos(np.dot(w_fit, np.array([0, 0, 1])))
phi = np.arctan2(w_fit[1], w_fit[0])
M = np.dot(rotation_matrix_from_axis_and_angle(np.array([0, 0, 1]), phi),
rotation_matrix_from_axis_and_angle(np.array([0, 1, 0]), theta))
# Plot the cylinder surface
delta = np.linspace(-np.pi, np.pi, 20)
z = np.linspace(-10, 10, 20)
Delta, Z = np.meshgrid(delta, z)
X = r_fit * np.cos(Delta)
Y = r_fit * np.sin(Delta)
for i in range(len(X)):
for j in range(len(X[i])):
p = np.dot(M, np.array([X[i][j], Y[i][j], Z[i][j]])) + C_fit
X[i][j] = p[0]
Y[i][j] = p[1]
Z[i][j] = p[2]
ax.plot_surface(X, Y, Z, alpha=0.2)
# Plot the center and direction
ax.quiver(C_fit[0], C_fit[1], C_fit[2],
r_fit * w_fit[0], r_fit * w_fit[1], r_fit * w_fit[2], color='red')
plt.show()
|
[
"def",
"show_fit",
"(",
"w_fit",
",",
"C_fit",
",",
"r_fit",
",",
"Xs",
")",
":",
"fig",
"=",
"plt",
".",
"figure",
"(",
")",
"ax",
"=",
"fig",
".",
"gca",
"(",
"projection",
"=",
"'3d'",
")",
"# Plot the data points",
"ax",
".",
"scatter",
"(",
"[",
"X",
"[",
"0",
"]",
"for",
"X",
"in",
"Xs",
"]",
",",
"[",
"X",
"[",
"1",
"]",
"for",
"X",
"in",
"Xs",
"]",
",",
"[",
"X",
"[",
"2",
"]",
"for",
"X",
"in",
"Xs",
"]",
")",
"# Get the transformation matrix",
"theta",
"=",
"np",
".",
"arccos",
"(",
"np",
".",
"dot",
"(",
"w_fit",
",",
"np",
".",
"array",
"(",
"[",
"0",
",",
"0",
",",
"1",
"]",
")",
")",
")",
"phi",
"=",
"np",
".",
"arctan2",
"(",
"w_fit",
"[",
"1",
"]",
",",
"w_fit",
"[",
"0",
"]",
")",
"M",
"=",
"np",
".",
"dot",
"(",
"rotation_matrix_from_axis_and_angle",
"(",
"np",
".",
"array",
"(",
"[",
"0",
",",
"0",
",",
"1",
"]",
")",
",",
"phi",
")",
",",
"rotation_matrix_from_axis_and_angle",
"(",
"np",
".",
"array",
"(",
"[",
"0",
",",
"1",
",",
"0",
"]",
")",
",",
"theta",
")",
")",
"# Plot the cylinder surface",
"delta",
"=",
"np",
".",
"linspace",
"(",
"-",
"np",
".",
"pi",
",",
"np",
".",
"pi",
",",
"20",
")",
"z",
"=",
"np",
".",
"linspace",
"(",
"-",
"10",
",",
"10",
",",
"20",
")",
"Delta",
",",
"Z",
"=",
"np",
".",
"meshgrid",
"(",
"delta",
",",
"z",
")",
"X",
"=",
"r_fit",
"*",
"np",
".",
"cos",
"(",
"Delta",
")",
"Y",
"=",
"r_fit",
"*",
"np",
".",
"sin",
"(",
"Delta",
")",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"X",
")",
")",
":",
"for",
"j",
"in",
"range",
"(",
"len",
"(",
"X",
"[",
"i",
"]",
")",
")",
":",
"p",
"=",
"np",
".",
"dot",
"(",
"M",
",",
"np",
".",
"array",
"(",
"[",
"X",
"[",
"i",
"]",
"[",
"j",
"]",
",",
"Y",
"[",
"i",
"]",
"[",
"j",
"]",
",",
"Z",
"[",
"i",
"]",
"[",
"j",
"]",
"]",
")",
")",
"+",
"C_fit",
"X",
"[",
"i",
"]",
"[",
"j",
"]",
"=",
"p",
"[",
"0",
"]",
"Y",
"[",
"i",
"]",
"[",
"j",
"]",
"=",
"p",
"[",
"1",
"]",
"Z",
"[",
"i",
"]",
"[",
"j",
"]",
"=",
"p",
"[",
"2",
"]",
"ax",
".",
"plot_surface",
"(",
"X",
",",
"Y",
",",
"Z",
",",
"alpha",
"=",
"0.2",
")",
"# Plot the center and direction",
"ax",
".",
"quiver",
"(",
"C_fit",
"[",
"0",
"]",
",",
"C_fit",
"[",
"1",
"]",
",",
"C_fit",
"[",
"2",
"]",
",",
"r_fit",
"*",
"w_fit",
"[",
"0",
"]",
",",
"r_fit",
"*",
"w_fit",
"[",
"1",
"]",
",",
"r_fit",
"*",
"w_fit",
"[",
"2",
"]",
",",
"color",
"=",
"'red'",
")",
"plt",
".",
"show",
"(",
")"
] |
Plot the fitting given the fitted axis direction, the fitted
center, the fitted radius and the data points.
|
[
"Plot",
"the",
"fitting",
"given",
"the",
"fitted",
"axis",
"direction",
"the",
"fitted",
"center",
"the",
"fitted",
"radius",
"and",
"the",
"data",
"points",
"."
] |
f96d79732bc49cbc0cb4b39f008af7ce42aeb213
|
https://github.com/xingjiepan/cylinder_fitting/blob/f96d79732bc49cbc0cb4b39f008af7ce42aeb213/cylinder_fitting/visualize.py#L27-L72
|
train
|
colab/colab
|
colab/utils/highlighting.py
|
ColabHighlighter.find_window
|
def find_window(self, highlight_locations):
"""Getting the HIGHLIGHT_NUM_CHARS_BEFORE_MATCH setting
to find how many characters before the first word found should
be removed from the window
"""
if len(self.text_block) <= self.max_length:
return (0, self.max_length)
num_chars_before = getattr(
settings,
'HIGHLIGHT_NUM_CHARS_BEFORE_MATCH',
0
)
best_start, best_end = super(ColabHighlighter, self).find_window(
highlight_locations
)
if best_start <= num_chars_before:
best_end -= best_start
best_start = 0
else:
best_start -= num_chars_before
best_end -= num_chars_before
return (best_start, best_end)
|
python
|
def find_window(self, highlight_locations):
"""Getting the HIGHLIGHT_NUM_CHARS_BEFORE_MATCH setting
to find how many characters before the first word found should
be removed from the window
"""
if len(self.text_block) <= self.max_length:
return (0, self.max_length)
num_chars_before = getattr(
settings,
'HIGHLIGHT_NUM_CHARS_BEFORE_MATCH',
0
)
best_start, best_end = super(ColabHighlighter, self).find_window(
highlight_locations
)
if best_start <= num_chars_before:
best_end -= best_start
best_start = 0
else:
best_start -= num_chars_before
best_end -= num_chars_before
return (best_start, best_end)
|
[
"def",
"find_window",
"(",
"self",
",",
"highlight_locations",
")",
":",
"if",
"len",
"(",
"self",
".",
"text_block",
")",
"<=",
"self",
".",
"max_length",
":",
"return",
"(",
"0",
",",
"self",
".",
"max_length",
")",
"num_chars_before",
"=",
"getattr",
"(",
"settings",
",",
"'HIGHLIGHT_NUM_CHARS_BEFORE_MATCH'",
",",
"0",
")",
"best_start",
",",
"best_end",
"=",
"super",
"(",
"ColabHighlighter",
",",
"self",
")",
".",
"find_window",
"(",
"highlight_locations",
")",
"if",
"best_start",
"<=",
"num_chars_before",
":",
"best_end",
"-=",
"best_start",
"best_start",
"=",
"0",
"else",
":",
"best_start",
"-=",
"num_chars_before",
"best_end",
"-=",
"num_chars_before",
"return",
"(",
"best_start",
",",
"best_end",
")"
] |
Getting the HIGHLIGHT_NUM_CHARS_BEFORE_MATCH setting
to find how many characters before the first word found should
be removed from the window
|
[
"Getting",
"the",
"HIGHLIGHT_NUM_CHARS_BEFORE_MATCH",
"setting",
"to",
"find",
"how",
"many",
"characters",
"before",
"the",
"first",
"word",
"found",
"should",
"be",
"removed",
"from",
"the",
"window"
] |
2ad099231e620bec647363b27d38006eca71e13b
|
https://github.com/colab/colab/blob/2ad099231e620bec647363b27d38006eca71e13b/colab/utils/highlighting.py#L13-L38
|
train
|
colab/colab
|
colab/middlewares/cookie_middleware.py
|
CookieHandler.__set
|
def __set(self, key, real_value, coded_value):
"""Private method for setting a cookie's value"""
morse_set = self.get(key, StringMorsel())
morse_set.set(key, real_value, coded_value)
dict.__setitem__(self, key, morse_set)
|
python
|
def __set(self, key, real_value, coded_value):
"""Private method for setting a cookie's value"""
morse_set = self.get(key, StringMorsel())
morse_set.set(key, real_value, coded_value)
dict.__setitem__(self, key, morse_set)
|
[
"def",
"__set",
"(",
"self",
",",
"key",
",",
"real_value",
",",
"coded_value",
")",
":",
"morse_set",
"=",
"self",
".",
"get",
"(",
"key",
",",
"StringMorsel",
"(",
")",
")",
"morse_set",
".",
"set",
"(",
"key",
",",
"real_value",
",",
"coded_value",
")",
"dict",
".",
"__setitem__",
"(",
"self",
",",
"key",
",",
"morse_set",
")"
] |
Private method for setting a cookie's value
|
[
"Private",
"method",
"for",
"setting",
"a",
"cookie",
"s",
"value"
] |
2ad099231e620bec647363b27d38006eca71e13b
|
https://github.com/colab/colab/blob/2ad099231e620bec647363b27d38006eca71e13b/colab/middlewares/cookie_middleware.py#L67-L71
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.login
|
def login(self):
"""
Try to login and set the internal session id.
Please note:
- Any failed login resets all existing session ids, even of
other users.
- SIDs expire after some time
"""
response = self.session.get(self.base_url + '/login_sid.lua', timeout=10)
xml = ET.fromstring(response.text)
if xml.find('SID').text == "0000000000000000":
challenge = xml.find('Challenge').text
url = self.base_url + "/login_sid.lua"
response = self.session.get(url, params={
"username": self.username,
"response": self.calculate_response(challenge, self.password),
}, timeout=10)
xml = ET.fromstring(response.text)
sid = xml.find('SID').text
if xml.find('SID').text == "0000000000000000":
blocktime = int(xml.find('BlockTime').text)
exc = Exception("Login failed, please wait {} seconds".format(
blocktime
))
exc.blocktime = blocktime
raise exc
self.sid = sid
return sid
|
python
|
def login(self):
"""
Try to login and set the internal session id.
Please note:
- Any failed login resets all existing session ids, even of
other users.
- SIDs expire after some time
"""
response = self.session.get(self.base_url + '/login_sid.lua', timeout=10)
xml = ET.fromstring(response.text)
if xml.find('SID').text == "0000000000000000":
challenge = xml.find('Challenge').text
url = self.base_url + "/login_sid.lua"
response = self.session.get(url, params={
"username": self.username,
"response": self.calculate_response(challenge, self.password),
}, timeout=10)
xml = ET.fromstring(response.text)
sid = xml.find('SID').text
if xml.find('SID').text == "0000000000000000":
blocktime = int(xml.find('BlockTime').text)
exc = Exception("Login failed, please wait {} seconds".format(
blocktime
))
exc.blocktime = blocktime
raise exc
self.sid = sid
return sid
|
[
"def",
"login",
"(",
"self",
")",
":",
"response",
"=",
"self",
".",
"session",
".",
"get",
"(",
"self",
".",
"base_url",
"+",
"'/login_sid.lua'",
",",
"timeout",
"=",
"10",
")",
"xml",
"=",
"ET",
".",
"fromstring",
"(",
"response",
".",
"text",
")",
"if",
"xml",
".",
"find",
"(",
"'SID'",
")",
".",
"text",
"==",
"\"0000000000000000\"",
":",
"challenge",
"=",
"xml",
".",
"find",
"(",
"'Challenge'",
")",
".",
"text",
"url",
"=",
"self",
".",
"base_url",
"+",
"\"/login_sid.lua\"",
"response",
"=",
"self",
".",
"session",
".",
"get",
"(",
"url",
",",
"params",
"=",
"{",
"\"username\"",
":",
"self",
".",
"username",
",",
"\"response\"",
":",
"self",
".",
"calculate_response",
"(",
"challenge",
",",
"self",
".",
"password",
")",
",",
"}",
",",
"timeout",
"=",
"10",
")",
"xml",
"=",
"ET",
".",
"fromstring",
"(",
"response",
".",
"text",
")",
"sid",
"=",
"xml",
".",
"find",
"(",
"'SID'",
")",
".",
"text",
"if",
"xml",
".",
"find",
"(",
"'SID'",
")",
".",
"text",
"==",
"\"0000000000000000\"",
":",
"blocktime",
"=",
"int",
"(",
"xml",
".",
"find",
"(",
"'BlockTime'",
")",
".",
"text",
")",
"exc",
"=",
"Exception",
"(",
"\"Login failed, please wait {} seconds\"",
".",
"format",
"(",
"blocktime",
")",
")",
"exc",
".",
"blocktime",
"=",
"blocktime",
"raise",
"exc",
"self",
".",
"sid",
"=",
"sid",
"return",
"sid"
] |
Try to login and set the internal session id.
Please note:
- Any failed login resets all existing session ids, even of
other users.
- SIDs expire after some time
|
[
"Try",
"to",
"login",
"and",
"set",
"the",
"internal",
"session",
"id",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L53-L81
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.calculate_response
|
def calculate_response(self, challenge, password):
"""Calculate response for the challenge-response authentication"""
to_hash = (challenge + "-" + password).encode("UTF-16LE")
hashed = hashlib.md5(to_hash).hexdigest()
return "{0}-{1}".format(challenge, hashed)
|
python
|
def calculate_response(self, challenge, password):
"""Calculate response for the challenge-response authentication"""
to_hash = (challenge + "-" + password).encode("UTF-16LE")
hashed = hashlib.md5(to_hash).hexdigest()
return "{0}-{1}".format(challenge, hashed)
|
[
"def",
"calculate_response",
"(",
"self",
",",
"challenge",
",",
"password",
")",
":",
"to_hash",
"=",
"(",
"challenge",
"+",
"\"-\"",
"+",
"password",
")",
".",
"encode",
"(",
"\"UTF-16LE\"",
")",
"hashed",
"=",
"hashlib",
".",
"md5",
"(",
"to_hash",
")",
".",
"hexdigest",
"(",
")",
"return",
"\"{0}-{1}\"",
".",
"format",
"(",
"challenge",
",",
"hashed",
")"
] |
Calculate response for the challenge-response authentication
|
[
"Calculate",
"response",
"for",
"the",
"challenge",
"-",
"response",
"authentication"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L83-L87
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.get_actors
|
def get_actors(self):
"""
Returns a list of Actor objects for querying SmartHome devices.
This is currently the only working method for getting temperature data.
"""
devices = self.homeautoswitch("getdevicelistinfos")
xml = ET.fromstring(devices)
actors = []
for device in xml.findall('device'):
actors.append(Actor(fritzbox=self, device=device))
return actors
|
python
|
def get_actors(self):
"""
Returns a list of Actor objects for querying SmartHome devices.
This is currently the only working method for getting temperature data.
"""
devices = self.homeautoswitch("getdevicelistinfos")
xml = ET.fromstring(devices)
actors = []
for device in xml.findall('device'):
actors.append(Actor(fritzbox=self, device=device))
return actors
|
[
"def",
"get_actors",
"(",
"self",
")",
":",
"devices",
"=",
"self",
".",
"homeautoswitch",
"(",
"\"getdevicelistinfos\"",
")",
"xml",
"=",
"ET",
".",
"fromstring",
"(",
"devices",
")",
"actors",
"=",
"[",
"]",
"for",
"device",
"in",
"xml",
".",
"findall",
"(",
"'device'",
")",
":",
"actors",
".",
"append",
"(",
"Actor",
"(",
"fritzbox",
"=",
"self",
",",
"device",
"=",
"device",
")",
")",
"return",
"actors"
] |
Returns a list of Actor objects for querying SmartHome devices.
This is currently the only working method for getting temperature data.
|
[
"Returns",
"a",
"list",
"of",
"Actor",
"objects",
"for",
"querying",
"SmartHome",
"devices",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L93-L106
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.get_actor_by_ain
|
def get_actor_by_ain(self, ain):
"""
Return a actor identified by it's ain or return None
"""
for actor in self.get_actors():
if actor.actor_id == ain:
return actor
|
python
|
def get_actor_by_ain(self, ain):
"""
Return a actor identified by it's ain or return None
"""
for actor in self.get_actors():
if actor.actor_id == ain:
return actor
|
[
"def",
"get_actor_by_ain",
"(",
"self",
",",
"ain",
")",
":",
"for",
"actor",
"in",
"self",
".",
"get_actors",
"(",
")",
":",
"if",
"actor",
".",
"actor_id",
"==",
"ain",
":",
"return",
"actor"
] |
Return a actor identified by it's ain or return None
|
[
"Return",
"a",
"actor",
"identified",
"by",
"it",
"s",
"ain",
"or",
"return",
"None"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L108-L114
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.homeautoswitch
|
def homeautoswitch(self, cmd, ain=None, param=None):
"""
Call a switch method.
Should only be used by internal library functions.
"""
assert self.sid, "Not logged in"
params = {
'switchcmd': cmd,
'sid': self.sid,
}
if param is not None:
params['param'] = param
if ain:
params['ain'] = ain
url = self.base_url + '/webservices/homeautoswitch.lua'
response = self.session.get(url, params=params, timeout=10)
response.raise_for_status()
return response.text.strip().encode('utf-8')
|
python
|
def homeautoswitch(self, cmd, ain=None, param=None):
"""
Call a switch method.
Should only be used by internal library functions.
"""
assert self.sid, "Not logged in"
params = {
'switchcmd': cmd,
'sid': self.sid,
}
if param is not None:
params['param'] = param
if ain:
params['ain'] = ain
url = self.base_url + '/webservices/homeautoswitch.lua'
response = self.session.get(url, params=params, timeout=10)
response.raise_for_status()
return response.text.strip().encode('utf-8')
|
[
"def",
"homeautoswitch",
"(",
"self",
",",
"cmd",
",",
"ain",
"=",
"None",
",",
"param",
"=",
"None",
")",
":",
"assert",
"self",
".",
"sid",
",",
"\"Not logged in\"",
"params",
"=",
"{",
"'switchcmd'",
":",
"cmd",
",",
"'sid'",
":",
"self",
".",
"sid",
",",
"}",
"if",
"param",
"is",
"not",
"None",
":",
"params",
"[",
"'param'",
"]",
"=",
"param",
"if",
"ain",
":",
"params",
"[",
"'ain'",
"]",
"=",
"ain",
"url",
"=",
"self",
".",
"base_url",
"+",
"'/webservices/homeautoswitch.lua'",
"response",
"=",
"self",
".",
"session",
".",
"get",
"(",
"url",
",",
"params",
"=",
"params",
",",
"timeout",
"=",
"10",
")",
"response",
".",
"raise_for_status",
"(",
")",
"return",
"response",
".",
"text",
".",
"strip",
"(",
")",
".",
"encode",
"(",
"'utf-8'",
")"
] |
Call a switch method.
Should only be used by internal library functions.
|
[
"Call",
"a",
"switch",
"method",
".",
"Should",
"only",
"be",
"used",
"by",
"internal",
"library",
"functions",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L120-L138
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.get_switch_actors
|
def get_switch_actors(self):
"""
Get information about all actors
This needs 1+(5n) requests where n = number of actors registered
Deprecated, use get_actors instead.
Returns a dict:
[ain] = {
'name': Name of actor,
'state': Powerstate (boolean)
'present': Connected to server? (boolean)
'power': Current power consumption in mW
'energy': Used energy in Wh since last energy reset
'temperature': Current environment temperature in celsius
}
"""
actors = {}
for ain in self.homeautoswitch("getswitchlist").split(','):
actors[ain] = {
'name': self.homeautoswitch("getswitchname", ain),
'state': bool(self.homeautoswitch("getswitchstate", ain)),
'present': bool(self.homeautoswitch("getswitchpresent", ain)),
'power': self.homeautoswitch("getswitchpower", ain),
'energy': self.homeautoswitch("getswitchenergy", ain),
'temperature': self.homeautoswitch("getswitchtemperature", ain),
}
return actors
|
python
|
def get_switch_actors(self):
"""
Get information about all actors
This needs 1+(5n) requests where n = number of actors registered
Deprecated, use get_actors instead.
Returns a dict:
[ain] = {
'name': Name of actor,
'state': Powerstate (boolean)
'present': Connected to server? (boolean)
'power': Current power consumption in mW
'energy': Used energy in Wh since last energy reset
'temperature': Current environment temperature in celsius
}
"""
actors = {}
for ain in self.homeautoswitch("getswitchlist").split(','):
actors[ain] = {
'name': self.homeautoswitch("getswitchname", ain),
'state': bool(self.homeautoswitch("getswitchstate", ain)),
'present': bool(self.homeautoswitch("getswitchpresent", ain)),
'power': self.homeautoswitch("getswitchpower", ain),
'energy': self.homeautoswitch("getswitchenergy", ain),
'temperature': self.homeautoswitch("getswitchtemperature", ain),
}
return actors
|
[
"def",
"get_switch_actors",
"(",
"self",
")",
":",
"actors",
"=",
"{",
"}",
"for",
"ain",
"in",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchlist\"",
")",
".",
"split",
"(",
"','",
")",
":",
"actors",
"[",
"ain",
"]",
"=",
"{",
"'name'",
":",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchname\"",
",",
"ain",
")",
",",
"'state'",
":",
"bool",
"(",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchstate\"",
",",
"ain",
")",
")",
",",
"'present'",
":",
"bool",
"(",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchpresent\"",
",",
"ain",
")",
")",
",",
"'power'",
":",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchpower\"",
",",
"ain",
")",
",",
"'energy'",
":",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchenergy\"",
",",
"ain",
")",
",",
"'temperature'",
":",
"self",
".",
"homeautoswitch",
"(",
"\"getswitchtemperature\"",
",",
"ain",
")",
",",
"}",
"return",
"actors"
] |
Get information about all actors
This needs 1+(5n) requests where n = number of actors registered
Deprecated, use get_actors instead.
Returns a dict:
[ain] = {
'name': Name of actor,
'state': Powerstate (boolean)
'present': Connected to server? (boolean)
'power': Current power consumption in mW
'energy': Used energy in Wh since last energy reset
'temperature': Current environment temperature in celsius
}
|
[
"Get",
"information",
"about",
"all",
"actors"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L140-L168
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.get_devices
|
def get_devices(self):
"""
Return a list of devices.
Deprecated, use get_actors instead.
"""
url = self.base_url + '/net/home_auto_query.lua'
response = self.session.get(url, params={
'sid': self.sid,
'command': 'AllOutletStates',
'xhr': 0,
}, timeout=15)
response.raise_for_status()
data = response.json()
count = int(data["Outlet_count"])
devices = []
for i in range(1, count + 1):
device = Device(
int(data["DeviceID_{0}".format(i)]),
int(data["DeviceConnectState_{0}".format(i)]),
int(data["DeviceSwitchState_{0}".format(i)])
)
devices.append(device)
return devices
|
python
|
def get_devices(self):
"""
Return a list of devices.
Deprecated, use get_actors instead.
"""
url = self.base_url + '/net/home_auto_query.lua'
response = self.session.get(url, params={
'sid': self.sid,
'command': 'AllOutletStates',
'xhr': 0,
}, timeout=15)
response.raise_for_status()
data = response.json()
count = int(data["Outlet_count"])
devices = []
for i in range(1, count + 1):
device = Device(
int(data["DeviceID_{0}".format(i)]),
int(data["DeviceConnectState_{0}".format(i)]),
int(data["DeviceSwitchState_{0}".format(i)])
)
devices.append(device)
return devices
|
[
"def",
"get_devices",
"(",
"self",
")",
":",
"url",
"=",
"self",
".",
"base_url",
"+",
"'/net/home_auto_query.lua'",
"response",
"=",
"self",
".",
"session",
".",
"get",
"(",
"url",
",",
"params",
"=",
"{",
"'sid'",
":",
"self",
".",
"sid",
",",
"'command'",
":",
"'AllOutletStates'",
",",
"'xhr'",
":",
"0",
",",
"}",
",",
"timeout",
"=",
"15",
")",
"response",
".",
"raise_for_status",
"(",
")",
"data",
"=",
"response",
".",
"json",
"(",
")",
"count",
"=",
"int",
"(",
"data",
"[",
"\"Outlet_count\"",
"]",
")",
"devices",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"count",
"+",
"1",
")",
":",
"device",
"=",
"Device",
"(",
"int",
"(",
"data",
"[",
"\"DeviceID_{0}\"",
".",
"format",
"(",
"i",
")",
"]",
")",
",",
"int",
"(",
"data",
"[",
"\"DeviceConnectState_{0}\"",
".",
"format",
"(",
"i",
")",
"]",
")",
",",
"int",
"(",
"data",
"[",
"\"DeviceSwitchState_{0}\"",
".",
"format",
"(",
"i",
")",
"]",
")",
")",
"devices",
".",
"append",
"(",
"device",
")",
"return",
"devices"
] |
Return a list of devices.
Deprecated, use get_actors instead.
|
[
"Return",
"a",
"list",
"of",
"devices",
".",
"Deprecated",
"use",
"get_actors",
"instead",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L189-L211
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.get_consumption
|
def get_consumption(self, deviceid, timerange="10"):
"""
Return all available energy consumption data for the device.
You need to divice watt_values by 100 and volt_values by 1000
to get the "real" values.
:return: dict
"""
tranges = ("10", "24h", "month", "year")
if timerange not in tranges:
raise ValueError(
"Unknown timerange. Possible values are: {0}".format(tranges)
)
url = self.base_url + "/net/home_auto_query.lua"
response = self.session.get(url, params={
'sid': self.sid,
'command': 'EnergyStats_{0}'.format(timerange),
'id': deviceid,
'xhr': 0,
}, timeout=15)
response.raise_for_status()
data = response.json()
result = {}
# Single result values
values_map = {
'MM_Value_Amp': 'mm_value_amp',
'MM_Value_Power': 'mm_value_power',
'MM_Value_Volt': 'mm_value_volt',
'EnStats_average_value': 'enstats_average_value',
'EnStats_max_value': 'enstats_max_value',
'EnStats_min_value': 'enstats_min_value',
'EnStats_timer_type': 'enstats_timer_type',
'sum_Day': 'sum_day',
'sum_Month': 'sum_month',
'sum_Year': 'sum_year',
}
for avm_key, py_key in values_map.items():
result[py_key] = int(data[avm_key])
# Stats counts
count = int(data["EnStats_count"])
watt_values = [None for i in range(count)]
volt_values = [None for i in range(count)]
for i in range(1, count + 1):
watt_values[i - 1] = int(data["EnStats_watt_value_{}".format(i)])
volt_values[i - 1] = int(data["EnStats_volt_value_{}".format(i)])
result['watt_values'] = watt_values
result['volt_values'] = volt_values
return result
|
python
|
def get_consumption(self, deviceid, timerange="10"):
"""
Return all available energy consumption data for the device.
You need to divice watt_values by 100 and volt_values by 1000
to get the "real" values.
:return: dict
"""
tranges = ("10", "24h", "month", "year")
if timerange not in tranges:
raise ValueError(
"Unknown timerange. Possible values are: {0}".format(tranges)
)
url = self.base_url + "/net/home_auto_query.lua"
response = self.session.get(url, params={
'sid': self.sid,
'command': 'EnergyStats_{0}'.format(timerange),
'id': deviceid,
'xhr': 0,
}, timeout=15)
response.raise_for_status()
data = response.json()
result = {}
# Single result values
values_map = {
'MM_Value_Amp': 'mm_value_amp',
'MM_Value_Power': 'mm_value_power',
'MM_Value_Volt': 'mm_value_volt',
'EnStats_average_value': 'enstats_average_value',
'EnStats_max_value': 'enstats_max_value',
'EnStats_min_value': 'enstats_min_value',
'EnStats_timer_type': 'enstats_timer_type',
'sum_Day': 'sum_day',
'sum_Month': 'sum_month',
'sum_Year': 'sum_year',
}
for avm_key, py_key in values_map.items():
result[py_key] = int(data[avm_key])
# Stats counts
count = int(data["EnStats_count"])
watt_values = [None for i in range(count)]
volt_values = [None for i in range(count)]
for i in range(1, count + 1):
watt_values[i - 1] = int(data["EnStats_watt_value_{}".format(i)])
volt_values[i - 1] = int(data["EnStats_volt_value_{}".format(i)])
result['watt_values'] = watt_values
result['volt_values'] = volt_values
return result
|
[
"def",
"get_consumption",
"(",
"self",
",",
"deviceid",
",",
"timerange",
"=",
"\"10\"",
")",
":",
"tranges",
"=",
"(",
"\"10\"",
",",
"\"24h\"",
",",
"\"month\"",
",",
"\"year\"",
")",
"if",
"timerange",
"not",
"in",
"tranges",
":",
"raise",
"ValueError",
"(",
"\"Unknown timerange. Possible values are: {0}\"",
".",
"format",
"(",
"tranges",
")",
")",
"url",
"=",
"self",
".",
"base_url",
"+",
"\"/net/home_auto_query.lua\"",
"response",
"=",
"self",
".",
"session",
".",
"get",
"(",
"url",
",",
"params",
"=",
"{",
"'sid'",
":",
"self",
".",
"sid",
",",
"'command'",
":",
"'EnergyStats_{0}'",
".",
"format",
"(",
"timerange",
")",
",",
"'id'",
":",
"deviceid",
",",
"'xhr'",
":",
"0",
",",
"}",
",",
"timeout",
"=",
"15",
")",
"response",
".",
"raise_for_status",
"(",
")",
"data",
"=",
"response",
".",
"json",
"(",
")",
"result",
"=",
"{",
"}",
"# Single result values",
"values_map",
"=",
"{",
"'MM_Value_Amp'",
":",
"'mm_value_amp'",
",",
"'MM_Value_Power'",
":",
"'mm_value_power'",
",",
"'MM_Value_Volt'",
":",
"'mm_value_volt'",
",",
"'EnStats_average_value'",
":",
"'enstats_average_value'",
",",
"'EnStats_max_value'",
":",
"'enstats_max_value'",
",",
"'EnStats_min_value'",
":",
"'enstats_min_value'",
",",
"'EnStats_timer_type'",
":",
"'enstats_timer_type'",
",",
"'sum_Day'",
":",
"'sum_day'",
",",
"'sum_Month'",
":",
"'sum_month'",
",",
"'sum_Year'",
":",
"'sum_year'",
",",
"}",
"for",
"avm_key",
",",
"py_key",
"in",
"values_map",
".",
"items",
"(",
")",
":",
"result",
"[",
"py_key",
"]",
"=",
"int",
"(",
"data",
"[",
"avm_key",
"]",
")",
"# Stats counts",
"count",
"=",
"int",
"(",
"data",
"[",
"\"EnStats_count\"",
"]",
")",
"watt_values",
"=",
"[",
"None",
"for",
"i",
"in",
"range",
"(",
"count",
")",
"]",
"volt_values",
"=",
"[",
"None",
"for",
"i",
"in",
"range",
"(",
"count",
")",
"]",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"count",
"+",
"1",
")",
":",
"watt_values",
"[",
"i",
"-",
"1",
"]",
"=",
"int",
"(",
"data",
"[",
"\"EnStats_watt_value_{}\"",
".",
"format",
"(",
"i",
")",
"]",
")",
"volt_values",
"[",
"i",
"-",
"1",
"]",
"=",
"int",
"(",
"data",
"[",
"\"EnStats_volt_value_{}\"",
".",
"format",
"(",
"i",
")",
"]",
")",
"result",
"[",
"'watt_values'",
"]",
"=",
"watt_values",
"result",
"[",
"'volt_values'",
"]",
"=",
"volt_values",
"return",
"result"
] |
Return all available energy consumption data for the device.
You need to divice watt_values by 100 and volt_values by 1000
to get the "real" values.
:return: dict
|
[
"Return",
"all",
"available",
"energy",
"consumption",
"data",
"for",
"the",
"device",
".",
"You",
"need",
"to",
"divice",
"watt_values",
"by",
"100",
"and",
"volt_values",
"by",
"1000",
"to",
"get",
"the",
"real",
"values",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L213-L268
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/fritz.py
|
FritzBox.get_logs
|
def get_logs(self):
"""
Return the system logs since the last reboot.
"""
assert BeautifulSoup, "Please install bs4 to use this method"
url = self.base_url + "/system/syslog.lua"
response = self.session.get(url, params={
'sid': self.sid,
'stylemode': 'print',
}, timeout=15)
response.raise_for_status()
entries = []
tree = BeautifulSoup(response.text)
rows = tree.find('table').find_all('tr')
for row in rows:
columns = row.find_all("td")
date = columns[0].string
time = columns[1].string
message = columns[2].find("a").string
merged = "{} {} {}".format(date, time, message.encode("UTF-8"))
msg_hash = hashlib.md5(merged).hexdigest()
entries.append(LogEntry(date, time, message, msg_hash))
return entries
|
python
|
def get_logs(self):
"""
Return the system logs since the last reboot.
"""
assert BeautifulSoup, "Please install bs4 to use this method"
url = self.base_url + "/system/syslog.lua"
response = self.session.get(url, params={
'sid': self.sid,
'stylemode': 'print',
}, timeout=15)
response.raise_for_status()
entries = []
tree = BeautifulSoup(response.text)
rows = tree.find('table').find_all('tr')
for row in rows:
columns = row.find_all("td")
date = columns[0].string
time = columns[1].string
message = columns[2].find("a").string
merged = "{} {} {}".format(date, time, message.encode("UTF-8"))
msg_hash = hashlib.md5(merged).hexdigest()
entries.append(LogEntry(date, time, message, msg_hash))
return entries
|
[
"def",
"get_logs",
"(",
"self",
")",
":",
"assert",
"BeautifulSoup",
",",
"\"Please install bs4 to use this method\"",
"url",
"=",
"self",
".",
"base_url",
"+",
"\"/system/syslog.lua\"",
"response",
"=",
"self",
".",
"session",
".",
"get",
"(",
"url",
",",
"params",
"=",
"{",
"'sid'",
":",
"self",
".",
"sid",
",",
"'stylemode'",
":",
"'print'",
",",
"}",
",",
"timeout",
"=",
"15",
")",
"response",
".",
"raise_for_status",
"(",
")",
"entries",
"=",
"[",
"]",
"tree",
"=",
"BeautifulSoup",
"(",
"response",
".",
"text",
")",
"rows",
"=",
"tree",
".",
"find",
"(",
"'table'",
")",
".",
"find_all",
"(",
"'tr'",
")",
"for",
"row",
"in",
"rows",
":",
"columns",
"=",
"row",
".",
"find_all",
"(",
"\"td\"",
")",
"date",
"=",
"columns",
"[",
"0",
"]",
".",
"string",
"time",
"=",
"columns",
"[",
"1",
"]",
".",
"string",
"message",
"=",
"columns",
"[",
"2",
"]",
".",
"find",
"(",
"\"a\"",
")",
".",
"string",
"merged",
"=",
"\"{} {} {}\"",
".",
"format",
"(",
"date",
",",
"time",
",",
"message",
".",
"encode",
"(",
"\"UTF-8\"",
")",
")",
"msg_hash",
"=",
"hashlib",
".",
"md5",
"(",
"merged",
")",
".",
"hexdigest",
"(",
")",
"entries",
".",
"append",
"(",
"LogEntry",
"(",
"date",
",",
"time",
",",
"message",
",",
"msg_hash",
")",
")",
"return",
"entries"
] |
Return the system logs since the last reboot.
|
[
"Return",
"the",
"system",
"logs",
"since",
"the",
"last",
"reboot",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/fritz.py#L270-L295
|
train
|
kumar303/hawkrest
|
hawkrest/__init__.py
|
seen_nonce
|
def seen_nonce(id, nonce, timestamp):
"""
Returns True if the Hawk nonce has been seen already.
"""
key = '{id}:{n}:{ts}'.format(id=id, n=nonce, ts=timestamp)
if cache.get(key):
log.warning('replay attack? already processed nonce {k}'
.format(k=key))
return True
else:
log.debug('caching nonce {k}'.format(k=key))
cache.set(key, True,
# We only need the nonce until the message itself expires.
# This also adds a little bit of padding.
timeout=getattr(settings, 'HAWK_MESSAGE_EXPIRATION',
default_message_expiration) + 5)
return False
|
python
|
def seen_nonce(id, nonce, timestamp):
"""
Returns True if the Hawk nonce has been seen already.
"""
key = '{id}:{n}:{ts}'.format(id=id, n=nonce, ts=timestamp)
if cache.get(key):
log.warning('replay attack? already processed nonce {k}'
.format(k=key))
return True
else:
log.debug('caching nonce {k}'.format(k=key))
cache.set(key, True,
# We only need the nonce until the message itself expires.
# This also adds a little bit of padding.
timeout=getattr(settings, 'HAWK_MESSAGE_EXPIRATION',
default_message_expiration) + 5)
return False
|
[
"def",
"seen_nonce",
"(",
"id",
",",
"nonce",
",",
"timestamp",
")",
":",
"key",
"=",
"'{id}:{n}:{ts}'",
".",
"format",
"(",
"id",
"=",
"id",
",",
"n",
"=",
"nonce",
",",
"ts",
"=",
"timestamp",
")",
"if",
"cache",
".",
"get",
"(",
"key",
")",
":",
"log",
".",
"warning",
"(",
"'replay attack? already processed nonce {k}'",
".",
"format",
"(",
"k",
"=",
"key",
")",
")",
"return",
"True",
"else",
":",
"log",
".",
"debug",
"(",
"'caching nonce {k}'",
".",
"format",
"(",
"k",
"=",
"key",
")",
")",
"cache",
".",
"set",
"(",
"key",
",",
"True",
",",
"# We only need the nonce until the message itself expires.",
"# This also adds a little bit of padding.",
"timeout",
"=",
"getattr",
"(",
"settings",
",",
"'HAWK_MESSAGE_EXPIRATION'",
",",
"default_message_expiration",
")",
"+",
"5",
")",
"return",
"False"
] |
Returns True if the Hawk nonce has been seen already.
|
[
"Returns",
"True",
"if",
"the",
"Hawk",
"nonce",
"has",
"been",
"seen",
"already",
"."
] |
df4de1a068dc2624bf716ed244ab55e8ad4fef7e
|
https://github.com/kumar303/hawkrest/blob/df4de1a068dc2624bf716ed244ab55e8ad4fef7e/hawkrest/__init__.py#L187-L203
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
cli
|
def cli(context, host, username, password):
"""
FritzBox SmartHome Tool
\b
Provides the following functions:
- A easy to use library for querying SmartHome actors
- This CLI tool for testing
- A carbon client for pipeing data into graphite
"""
context.obj = FritzBox(host, username, password)
|
python
|
def cli(context, host, username, password):
"""
FritzBox SmartHome Tool
\b
Provides the following functions:
- A easy to use library for querying SmartHome actors
- This CLI tool for testing
- A carbon client for pipeing data into graphite
"""
context.obj = FritzBox(host, username, password)
|
[
"def",
"cli",
"(",
"context",
",",
"host",
",",
"username",
",",
"password",
")",
":",
"context",
".",
"obj",
"=",
"FritzBox",
"(",
"host",
",",
"username",
",",
"password",
")"
] |
FritzBox SmartHome Tool
\b
Provides the following functions:
- A easy to use library for querying SmartHome actors
- This CLI tool for testing
- A carbon client for pipeing data into graphite
|
[
"FritzBox",
"SmartHome",
"Tool"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L25-L35
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
actors
|
def actors(context):
"""Display a list of actors"""
fritz = context.obj
fritz.login()
for actor in fritz.get_actors():
click.echo("{} ({} {}; AIN {} )".format(
actor.name,
actor.manufacturer,
actor.productname,
actor.actor_id,
))
if actor.has_temperature:
click.echo("Temp: act {} target {}; battery (low): {}".format(
actor.temperature,
actor.target_temperature,
actor.battery_low,
))
click.echo("Temp (via get): act {} target {}".format(
actor.get_temperature(),
actor.get_target_temperature(),
))
|
python
|
def actors(context):
"""Display a list of actors"""
fritz = context.obj
fritz.login()
for actor in fritz.get_actors():
click.echo("{} ({} {}; AIN {} )".format(
actor.name,
actor.manufacturer,
actor.productname,
actor.actor_id,
))
if actor.has_temperature:
click.echo("Temp: act {} target {}; battery (low): {}".format(
actor.temperature,
actor.target_temperature,
actor.battery_low,
))
click.echo("Temp (via get): act {} target {}".format(
actor.get_temperature(),
actor.get_target_temperature(),
))
|
[
"def",
"actors",
"(",
"context",
")",
":",
"fritz",
"=",
"context",
".",
"obj",
"fritz",
".",
"login",
"(",
")",
"for",
"actor",
"in",
"fritz",
".",
"get_actors",
"(",
")",
":",
"click",
".",
"echo",
"(",
"\"{} ({} {}; AIN {} )\"",
".",
"format",
"(",
"actor",
".",
"name",
",",
"actor",
".",
"manufacturer",
",",
"actor",
".",
"productname",
",",
"actor",
".",
"actor_id",
",",
")",
")",
"if",
"actor",
".",
"has_temperature",
":",
"click",
".",
"echo",
"(",
"\"Temp: act {} target {}; battery (low): {}\"",
".",
"format",
"(",
"actor",
".",
"temperature",
",",
"actor",
".",
"target_temperature",
",",
"actor",
".",
"battery_low",
",",
")",
")",
"click",
".",
"echo",
"(",
"\"Temp (via get): act {} target {}\"",
".",
"format",
"(",
"actor",
".",
"get_temperature",
"(",
")",
",",
"actor",
".",
"get_target_temperature",
"(",
")",
",",
")",
")"
] |
Display a list of actors
|
[
"Display",
"a",
"list",
"of",
"actors"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L40-L62
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
energy
|
def energy(context, features):
"""Display energy stats of all actors"""
fritz = context.obj
fritz.login()
for actor in fritz.get_actors():
if actor.temperature is not None:
click.echo("{} ({}): {:.2f} Watt current, {:.3f} wH total, {:.2f} °C".format(
actor.name.encode('utf-8'),
actor.actor_id,
(actor.get_power() or 0.0) / 1000,
(actor.get_energy() or 0.0) / 100,
actor.temperature
))
else:
click.echo("{} ({}): {:.2f} Watt current, {:.3f} wH total, offline".format(
actor.name.encode('utf-8'),
actor.actor_id,
(actor.get_power() or 0.0) / 1000,
(actor.get_energy() or 0.0) / 100
))
if features:
click.echo(" Features: PowerMeter: {}, Temperatur: {}, Switch: {}".format(
actor.has_powermeter, actor.has_temperature, actor.has_switch
))
|
python
|
def energy(context, features):
"""Display energy stats of all actors"""
fritz = context.obj
fritz.login()
for actor in fritz.get_actors():
if actor.temperature is not None:
click.echo("{} ({}): {:.2f} Watt current, {:.3f} wH total, {:.2f} °C".format(
actor.name.encode('utf-8'),
actor.actor_id,
(actor.get_power() or 0.0) / 1000,
(actor.get_energy() or 0.0) / 100,
actor.temperature
))
else:
click.echo("{} ({}): {:.2f} Watt current, {:.3f} wH total, offline".format(
actor.name.encode('utf-8'),
actor.actor_id,
(actor.get_power() or 0.0) / 1000,
(actor.get_energy() or 0.0) / 100
))
if features:
click.echo(" Features: PowerMeter: {}, Temperatur: {}, Switch: {}".format(
actor.has_powermeter, actor.has_temperature, actor.has_switch
))
|
[
"def",
"energy",
"(",
"context",
",",
"features",
")",
":",
"fritz",
"=",
"context",
".",
"obj",
"fritz",
".",
"login",
"(",
")",
"for",
"actor",
"in",
"fritz",
".",
"get_actors",
"(",
")",
":",
"if",
"actor",
".",
"temperature",
"is",
"not",
"None",
":",
"click",
".",
"echo",
"(",
"\"{} ({}): {:.2f} Watt current, {:.3f} wH total, {:.2f} °C\".",
"f",
"ormat(",
"",
"actor",
".",
"name",
".",
"encode",
"(",
"'utf-8'",
")",
",",
"actor",
".",
"actor_id",
",",
"(",
"actor",
".",
"get_power",
"(",
")",
"or",
"0.0",
")",
"/",
"1000",
",",
"(",
"actor",
".",
"get_energy",
"(",
")",
"or",
"0.0",
")",
"/",
"100",
",",
"actor",
".",
"temperature",
")",
")",
"else",
":",
"click",
".",
"echo",
"(",
"\"{} ({}): {:.2f} Watt current, {:.3f} wH total, offline\"",
".",
"format",
"(",
"actor",
".",
"name",
".",
"encode",
"(",
"'utf-8'",
")",
",",
"actor",
".",
"actor_id",
",",
"(",
"actor",
".",
"get_power",
"(",
")",
"or",
"0.0",
")",
"/",
"1000",
",",
"(",
"actor",
".",
"get_energy",
"(",
")",
"or",
"0.0",
")",
"/",
"100",
")",
")",
"if",
"features",
":",
"click",
".",
"echo",
"(",
"\" Features: PowerMeter: {}, Temperatur: {}, Switch: {}\"",
".",
"format",
"(",
"actor",
".",
"has_powermeter",
",",
"actor",
".",
"has_temperature",
",",
"actor",
".",
"has_switch",
")",
")"
] |
Display energy stats of all actors
|
[
"Display",
"energy",
"stats",
"of",
"all",
"actors"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L68-L92
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
graphite
|
def graphite(context, server, port, interval, prefix):
"""Display energy stats of all actors"""
fritz = context.obj
fritz.login()
sid_ttl = time.time() + 600
# Find actors and create carbon keys
click.echo(" * Requesting actors list")
simple_chars = re.compile('[^A-Za-z0-9]+')
actors = fritz.get_actors()
keys = {}
for actor in actors:
keys[actor.name] = "{}.{}".format(
prefix,
simple_chars.sub('_', actor.name)
)
# Connect to carbon
click.echo(" * Trying to connect to carbon")
timeout = 2
sock = socket.socket()
sock.settimeout(timeout)
try:
sock.connect((server, port))
except socket.timeout:
raise Exception("Took over {} second(s) to connect to {}".format(
timeout, server
))
except Exception as error:
raise Exception("unknown exception while connecting to {} - {}".format(
server, error
))
def send(key, value):
"""Send a key-value-pair to carbon"""
now = int(time.time())
payload = "{} {} {}\n".format(key, value, now)
sock.sendall(payload)
while True:
if time.time() > sid_ttl:
click.echo(" * Requesting new SID")
fritz.login()
sid_ttl = time.time() + 600
click.echo(" * Requesting statistics")
for actor in actors:
power = actor.get_power()
total = actor.get_energy()
click.echo(" -> {}: {:.2f} Watt current, {:.3f} wH total".format(
actor.name, power / 1000, total / 100
))
send(keys[actor.name] + '.current', power)
send(keys[actor.name] + '.total', total)
time.sleep(interval)
|
python
|
def graphite(context, server, port, interval, prefix):
"""Display energy stats of all actors"""
fritz = context.obj
fritz.login()
sid_ttl = time.time() + 600
# Find actors and create carbon keys
click.echo(" * Requesting actors list")
simple_chars = re.compile('[^A-Za-z0-9]+')
actors = fritz.get_actors()
keys = {}
for actor in actors:
keys[actor.name] = "{}.{}".format(
prefix,
simple_chars.sub('_', actor.name)
)
# Connect to carbon
click.echo(" * Trying to connect to carbon")
timeout = 2
sock = socket.socket()
sock.settimeout(timeout)
try:
sock.connect((server, port))
except socket.timeout:
raise Exception("Took over {} second(s) to connect to {}".format(
timeout, server
))
except Exception as error:
raise Exception("unknown exception while connecting to {} - {}".format(
server, error
))
def send(key, value):
"""Send a key-value-pair to carbon"""
now = int(time.time())
payload = "{} {} {}\n".format(key, value, now)
sock.sendall(payload)
while True:
if time.time() > sid_ttl:
click.echo(" * Requesting new SID")
fritz.login()
sid_ttl = time.time() + 600
click.echo(" * Requesting statistics")
for actor in actors:
power = actor.get_power()
total = actor.get_energy()
click.echo(" -> {}: {:.2f} Watt current, {:.3f} wH total".format(
actor.name, power / 1000, total / 100
))
send(keys[actor.name] + '.current', power)
send(keys[actor.name] + '.total', total)
time.sleep(interval)
|
[
"def",
"graphite",
"(",
"context",
",",
"server",
",",
"port",
",",
"interval",
",",
"prefix",
")",
":",
"fritz",
"=",
"context",
".",
"obj",
"fritz",
".",
"login",
"(",
")",
"sid_ttl",
"=",
"time",
".",
"time",
"(",
")",
"+",
"600",
"# Find actors and create carbon keys",
"click",
".",
"echo",
"(",
"\" * Requesting actors list\"",
")",
"simple_chars",
"=",
"re",
".",
"compile",
"(",
"'[^A-Za-z0-9]+'",
")",
"actors",
"=",
"fritz",
".",
"get_actors",
"(",
")",
"keys",
"=",
"{",
"}",
"for",
"actor",
"in",
"actors",
":",
"keys",
"[",
"actor",
".",
"name",
"]",
"=",
"\"{}.{}\"",
".",
"format",
"(",
"prefix",
",",
"simple_chars",
".",
"sub",
"(",
"'_'",
",",
"actor",
".",
"name",
")",
")",
"# Connect to carbon",
"click",
".",
"echo",
"(",
"\" * Trying to connect to carbon\"",
")",
"timeout",
"=",
"2",
"sock",
"=",
"socket",
".",
"socket",
"(",
")",
"sock",
".",
"settimeout",
"(",
"timeout",
")",
"try",
":",
"sock",
".",
"connect",
"(",
"(",
"server",
",",
"port",
")",
")",
"except",
"socket",
".",
"timeout",
":",
"raise",
"Exception",
"(",
"\"Took over {} second(s) to connect to {}\"",
".",
"format",
"(",
"timeout",
",",
"server",
")",
")",
"except",
"Exception",
"as",
"error",
":",
"raise",
"Exception",
"(",
"\"unknown exception while connecting to {} - {}\"",
".",
"format",
"(",
"server",
",",
"error",
")",
")",
"def",
"send",
"(",
"key",
",",
"value",
")",
":",
"\"\"\"Send a key-value-pair to carbon\"\"\"",
"now",
"=",
"int",
"(",
"time",
".",
"time",
"(",
")",
")",
"payload",
"=",
"\"{} {} {}\\n\"",
".",
"format",
"(",
"key",
",",
"value",
",",
"now",
")",
"sock",
".",
"sendall",
"(",
"payload",
")",
"while",
"True",
":",
"if",
"time",
".",
"time",
"(",
")",
">",
"sid_ttl",
":",
"click",
".",
"echo",
"(",
"\" * Requesting new SID\"",
")",
"fritz",
".",
"login",
"(",
")",
"sid_ttl",
"=",
"time",
".",
"time",
"(",
")",
"+",
"600",
"click",
".",
"echo",
"(",
"\" * Requesting statistics\"",
")",
"for",
"actor",
"in",
"actors",
":",
"power",
"=",
"actor",
".",
"get_power",
"(",
")",
"total",
"=",
"actor",
".",
"get_energy",
"(",
")",
"click",
".",
"echo",
"(",
"\" -> {}: {:.2f} Watt current, {:.3f} wH total\"",
".",
"format",
"(",
"actor",
".",
"name",
",",
"power",
"/",
"1000",
",",
"total",
"/",
"100",
")",
")",
"send",
"(",
"keys",
"[",
"actor",
".",
"name",
"]",
"+",
"'.current'",
",",
"power",
")",
"send",
"(",
"keys",
"[",
"actor",
".",
"name",
"]",
"+",
"'.total'",
",",
"total",
")",
"time",
".",
"sleep",
"(",
"interval",
")"
] |
Display energy stats of all actors
|
[
"Display",
"energy",
"stats",
"of",
"all",
"actors"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L101-L157
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
switch_on
|
def switch_on(context, ain):
"""Switch an actor's power to ON"""
context.obj.login()
actor = context.obj.get_actor_by_ain(ain)
if actor:
click.echo("Switching {} on".format(actor.name))
actor.switch_on()
else:
click.echo("Actor not found: {}".format(ain))
|
python
|
def switch_on(context, ain):
"""Switch an actor's power to ON"""
context.obj.login()
actor = context.obj.get_actor_by_ain(ain)
if actor:
click.echo("Switching {} on".format(actor.name))
actor.switch_on()
else:
click.echo("Actor not found: {}".format(ain))
|
[
"def",
"switch_on",
"(",
"context",
",",
"ain",
")",
":",
"context",
".",
"obj",
".",
"login",
"(",
")",
"actor",
"=",
"context",
".",
"obj",
".",
"get_actor_by_ain",
"(",
"ain",
")",
"if",
"actor",
":",
"click",
".",
"echo",
"(",
"\"Switching {} on\"",
".",
"format",
"(",
"actor",
".",
"name",
")",
")",
"actor",
".",
"switch_on",
"(",
")",
"else",
":",
"click",
".",
"echo",
"(",
"\"Actor not found: {}\"",
".",
"format",
"(",
"ain",
")",
")"
] |
Switch an actor's power to ON
|
[
"Switch",
"an",
"actor",
"s",
"power",
"to",
"ON"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L163-L171
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
switch_state
|
def switch_state(context, ain):
"""Get an actor's power state"""
context.obj.login()
actor = context.obj.get_actor_by_ain(ain)
if actor:
click.echo("State for {} is: {}".format(ain,'ON' if actor.get_state() else 'OFF'))
else:
click.echo("Actor not found: {}".format(ain))
|
python
|
def switch_state(context, ain):
"""Get an actor's power state"""
context.obj.login()
actor = context.obj.get_actor_by_ain(ain)
if actor:
click.echo("State for {} is: {}".format(ain,'ON' if actor.get_state() else 'OFF'))
else:
click.echo("Actor not found: {}".format(ain))
|
[
"def",
"switch_state",
"(",
"context",
",",
"ain",
")",
":",
"context",
".",
"obj",
".",
"login",
"(",
")",
"actor",
"=",
"context",
".",
"obj",
".",
"get_actor_by_ain",
"(",
"ain",
")",
"if",
"actor",
":",
"click",
".",
"echo",
"(",
"\"State for {} is: {}\"",
".",
"format",
"(",
"ain",
",",
"'ON'",
"if",
"actor",
".",
"get_state",
"(",
")",
"else",
"'OFF'",
")",
")",
"else",
":",
"click",
".",
"echo",
"(",
"\"Actor not found: {}\"",
".",
"format",
"(",
"ain",
")",
")"
] |
Get an actor's power state
|
[
"Get",
"an",
"actor",
"s",
"power",
"state"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L191-L198
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
switch_toggle
|
def switch_toggle(context, ain):
"""Toggle an actor's power state"""
context.obj.login()
actor = context.obj.get_actor_by_ain(ain)
if actor:
if actor.get_state():
actor.switch_off()
click.echo("State for {} is now OFF".format(ain))
else:
actor.switch_on()
click.echo("State for {} is now ON".format(ain))
else:
click.echo("Actor not found: {}".format(ain))
|
python
|
def switch_toggle(context, ain):
"""Toggle an actor's power state"""
context.obj.login()
actor = context.obj.get_actor_by_ain(ain)
if actor:
if actor.get_state():
actor.switch_off()
click.echo("State for {} is now OFF".format(ain))
else:
actor.switch_on()
click.echo("State for {} is now ON".format(ain))
else:
click.echo("Actor not found: {}".format(ain))
|
[
"def",
"switch_toggle",
"(",
"context",
",",
"ain",
")",
":",
"context",
".",
"obj",
".",
"login",
"(",
")",
"actor",
"=",
"context",
".",
"obj",
".",
"get_actor_by_ain",
"(",
"ain",
")",
"if",
"actor",
":",
"if",
"actor",
".",
"get_state",
"(",
")",
":",
"actor",
".",
"switch_off",
"(",
")",
"click",
".",
"echo",
"(",
"\"State for {} is now OFF\"",
".",
"format",
"(",
"ain",
")",
")",
"else",
":",
"actor",
".",
"switch_on",
"(",
")",
"click",
".",
"echo",
"(",
"\"State for {} is now ON\"",
".",
"format",
"(",
"ain",
")",
")",
"else",
":",
"click",
".",
"echo",
"(",
"\"Actor not found: {}\"",
".",
"format",
"(",
"ain",
")",
")"
] |
Toggle an actor's power state
|
[
"Toggle",
"an",
"actor",
"s",
"power",
"state"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L204-L216
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/__main__.py
|
logs
|
def logs(context, format):
"""Show system logs since last reboot"""
fritz = context.obj
fritz.login()
messages = fritz.get_logs()
if format == "plain":
for msg in messages:
merged = "{} {} {}".format(msg.date, msg.time, msg.message.encode("UTF-8"))
click.echo(merged)
if format == "json":
entries = [msg._asdict() for msg in messages]
click.echo(json.dumps({
"entries": entries,
}))
|
python
|
def logs(context, format):
"""Show system logs since last reboot"""
fritz = context.obj
fritz.login()
messages = fritz.get_logs()
if format == "plain":
for msg in messages:
merged = "{} {} {}".format(msg.date, msg.time, msg.message.encode("UTF-8"))
click.echo(merged)
if format == "json":
entries = [msg._asdict() for msg in messages]
click.echo(json.dumps({
"entries": entries,
}))
|
[
"def",
"logs",
"(",
"context",
",",
"format",
")",
":",
"fritz",
"=",
"context",
".",
"obj",
"fritz",
".",
"login",
"(",
")",
"messages",
"=",
"fritz",
".",
"get_logs",
"(",
")",
"if",
"format",
"==",
"\"plain\"",
":",
"for",
"msg",
"in",
"messages",
":",
"merged",
"=",
"\"{} {} {}\"",
".",
"format",
"(",
"msg",
".",
"date",
",",
"msg",
".",
"time",
",",
"msg",
".",
"message",
".",
"encode",
"(",
"\"UTF-8\"",
")",
")",
"click",
".",
"echo",
"(",
"merged",
")",
"if",
"format",
"==",
"\"json\"",
":",
"entries",
"=",
"[",
"msg",
".",
"_asdict",
"(",
")",
"for",
"msg",
"in",
"messages",
"]",
"click",
".",
"echo",
"(",
"json",
".",
"dumps",
"(",
"{",
"\"entries\"",
":",
"entries",
",",
"}",
")",
")"
] |
Show system logs since last reboot
|
[
"Show",
"system",
"logs",
"since",
"last",
"reboot"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/__main__.py#L223-L238
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/actor.py
|
Actor.get_power
|
def get_power(self):
"""
Returns the current power usage in milliWatts.
Attention: Returns None if the value can't be queried or is unknown.
"""
value = self.box.homeautoswitch("getswitchpower", self.actor_id)
return int(value) if value.isdigit() else None
|
python
|
def get_power(self):
"""
Returns the current power usage in milliWatts.
Attention: Returns None if the value can't be queried or is unknown.
"""
value = self.box.homeautoswitch("getswitchpower", self.actor_id)
return int(value) if value.isdigit() else None
|
[
"def",
"get_power",
"(",
"self",
")",
":",
"value",
"=",
"self",
".",
"box",
".",
"homeautoswitch",
"(",
"\"getswitchpower\"",
",",
"self",
".",
"actor_id",
")",
"return",
"int",
"(",
"value",
")",
"if",
"value",
".",
"isdigit",
"(",
")",
"else",
"None"
] |
Returns the current power usage in milliWatts.
Attention: Returns None if the value can't be queried or is unknown.
|
[
"Returns",
"the",
"current",
"power",
"usage",
"in",
"milliWatts",
".",
"Attention",
":",
"Returns",
"None",
"if",
"the",
"value",
"can",
"t",
"be",
"queried",
"or",
"is",
"unknown",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/actor.py#L83-L89
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/actor.py
|
Actor.get_energy
|
def get_energy(self):
"""
Returns the consumed energy since the start of the statistics in Wh.
Attention: Returns None if the value can't be queried or is unknown.
"""
value = self.box.homeautoswitch("getswitchenergy", self.actor_id)
return int(value) if value.isdigit() else None
|
python
|
def get_energy(self):
"""
Returns the consumed energy since the start of the statistics in Wh.
Attention: Returns None if the value can't be queried or is unknown.
"""
value = self.box.homeautoswitch("getswitchenergy", self.actor_id)
return int(value) if value.isdigit() else None
|
[
"def",
"get_energy",
"(",
"self",
")",
":",
"value",
"=",
"self",
".",
"box",
".",
"homeautoswitch",
"(",
"\"getswitchenergy\"",
",",
"self",
".",
"actor_id",
")",
"return",
"int",
"(",
"value",
")",
"if",
"value",
".",
"isdigit",
"(",
")",
"else",
"None"
] |
Returns the consumed energy since the start of the statistics in Wh.
Attention: Returns None if the value can't be queried or is unknown.
|
[
"Returns",
"the",
"consumed",
"energy",
"since",
"the",
"start",
"of",
"the",
"statistics",
"in",
"Wh",
".",
"Attention",
":",
"Returns",
"None",
"if",
"the",
"value",
"can",
"t",
"be",
"queried",
"or",
"is",
"unknown",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/actor.py#L91-L97
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/actor.py
|
Actor.get_temperature
|
def get_temperature(self):
"""
Returns the current environment temperature.
Attention: Returns None if the value can't be queried or is unknown.
"""
#raise NotImplementedError("This should work according to the AVM docs, but don't...")
value = self.box.homeautoswitch("gettemperature", self.actor_id)
if value.isdigit():
self.temperature = float(value)/10
else:
self.temperature = None
return self.temperature
|
python
|
def get_temperature(self):
"""
Returns the current environment temperature.
Attention: Returns None if the value can't be queried or is unknown.
"""
#raise NotImplementedError("This should work according to the AVM docs, but don't...")
value = self.box.homeautoswitch("gettemperature", self.actor_id)
if value.isdigit():
self.temperature = float(value)/10
else:
self.temperature = None
return self.temperature
|
[
"def",
"get_temperature",
"(",
"self",
")",
":",
"#raise NotImplementedError(\"This should work according to the AVM docs, but don't...\")",
"value",
"=",
"self",
".",
"box",
".",
"homeautoswitch",
"(",
"\"gettemperature\"",
",",
"self",
".",
"actor_id",
")",
"if",
"value",
".",
"isdigit",
"(",
")",
":",
"self",
".",
"temperature",
"=",
"float",
"(",
"value",
")",
"/",
"10",
"else",
":",
"self",
".",
"temperature",
"=",
"None",
"return",
"self",
".",
"temperature"
] |
Returns the current environment temperature.
Attention: Returns None if the value can't be queried or is unknown.
|
[
"Returns",
"the",
"current",
"environment",
"temperature",
".",
"Attention",
":",
"Returns",
"None",
"if",
"the",
"value",
"can",
"t",
"be",
"queried",
"or",
"is",
"unknown",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/actor.py#L99-L110
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/actor.py
|
Actor.get_target_temperature
|
def get_target_temperature(self):
"""
Returns the actual target temperature.
Attention: Returns None if the value can't be queried or is unknown.
"""
value = self.box.homeautoswitch("gethkrtsoll", self.actor_id)
self.target_temperature = self.__get_temp(value)
return self.target_temperature
|
python
|
def get_target_temperature(self):
"""
Returns the actual target temperature.
Attention: Returns None if the value can't be queried or is unknown.
"""
value = self.box.homeautoswitch("gethkrtsoll", self.actor_id)
self.target_temperature = self.__get_temp(value)
return self.target_temperature
|
[
"def",
"get_target_temperature",
"(",
"self",
")",
":",
"value",
"=",
"self",
".",
"box",
".",
"homeautoswitch",
"(",
"\"gethkrtsoll\"",
",",
"self",
".",
"actor_id",
")",
"self",
".",
"target_temperature",
"=",
"self",
".",
"__get_temp",
"(",
"value",
")",
"return",
"self",
".",
"target_temperature"
] |
Returns the actual target temperature.
Attention: Returns None if the value can't be queried or is unknown.
|
[
"Returns",
"the",
"actual",
"target",
"temperature",
".",
"Attention",
":",
"Returns",
"None",
"if",
"the",
"value",
"can",
"t",
"be",
"queried",
"or",
"is",
"unknown",
"."
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/actor.py#L125-L132
|
train
|
DerMitch/fritzbox-smarthome
|
fritzhome/actor.py
|
Actor.set_temperature
|
def set_temperature(self, temp):
"""
Sets the temperature in celcius
"""
# Temperature is send to fritz.box a little weird
param = 16 + ( ( temp - 8 ) * 2 )
if param < 16:
param = 253
logger.info("Actor " + self.name + ": Temperature control set to off")
elif param >= 56:
param = 254
logger.info("Actor " + self.name + ": Temperature control set to on")
else:
logger.info("Actor " + self.name + ": Temperature control set to " + str(temp))
return self.box.homeautoswitch("sethkrtsoll", self.actor_id, param)
|
python
|
def set_temperature(self, temp):
"""
Sets the temperature in celcius
"""
# Temperature is send to fritz.box a little weird
param = 16 + ( ( temp - 8 ) * 2 )
if param < 16:
param = 253
logger.info("Actor " + self.name + ": Temperature control set to off")
elif param >= 56:
param = 254
logger.info("Actor " + self.name + ": Temperature control set to on")
else:
logger.info("Actor " + self.name + ": Temperature control set to " + str(temp))
return self.box.homeautoswitch("sethkrtsoll", self.actor_id, param)
|
[
"def",
"set_temperature",
"(",
"self",
",",
"temp",
")",
":",
"# Temperature is send to fritz.box a little weird",
"param",
"=",
"16",
"+",
"(",
"(",
"temp",
"-",
"8",
")",
"*",
"2",
")",
"if",
"param",
"<",
"16",
":",
"param",
"=",
"253",
"logger",
".",
"info",
"(",
"\"Actor \"",
"+",
"self",
".",
"name",
"+",
"\": Temperature control set to off\"",
")",
"elif",
"param",
">=",
"56",
":",
"param",
"=",
"254",
"logger",
".",
"info",
"(",
"\"Actor \"",
"+",
"self",
".",
"name",
"+",
"\": Temperature control set to on\"",
")",
"else",
":",
"logger",
".",
"info",
"(",
"\"Actor \"",
"+",
"self",
".",
"name",
"+",
"\": Temperature control set to \"",
"+",
"str",
"(",
"temp",
")",
")",
"return",
"self",
".",
"box",
".",
"homeautoswitch",
"(",
"\"sethkrtsoll\"",
",",
"self",
".",
"actor_id",
",",
"param",
")"
] |
Sets the temperature in celcius
|
[
"Sets",
"the",
"temperature",
"in",
"celcius"
] |
84cbd7c1b33e6256add041b0395ff5fccc01f103
|
https://github.com/DerMitch/fritzbox-smarthome/blob/84cbd7c1b33e6256add041b0395ff5fccc01f103/fritzhome/actor.py#L134-L150
|
train
|
projectatomic/osbs-client
|
osbs/conf.py
|
Configuration.get_openshift_base_uri
|
def get_openshift_base_uri(self):
"""
https://<host>[:<port>]/
:return: str
"""
deprecated_key = "openshift_uri"
key = "openshift_url"
val = self._get_value(deprecated_key, self.conf_section, deprecated_key)
if val is not None:
warnings.warn("%r is deprecated, use %r instead" % (deprecated_key, key))
return val
return self._get_value(key, self.conf_section, key)
|
python
|
def get_openshift_base_uri(self):
"""
https://<host>[:<port>]/
:return: str
"""
deprecated_key = "openshift_uri"
key = "openshift_url"
val = self._get_value(deprecated_key, self.conf_section, deprecated_key)
if val is not None:
warnings.warn("%r is deprecated, use %r instead" % (deprecated_key, key))
return val
return self._get_value(key, self.conf_section, key)
|
[
"def",
"get_openshift_base_uri",
"(",
"self",
")",
":",
"deprecated_key",
"=",
"\"openshift_uri\"",
"key",
"=",
"\"openshift_url\"",
"val",
"=",
"self",
".",
"_get_value",
"(",
"deprecated_key",
",",
"self",
".",
"conf_section",
",",
"deprecated_key",
")",
"if",
"val",
"is",
"not",
"None",
":",
"warnings",
".",
"warn",
"(",
"\"%r is deprecated, use %r instead\"",
"%",
"(",
"deprecated_key",
",",
"key",
")",
")",
"return",
"val",
"return",
"self",
".",
"_get_value",
"(",
"key",
",",
"self",
".",
"conf_section",
",",
"key",
")"
] |
https://<host>[:<port>]/
:return: str
|
[
"https",
":",
"//",
"<host",
">",
"[",
":",
"<port",
">",
"]",
"/"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/conf.py#L130-L142
|
train
|
projectatomic/osbs-client
|
osbs/conf.py
|
Configuration.get_builder_openshift_url
|
def get_builder_openshift_url(self):
""" url of OpenShift where builder will connect """
key = "builder_openshift_url"
url = self._get_deprecated(key, self.conf_section, key)
if url is None:
logger.warning("%r not found, falling back to get_openshift_base_uri()", key)
url = self.get_openshift_base_uri()
return url
|
python
|
def get_builder_openshift_url(self):
""" url of OpenShift where builder will connect """
key = "builder_openshift_url"
url = self._get_deprecated(key, self.conf_section, key)
if url is None:
logger.warning("%r not found, falling back to get_openshift_base_uri()", key)
url = self.get_openshift_base_uri()
return url
|
[
"def",
"get_builder_openshift_url",
"(",
"self",
")",
":",
"key",
"=",
"\"builder_openshift_url\"",
"url",
"=",
"self",
".",
"_get_deprecated",
"(",
"key",
",",
"self",
".",
"conf_section",
",",
"key",
")",
"if",
"url",
"is",
"None",
":",
"logger",
".",
"warning",
"(",
"\"%r not found, falling back to get_openshift_base_uri()\"",
",",
"key",
")",
"url",
"=",
"self",
".",
"get_openshift_base_uri",
"(",
")",
"return",
"url"
] |
url of OpenShift where builder will connect
|
[
"url",
"of",
"OpenShift",
"where",
"builder",
"will",
"connect"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/conf.py#L363-L370
|
train
|
projectatomic/osbs-client
|
osbs/conf.py
|
Configuration.generate_nodeselector_dict
|
def generate_nodeselector_dict(self, nodeselector_str):
"""
helper method for generating nodeselector dict
:param nodeselector_str:
:return: dict
"""
nodeselector = {}
if nodeselector_str and nodeselector_str != 'none':
constraints = [x.strip() for x in nodeselector_str.split(',')]
raw_nodeselector = dict([constraint.split('=', 1) for constraint in constraints])
nodeselector = dict([k.strip(), v.strip()] for (k, v) in raw_nodeselector.items())
return nodeselector
|
python
|
def generate_nodeselector_dict(self, nodeselector_str):
"""
helper method for generating nodeselector dict
:param nodeselector_str:
:return: dict
"""
nodeselector = {}
if nodeselector_str and nodeselector_str != 'none':
constraints = [x.strip() for x in nodeselector_str.split(',')]
raw_nodeselector = dict([constraint.split('=', 1) for constraint in constraints])
nodeselector = dict([k.strip(), v.strip()] for (k, v) in raw_nodeselector.items())
return nodeselector
|
[
"def",
"generate_nodeselector_dict",
"(",
"self",
",",
"nodeselector_str",
")",
":",
"nodeselector",
"=",
"{",
"}",
"if",
"nodeselector_str",
"and",
"nodeselector_str",
"!=",
"'none'",
":",
"constraints",
"=",
"[",
"x",
".",
"strip",
"(",
")",
"for",
"x",
"in",
"nodeselector_str",
".",
"split",
"(",
"','",
")",
"]",
"raw_nodeselector",
"=",
"dict",
"(",
"[",
"constraint",
".",
"split",
"(",
"'='",
",",
"1",
")",
"for",
"constraint",
"in",
"constraints",
"]",
")",
"nodeselector",
"=",
"dict",
"(",
"[",
"k",
".",
"strip",
"(",
")",
",",
"v",
".",
"strip",
"(",
")",
"]",
"for",
"(",
"k",
",",
"v",
")",
"in",
"raw_nodeselector",
".",
"items",
"(",
")",
")",
"return",
"nodeselector"
] |
helper method for generating nodeselector dict
:param nodeselector_str:
:return: dict
|
[
"helper",
"method",
"for",
"generating",
"nodeselector",
"dict",
":",
"param",
"nodeselector_str",
":",
":",
"return",
":",
"dict"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/conf.py#L565-L577
|
train
|
projectatomic/osbs-client
|
osbs/conf.py
|
Configuration.get_platform_node_selector
|
def get_platform_node_selector(self, platform):
"""
search the configuration for entries of the form node_selector.platform
:param platform: str, platform to search for, can be null
:return dict
"""
nodeselector = {}
if platform:
nodeselector_str = self._get_value("node_selector." + platform, self.conf_section,
"node_selector." + platform)
nodeselector = self.generate_nodeselector_dict(nodeselector_str)
return nodeselector
|
python
|
def get_platform_node_selector(self, platform):
"""
search the configuration for entries of the form node_selector.platform
:param platform: str, platform to search for, can be null
:return dict
"""
nodeselector = {}
if platform:
nodeselector_str = self._get_value("node_selector." + platform, self.conf_section,
"node_selector." + platform)
nodeselector = self.generate_nodeselector_dict(nodeselector_str)
return nodeselector
|
[
"def",
"get_platform_node_selector",
"(",
"self",
",",
"platform",
")",
":",
"nodeselector",
"=",
"{",
"}",
"if",
"platform",
":",
"nodeselector_str",
"=",
"self",
".",
"_get_value",
"(",
"\"node_selector.\"",
"+",
"platform",
",",
"self",
".",
"conf_section",
",",
"\"node_selector.\"",
"+",
"platform",
")",
"nodeselector",
"=",
"self",
".",
"generate_nodeselector_dict",
"(",
"nodeselector_str",
")",
"return",
"nodeselector"
] |
search the configuration for entries of the form node_selector.platform
:param platform: str, platform to search for, can be null
:return dict
|
[
"search",
"the",
"configuration",
"for",
"entries",
"of",
"the",
"form",
"node_selector",
".",
"platform",
":",
"param",
"platform",
":",
"str",
"platform",
"to",
"search",
"for",
"can",
"be",
"null",
":",
"return",
"dict"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/conf.py#L579-L591
|
train
|
thombashi/pytablereader
|
pytablereader/csv/core.py
|
CsvTableFileLoader.load
|
def load(self):
"""
Extract tabular data as |TableData| instances from a CSV file.
|load_source_desc_file|
:return:
Loaded table data.
|load_table_name_desc|
=================== ========================================
Format specifier Value after the replacement
=================== ========================================
``%(filename)s`` |filename_desc|
``%(format_name)s`` ``"csv"``
``%(format_id)s`` |format_id_desc|
``%(global_id)s`` |global_id|
=================== ========================================
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the CSV data is invalid.
.. seealso::
:py:func:`csv.reader`
"""
self._validate()
self._logger.logging_load()
self.encoding = get_file_encoding(self.source, self.encoding)
if six.PY3:
self._csv_reader = csv.reader(
io.open(self.source, "r", encoding=self.encoding),
delimiter=self.delimiter,
quotechar=self.quotechar,
strict=True,
skipinitialspace=True,
)
else:
self._csv_reader = csv.reader(
_utf_8_encoder(io.open(self.source, "r", encoding=self.encoding)),
delimiter=self.delimiter,
quotechar=self.quotechar,
strict=True,
skipinitialspace=True,
)
formatter = CsvTableFormatter(self._to_data_matrix())
formatter.accept(self)
return formatter.to_table_data()
|
python
|
def load(self):
"""
Extract tabular data as |TableData| instances from a CSV file.
|load_source_desc_file|
:return:
Loaded table data.
|load_table_name_desc|
=================== ========================================
Format specifier Value after the replacement
=================== ========================================
``%(filename)s`` |filename_desc|
``%(format_name)s`` ``"csv"``
``%(format_id)s`` |format_id_desc|
``%(global_id)s`` |global_id|
=================== ========================================
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the CSV data is invalid.
.. seealso::
:py:func:`csv.reader`
"""
self._validate()
self._logger.logging_load()
self.encoding = get_file_encoding(self.source, self.encoding)
if six.PY3:
self._csv_reader = csv.reader(
io.open(self.source, "r", encoding=self.encoding),
delimiter=self.delimiter,
quotechar=self.quotechar,
strict=True,
skipinitialspace=True,
)
else:
self._csv_reader = csv.reader(
_utf_8_encoder(io.open(self.source, "r", encoding=self.encoding)),
delimiter=self.delimiter,
quotechar=self.quotechar,
strict=True,
skipinitialspace=True,
)
formatter = CsvTableFormatter(self._to_data_matrix())
formatter.accept(self)
return formatter.to_table_data()
|
[
"def",
"load",
"(",
"self",
")",
":",
"self",
".",
"_validate",
"(",
")",
"self",
".",
"_logger",
".",
"logging_load",
"(",
")",
"self",
".",
"encoding",
"=",
"get_file_encoding",
"(",
"self",
".",
"source",
",",
"self",
".",
"encoding",
")",
"if",
"six",
".",
"PY3",
":",
"self",
".",
"_csv_reader",
"=",
"csv",
".",
"reader",
"(",
"io",
".",
"open",
"(",
"self",
".",
"source",
",",
"\"r\"",
",",
"encoding",
"=",
"self",
".",
"encoding",
")",
",",
"delimiter",
"=",
"self",
".",
"delimiter",
",",
"quotechar",
"=",
"self",
".",
"quotechar",
",",
"strict",
"=",
"True",
",",
"skipinitialspace",
"=",
"True",
",",
")",
"else",
":",
"self",
".",
"_csv_reader",
"=",
"csv",
".",
"reader",
"(",
"_utf_8_encoder",
"(",
"io",
".",
"open",
"(",
"self",
".",
"source",
",",
"\"r\"",
",",
"encoding",
"=",
"self",
".",
"encoding",
")",
")",
",",
"delimiter",
"=",
"self",
".",
"delimiter",
",",
"quotechar",
"=",
"self",
".",
"quotechar",
",",
"strict",
"=",
"True",
",",
"skipinitialspace",
"=",
"True",
",",
")",
"formatter",
"=",
"CsvTableFormatter",
"(",
"self",
".",
"_to_data_matrix",
"(",
")",
")",
"formatter",
".",
"accept",
"(",
"self",
")",
"return",
"formatter",
".",
"to_table_data",
"(",
")"
] |
Extract tabular data as |TableData| instances from a CSV file.
|load_source_desc_file|
:return:
Loaded table data.
|load_table_name_desc|
=================== ========================================
Format specifier Value after the replacement
=================== ========================================
``%(filename)s`` |filename_desc|
``%(format_name)s`` ``"csv"``
``%(format_id)s`` |format_id_desc|
``%(global_id)s`` |global_id|
=================== ========================================
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the CSV data is invalid.
.. seealso::
:py:func:`csv.reader`
|
[
"Extract",
"tabular",
"data",
"as",
"|TableData|",
"instances",
"from",
"a",
"CSV",
"file",
".",
"|load_source_desc_file|"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/csv/core.py#L145-L194
|
train
|
thombashi/pytablereader
|
pytablereader/csv/core.py
|
CsvTableTextLoader.load
|
def load(self):
"""
Extract tabular data as |TableData| instances from a CSV text object.
|load_source_desc_text|
:return:
Loaded table data.
|load_table_name_desc|
=================== ========================================
Format specifier Value after the replacement
=================== ========================================
``%(filename)s`` ``""``
``%(format_name)s`` ``"csv"``
``%(format_id)s`` |format_id_desc|
``%(global_id)s`` |global_id|
=================== ========================================
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the CSV data is invalid.
.. seealso::
:py:func:`csv.reader`
"""
self._validate()
self._logger.logging_load()
self._csv_reader = csv.reader(
six.StringIO(self.source.strip()),
delimiter=self.delimiter,
quotechar=self.quotechar,
strict=True,
skipinitialspace=True,
)
formatter = CsvTableFormatter(self._to_data_matrix())
formatter.accept(self)
return formatter.to_table_data()
|
python
|
def load(self):
"""
Extract tabular data as |TableData| instances from a CSV text object.
|load_source_desc_text|
:return:
Loaded table data.
|load_table_name_desc|
=================== ========================================
Format specifier Value after the replacement
=================== ========================================
``%(filename)s`` ``""``
``%(format_name)s`` ``"csv"``
``%(format_id)s`` |format_id_desc|
``%(global_id)s`` |global_id|
=================== ========================================
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the CSV data is invalid.
.. seealso::
:py:func:`csv.reader`
"""
self._validate()
self._logger.logging_load()
self._csv_reader = csv.reader(
six.StringIO(self.source.strip()),
delimiter=self.delimiter,
quotechar=self.quotechar,
strict=True,
skipinitialspace=True,
)
formatter = CsvTableFormatter(self._to_data_matrix())
formatter.accept(self)
return formatter.to_table_data()
|
[
"def",
"load",
"(",
"self",
")",
":",
"self",
".",
"_validate",
"(",
")",
"self",
".",
"_logger",
".",
"logging_load",
"(",
")",
"self",
".",
"_csv_reader",
"=",
"csv",
".",
"reader",
"(",
"six",
".",
"StringIO",
"(",
"self",
".",
"source",
".",
"strip",
"(",
")",
")",
",",
"delimiter",
"=",
"self",
".",
"delimiter",
",",
"quotechar",
"=",
"self",
".",
"quotechar",
",",
"strict",
"=",
"True",
",",
"skipinitialspace",
"=",
"True",
",",
")",
"formatter",
"=",
"CsvTableFormatter",
"(",
"self",
".",
"_to_data_matrix",
"(",
")",
")",
"formatter",
".",
"accept",
"(",
"self",
")",
"return",
"formatter",
".",
"to_table_data",
"(",
")"
] |
Extract tabular data as |TableData| instances from a CSV text object.
|load_source_desc_text|
:return:
Loaded table data.
|load_table_name_desc|
=================== ========================================
Format specifier Value after the replacement
=================== ========================================
``%(filename)s`` ``""``
``%(format_name)s`` ``"csv"``
``%(format_id)s`` |format_id_desc|
``%(global_id)s`` |global_id|
=================== ========================================
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the CSV data is invalid.
.. seealso::
:py:func:`csv.reader`
|
[
"Extract",
"tabular",
"data",
"as",
"|TableData|",
"instances",
"from",
"a",
"CSV",
"text",
"object",
".",
"|load_source_desc_text|"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/csv/core.py#L220-L258
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.set_params
|
def set_params(self, **kwargs):
"""
set parameters according to specification
these parameters are accepted:
:param pulp_secret: str, resource name of pulp secret
:param koji_target: str, koji tag with packages used to build the image
:param kojiroot: str, URL from which koji packages are fetched
:param kojihub: str, URL of the koji hub
:param koji_certs_secret: str, resource name of secret that holds the koji certificates
:param koji_task_id: int, Koji Task that created this build config
:param flatpak: if we should build a Flatpak OCI Image
:param filesystem_koji_task_id: int, Koji Task that created the base filesystem
:param pulp_registry: str, name of pulp registry in dockpulp.conf
:param sources_command: str, command used to fetch dist-git sources
:param architecture: str, architecture we are building for
:param vendor: str, vendor name
:param build_host: str, host the build will run on or None for auto
:param authoritative_registry: str, the docker registry authoritative for this image
:param distribution_scope: str, distribution scope for this image
(private, authoritative-source-only, restricted, public)
:param use_auth: bool, use auth from atomic-reactor?
:param platform_node_selector: dict, a nodeselector for a specific platform
:param platform_descriptors: dict, platforms and their archiectures and enable_v1 settings
:param scratch_build_node_selector: dict, a nodeselector for scratch builds
:param explicit_build_node_selector: dict, a nodeselector for explicit builds
:param auto_build_node_selector: dict, a nodeselector for auto builds
:param isolated_build_node_selector: dict, a nodeselector for isolated builds
:param is_auto: bool, indicates if build is auto build
:param parent_images_digests: dict, mapping image names with tags to platform specific
digests, example:
{'registry.fedorahosted.org/fedora:29': {
x86_64': 'registry.fedorahosted.org/fedora@sha256:....'}
}
"""
# Here we cater to the koji "scratch" build type, this will disable
# all plugins that might cause importing of data to koji
self.scratch = kwargs.pop('scratch', False)
# When true, it indicates build was automatically started by
# OpenShift via a trigger, for instance ImageChangeTrigger
self.is_auto = kwargs.pop('is_auto', False)
# An isolated build is meant to patch a certain release and not
# update transient tags in container registry
self.isolated = kwargs.pop('isolated', False)
self.validate_build_variation()
self.base_image = kwargs.get('base_image')
self.platform_node_selector = kwargs.get('platform_node_selector', {})
self.platform_descriptors = kwargs.get('platform_descriptors', {})
self.scratch_build_node_selector = kwargs.get('scratch_build_node_selector', {})
self.explicit_build_node_selector = kwargs.get('explicit_build_node_selector', {})
self.auto_build_node_selector = kwargs.get('auto_build_node_selector', {})
self.isolated_build_node_selector = kwargs.get('isolated_build_node_selector', {})
logger.debug("setting params '%s' for %s", kwargs, self.spec)
self.spec.set_params(**kwargs)
self.osbs_api = kwargs.pop('osbs_api')
|
python
|
def set_params(self, **kwargs):
"""
set parameters according to specification
these parameters are accepted:
:param pulp_secret: str, resource name of pulp secret
:param koji_target: str, koji tag with packages used to build the image
:param kojiroot: str, URL from which koji packages are fetched
:param kojihub: str, URL of the koji hub
:param koji_certs_secret: str, resource name of secret that holds the koji certificates
:param koji_task_id: int, Koji Task that created this build config
:param flatpak: if we should build a Flatpak OCI Image
:param filesystem_koji_task_id: int, Koji Task that created the base filesystem
:param pulp_registry: str, name of pulp registry in dockpulp.conf
:param sources_command: str, command used to fetch dist-git sources
:param architecture: str, architecture we are building for
:param vendor: str, vendor name
:param build_host: str, host the build will run on or None for auto
:param authoritative_registry: str, the docker registry authoritative for this image
:param distribution_scope: str, distribution scope for this image
(private, authoritative-source-only, restricted, public)
:param use_auth: bool, use auth from atomic-reactor?
:param platform_node_selector: dict, a nodeselector for a specific platform
:param platform_descriptors: dict, platforms and their archiectures and enable_v1 settings
:param scratch_build_node_selector: dict, a nodeselector for scratch builds
:param explicit_build_node_selector: dict, a nodeselector for explicit builds
:param auto_build_node_selector: dict, a nodeselector for auto builds
:param isolated_build_node_selector: dict, a nodeselector for isolated builds
:param is_auto: bool, indicates if build is auto build
:param parent_images_digests: dict, mapping image names with tags to platform specific
digests, example:
{'registry.fedorahosted.org/fedora:29': {
x86_64': 'registry.fedorahosted.org/fedora@sha256:....'}
}
"""
# Here we cater to the koji "scratch" build type, this will disable
# all plugins that might cause importing of data to koji
self.scratch = kwargs.pop('scratch', False)
# When true, it indicates build was automatically started by
# OpenShift via a trigger, for instance ImageChangeTrigger
self.is_auto = kwargs.pop('is_auto', False)
# An isolated build is meant to patch a certain release and not
# update transient tags in container registry
self.isolated = kwargs.pop('isolated', False)
self.validate_build_variation()
self.base_image = kwargs.get('base_image')
self.platform_node_selector = kwargs.get('platform_node_selector', {})
self.platform_descriptors = kwargs.get('platform_descriptors', {})
self.scratch_build_node_selector = kwargs.get('scratch_build_node_selector', {})
self.explicit_build_node_selector = kwargs.get('explicit_build_node_selector', {})
self.auto_build_node_selector = kwargs.get('auto_build_node_selector', {})
self.isolated_build_node_selector = kwargs.get('isolated_build_node_selector', {})
logger.debug("setting params '%s' for %s", kwargs, self.spec)
self.spec.set_params(**kwargs)
self.osbs_api = kwargs.pop('osbs_api')
|
[
"def",
"set_params",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"# Here we cater to the koji \"scratch\" build type, this will disable",
"# all plugins that might cause importing of data to koji",
"self",
".",
"scratch",
"=",
"kwargs",
".",
"pop",
"(",
"'scratch'",
",",
"False",
")",
"# When true, it indicates build was automatically started by",
"# OpenShift via a trigger, for instance ImageChangeTrigger",
"self",
".",
"is_auto",
"=",
"kwargs",
".",
"pop",
"(",
"'is_auto'",
",",
"False",
")",
"# An isolated build is meant to patch a certain release and not",
"# update transient tags in container registry",
"self",
".",
"isolated",
"=",
"kwargs",
".",
"pop",
"(",
"'isolated'",
",",
"False",
")",
"self",
".",
"validate_build_variation",
"(",
")",
"self",
".",
"base_image",
"=",
"kwargs",
".",
"get",
"(",
"'base_image'",
")",
"self",
".",
"platform_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'platform_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"platform_descriptors",
"=",
"kwargs",
".",
"get",
"(",
"'platform_descriptors'",
",",
"{",
"}",
")",
"self",
".",
"scratch_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'scratch_build_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"explicit_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'explicit_build_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"auto_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'auto_build_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"isolated_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'isolated_build_node_selector'",
",",
"{",
"}",
")",
"logger",
".",
"debug",
"(",
"\"setting params '%s' for %s\"",
",",
"kwargs",
",",
"self",
".",
"spec",
")",
"self",
".",
"spec",
".",
"set_params",
"(",
"*",
"*",
"kwargs",
")",
"self",
".",
"osbs_api",
"=",
"kwargs",
".",
"pop",
"(",
"'osbs_api'",
")"
] |
set parameters according to specification
these parameters are accepted:
:param pulp_secret: str, resource name of pulp secret
:param koji_target: str, koji tag with packages used to build the image
:param kojiroot: str, URL from which koji packages are fetched
:param kojihub: str, URL of the koji hub
:param koji_certs_secret: str, resource name of secret that holds the koji certificates
:param koji_task_id: int, Koji Task that created this build config
:param flatpak: if we should build a Flatpak OCI Image
:param filesystem_koji_task_id: int, Koji Task that created the base filesystem
:param pulp_registry: str, name of pulp registry in dockpulp.conf
:param sources_command: str, command used to fetch dist-git sources
:param architecture: str, architecture we are building for
:param vendor: str, vendor name
:param build_host: str, host the build will run on or None for auto
:param authoritative_registry: str, the docker registry authoritative for this image
:param distribution_scope: str, distribution scope for this image
(private, authoritative-source-only, restricted, public)
:param use_auth: bool, use auth from atomic-reactor?
:param platform_node_selector: dict, a nodeselector for a specific platform
:param platform_descriptors: dict, platforms and their archiectures and enable_v1 settings
:param scratch_build_node_selector: dict, a nodeselector for scratch builds
:param explicit_build_node_selector: dict, a nodeselector for explicit builds
:param auto_build_node_selector: dict, a nodeselector for auto builds
:param isolated_build_node_selector: dict, a nodeselector for isolated builds
:param is_auto: bool, indicates if build is auto build
:param parent_images_digests: dict, mapping image names with tags to platform specific
digests, example:
{'registry.fedorahosted.org/fedora:29': {
x86_64': 'registry.fedorahosted.org/fedora@sha256:....'}
}
|
[
"set",
"parameters",
"according",
"to",
"specification"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L72-L131
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.has_ist_trigger
|
def has_ist_trigger(self):
"""Return True if this BuildConfig has ImageStreamTag trigger."""
triggers = self.template['spec'].get('triggers', [])
if not triggers:
return False
for trigger in triggers:
if trigger['type'] == 'ImageChange' and \
trigger['imageChange']['from']['kind'] == 'ImageStreamTag':
return True
return False
|
python
|
def has_ist_trigger(self):
"""Return True if this BuildConfig has ImageStreamTag trigger."""
triggers = self.template['spec'].get('triggers', [])
if not triggers:
return False
for trigger in triggers:
if trigger['type'] == 'ImageChange' and \
trigger['imageChange']['from']['kind'] == 'ImageStreamTag':
return True
return False
|
[
"def",
"has_ist_trigger",
"(",
"self",
")",
":",
"triggers",
"=",
"self",
".",
"template",
"[",
"'spec'",
"]",
".",
"get",
"(",
"'triggers'",
",",
"[",
"]",
")",
"if",
"not",
"triggers",
":",
"return",
"False",
"for",
"trigger",
"in",
"triggers",
":",
"if",
"trigger",
"[",
"'type'",
"]",
"==",
"'ImageChange'",
"and",
"trigger",
"[",
"'imageChange'",
"]",
"[",
"'from'",
"]",
"[",
"'kind'",
"]",
"==",
"'ImageStreamTag'",
":",
"return",
"True",
"return",
"False"
] |
Return True if this BuildConfig has ImageStreamTag trigger.
|
[
"Return",
"True",
"if",
"this",
"BuildConfig",
"has",
"ImageStreamTag",
"trigger",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L210-L219
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.set_secret_for_plugin
|
def set_secret_for_plugin(self, secret, plugin=None, mount_path=None):
"""
Sets secret for plugin, if no plugin specified
it will also set general secret
:param secret: str, secret name
:param plugin: tuple, (plugin type, plugin name, argument name)
:param mount_path: str, mount path of secret
"""
has_plugin_conf = False
if plugin is not None:
has_plugin_conf = self.dj.dock_json_has_plugin_conf(plugin[0],
plugin[1])
if 'secrets' in self.template['spec']['strategy']['customStrategy']:
if not plugin or has_plugin_conf:
custom = self.template['spec']['strategy']['customStrategy']
if mount_path:
secret_path = mount_path
else:
secret_path = os.path.join(SECRETS_PATH, secret)
logger.info("Configuring %s secret at %s", secret, secret_path)
existing = [secret_mount for secret_mount in custom['secrets']
if secret_mount['secretSource']['name'] == secret]
if existing:
logger.debug("secret %s already set", secret)
else:
custom['secrets'].append({
'secretSource': {
'name': secret,
},
'mountPath': secret_path,
})
# there's no need to set args if no plugin secret specified
# this is used in tag_and_push plugin, as it sets secret path
# for each registry separately
if plugin and plugin[2] is not None:
self.dj.dock_json_set_arg(*(plugin + (secret_path,)))
else:
logger.debug("not setting secret for unused plugin %s",
plugin[1])
|
python
|
def set_secret_for_plugin(self, secret, plugin=None, mount_path=None):
"""
Sets secret for plugin, if no plugin specified
it will also set general secret
:param secret: str, secret name
:param plugin: tuple, (plugin type, plugin name, argument name)
:param mount_path: str, mount path of secret
"""
has_plugin_conf = False
if plugin is not None:
has_plugin_conf = self.dj.dock_json_has_plugin_conf(plugin[0],
plugin[1])
if 'secrets' in self.template['spec']['strategy']['customStrategy']:
if not plugin or has_plugin_conf:
custom = self.template['spec']['strategy']['customStrategy']
if mount_path:
secret_path = mount_path
else:
secret_path = os.path.join(SECRETS_PATH, secret)
logger.info("Configuring %s secret at %s", secret, secret_path)
existing = [secret_mount for secret_mount in custom['secrets']
if secret_mount['secretSource']['name'] == secret]
if existing:
logger.debug("secret %s already set", secret)
else:
custom['secrets'].append({
'secretSource': {
'name': secret,
},
'mountPath': secret_path,
})
# there's no need to set args if no plugin secret specified
# this is used in tag_and_push plugin, as it sets secret path
# for each registry separately
if plugin and plugin[2] is not None:
self.dj.dock_json_set_arg(*(plugin + (secret_path,)))
else:
logger.debug("not setting secret for unused plugin %s",
plugin[1])
|
[
"def",
"set_secret_for_plugin",
"(",
"self",
",",
"secret",
",",
"plugin",
"=",
"None",
",",
"mount_path",
"=",
"None",
")",
":",
"has_plugin_conf",
"=",
"False",
"if",
"plugin",
"is",
"not",
"None",
":",
"has_plugin_conf",
"=",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"plugin",
"[",
"0",
"]",
",",
"plugin",
"[",
"1",
"]",
")",
"if",
"'secrets'",
"in",
"self",
".",
"template",
"[",
"'spec'",
"]",
"[",
"'strategy'",
"]",
"[",
"'customStrategy'",
"]",
":",
"if",
"not",
"plugin",
"or",
"has_plugin_conf",
":",
"custom",
"=",
"self",
".",
"template",
"[",
"'spec'",
"]",
"[",
"'strategy'",
"]",
"[",
"'customStrategy'",
"]",
"if",
"mount_path",
":",
"secret_path",
"=",
"mount_path",
"else",
":",
"secret_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"SECRETS_PATH",
",",
"secret",
")",
"logger",
".",
"info",
"(",
"\"Configuring %s secret at %s\"",
",",
"secret",
",",
"secret_path",
")",
"existing",
"=",
"[",
"secret_mount",
"for",
"secret_mount",
"in",
"custom",
"[",
"'secrets'",
"]",
"if",
"secret_mount",
"[",
"'secretSource'",
"]",
"[",
"'name'",
"]",
"==",
"secret",
"]",
"if",
"existing",
":",
"logger",
".",
"debug",
"(",
"\"secret %s already set\"",
",",
"secret",
")",
"else",
":",
"custom",
"[",
"'secrets'",
"]",
".",
"append",
"(",
"{",
"'secretSource'",
":",
"{",
"'name'",
":",
"secret",
",",
"}",
",",
"'mountPath'",
":",
"secret_path",
",",
"}",
")",
"# there's no need to set args if no plugin secret specified",
"# this is used in tag_and_push plugin, as it sets secret path",
"# for each registry separately",
"if",
"plugin",
"and",
"plugin",
"[",
"2",
"]",
"is",
"not",
"None",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"*",
"(",
"plugin",
"+",
"(",
"secret_path",
",",
")",
")",
")",
"else",
":",
"logger",
".",
"debug",
"(",
"\"not setting secret for unused plugin %s\"",
",",
"plugin",
"[",
"1",
"]",
")"
] |
Sets secret for plugin, if no plugin specified
it will also set general secret
:param secret: str, secret name
:param plugin: tuple, (plugin type, plugin name, argument name)
:param mount_path: str, mount path of secret
|
[
"Sets",
"secret",
"for",
"plugin",
"if",
"no",
"plugin",
"specified",
"it",
"will",
"also",
"set",
"general",
"secret"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L403-L445
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.set_secrets
|
def set_secrets(self, secrets):
"""
:param secrets: dict, {(plugin type, plugin name, argument name): secret name}
for example {('exit_plugins', 'koji_promote', 'koji_ssl_certs'): 'koji_ssl_certs', ...}
"""
secret_set = False
for (plugin, secret) in secrets.items():
if not isinstance(plugin, tuple) or len(plugin) != 3:
raise ValueError('got "%s" as secrets key, need 3-tuple' % plugin)
if secret is not None:
if isinstance(secret, list):
for secret_item in secret:
self.set_secret_for_plugin(secret_item, plugin=plugin)
else:
self.set_secret_for_plugin(secret, plugin=plugin)
secret_set = True
if not secret_set:
# remove references to secret if no secret was set
if 'secrets' in self.template['spec']['strategy']['customStrategy']:
del self.template['spec']['strategy']['customStrategy']['secrets']
|
python
|
def set_secrets(self, secrets):
"""
:param secrets: dict, {(plugin type, plugin name, argument name): secret name}
for example {('exit_plugins', 'koji_promote', 'koji_ssl_certs'): 'koji_ssl_certs', ...}
"""
secret_set = False
for (plugin, secret) in secrets.items():
if not isinstance(plugin, tuple) or len(plugin) != 3:
raise ValueError('got "%s" as secrets key, need 3-tuple' % plugin)
if secret is not None:
if isinstance(secret, list):
for secret_item in secret:
self.set_secret_for_plugin(secret_item, plugin=plugin)
else:
self.set_secret_for_plugin(secret, plugin=plugin)
secret_set = True
if not secret_set:
# remove references to secret if no secret was set
if 'secrets' in self.template['spec']['strategy']['customStrategy']:
del self.template['spec']['strategy']['customStrategy']['secrets']
|
[
"def",
"set_secrets",
"(",
"self",
",",
"secrets",
")",
":",
"secret_set",
"=",
"False",
"for",
"(",
"plugin",
",",
"secret",
")",
"in",
"secrets",
".",
"items",
"(",
")",
":",
"if",
"not",
"isinstance",
"(",
"plugin",
",",
"tuple",
")",
"or",
"len",
"(",
"plugin",
")",
"!=",
"3",
":",
"raise",
"ValueError",
"(",
"'got \"%s\" as secrets key, need 3-tuple'",
"%",
"plugin",
")",
"if",
"secret",
"is",
"not",
"None",
":",
"if",
"isinstance",
"(",
"secret",
",",
"list",
")",
":",
"for",
"secret_item",
"in",
"secret",
":",
"self",
".",
"set_secret_for_plugin",
"(",
"secret_item",
",",
"plugin",
"=",
"plugin",
")",
"else",
":",
"self",
".",
"set_secret_for_plugin",
"(",
"secret",
",",
"plugin",
"=",
"plugin",
")",
"secret_set",
"=",
"True",
"if",
"not",
"secret_set",
":",
"# remove references to secret if no secret was set",
"if",
"'secrets'",
"in",
"self",
".",
"template",
"[",
"'spec'",
"]",
"[",
"'strategy'",
"]",
"[",
"'customStrategy'",
"]",
":",
"del",
"self",
".",
"template",
"[",
"'spec'",
"]",
"[",
"'strategy'",
"]",
"[",
"'customStrategy'",
"]",
"[",
"'secrets'",
"]"
] |
:param secrets: dict, {(plugin type, plugin name, argument name): secret name}
for example {('exit_plugins', 'koji_promote', 'koji_ssl_certs'): 'koji_ssl_certs', ...}
|
[
":",
"param",
"secrets",
":",
"dict",
"{",
"(",
"plugin",
"type",
"plugin",
"name",
"argument",
"name",
")",
":",
"secret",
"name",
"}",
"for",
"example",
"{",
"(",
"exit_plugins",
"koji_promote",
"koji_ssl_certs",
")",
":",
"koji_ssl_certs",
"...",
"}"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L447-L467
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.remove_tag_and_push_registries
|
def remove_tag_and_push_registries(tag_and_push_registries, version):
"""
Remove matching entries from tag_and_push_registries (in-place)
:param tag_and_push_registries: dict, uri -> dict
:param version: str, 'version' to match against
"""
registries = [uri
for uri, regdict in tag_and_push_registries.items()
if regdict['version'] == version]
for registry in registries:
logger.info("removing %s registry: %s", version, registry)
del tag_and_push_registries[registry]
|
python
|
def remove_tag_and_push_registries(tag_and_push_registries, version):
"""
Remove matching entries from tag_and_push_registries (in-place)
:param tag_and_push_registries: dict, uri -> dict
:param version: str, 'version' to match against
"""
registries = [uri
for uri, regdict in tag_and_push_registries.items()
if regdict['version'] == version]
for registry in registries:
logger.info("removing %s registry: %s", version, registry)
del tag_and_push_registries[registry]
|
[
"def",
"remove_tag_and_push_registries",
"(",
"tag_and_push_registries",
",",
"version",
")",
":",
"registries",
"=",
"[",
"uri",
"for",
"uri",
",",
"regdict",
"in",
"tag_and_push_registries",
".",
"items",
"(",
")",
"if",
"regdict",
"[",
"'version'",
"]",
"==",
"version",
"]",
"for",
"registry",
"in",
"registries",
":",
"logger",
".",
"info",
"(",
"\"removing %s registry: %s\"",
",",
"version",
",",
"registry",
")",
"del",
"tag_and_push_registries",
"[",
"registry",
"]"
] |
Remove matching entries from tag_and_push_registries (in-place)
:param tag_and_push_registries: dict, uri -> dict
:param version: str, 'version' to match against
|
[
"Remove",
"matching",
"entries",
"from",
"tag_and_push_registries",
"(",
"in",
"-",
"place",
")"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L488-L500
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.adjust_for_registry_api_versions
|
def adjust_for_registry_api_versions(self):
"""
Enable/disable plugins depending on supported registry API versions
"""
versions = self.spec.registry_api_versions.value
if 'v2' not in versions:
raise OsbsValidationException('v1-only docker registry API is not supported')
try:
push_conf = self.dj.dock_json_get_plugin_conf('postbuild_plugins',
'tag_and_push')
tag_and_push_registries = push_conf['args']['registries']
except (KeyError, IndexError):
tag_and_push_registries = {}
if 'v1' not in versions:
# Remove v1-only plugins
for phase, name in [('postbuild_plugins', 'pulp_push')]:
logger.info("removing v1-only plugin: %s", name)
self.dj.remove_plugin(phase, name)
# remove extra tag_and_push config
self.remove_tag_and_push_registries(tag_and_push_registries, 'v1')
# Remove 'version' from tag_and_push plugin config as it's no
# longer needed
for regdict in tag_and_push_registries.values():
if 'version' in regdict:
del regdict['version']
|
python
|
def adjust_for_registry_api_versions(self):
"""
Enable/disable plugins depending on supported registry API versions
"""
versions = self.spec.registry_api_versions.value
if 'v2' not in versions:
raise OsbsValidationException('v1-only docker registry API is not supported')
try:
push_conf = self.dj.dock_json_get_plugin_conf('postbuild_plugins',
'tag_and_push')
tag_and_push_registries = push_conf['args']['registries']
except (KeyError, IndexError):
tag_and_push_registries = {}
if 'v1' not in versions:
# Remove v1-only plugins
for phase, name in [('postbuild_plugins', 'pulp_push')]:
logger.info("removing v1-only plugin: %s", name)
self.dj.remove_plugin(phase, name)
# remove extra tag_and_push config
self.remove_tag_and_push_registries(tag_and_push_registries, 'v1')
# Remove 'version' from tag_and_push plugin config as it's no
# longer needed
for regdict in tag_and_push_registries.values():
if 'version' in regdict:
del regdict['version']
|
[
"def",
"adjust_for_registry_api_versions",
"(",
"self",
")",
":",
"versions",
"=",
"self",
".",
"spec",
".",
"registry_api_versions",
".",
"value",
"if",
"'v2'",
"not",
"in",
"versions",
":",
"raise",
"OsbsValidationException",
"(",
"'v1-only docker registry API is not supported'",
")",
"try",
":",
"push_conf",
"=",
"self",
".",
"dj",
".",
"dock_json_get_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'tag_and_push'",
")",
"tag_and_push_registries",
"=",
"push_conf",
"[",
"'args'",
"]",
"[",
"'registries'",
"]",
"except",
"(",
"KeyError",
",",
"IndexError",
")",
":",
"tag_and_push_registries",
"=",
"{",
"}",
"if",
"'v1'",
"not",
"in",
"versions",
":",
"# Remove v1-only plugins",
"for",
"phase",
",",
"name",
"in",
"[",
"(",
"'postbuild_plugins'",
",",
"'pulp_push'",
")",
"]",
":",
"logger",
".",
"info",
"(",
"\"removing v1-only plugin: %s\"",
",",
"name",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"name",
")",
"# remove extra tag_and_push config",
"self",
".",
"remove_tag_and_push_registries",
"(",
"tag_and_push_registries",
",",
"'v1'",
")",
"# Remove 'version' from tag_and_push plugin config as it's no",
"# longer needed",
"for",
"regdict",
"in",
"tag_and_push_registries",
".",
"values",
"(",
")",
":",
"if",
"'version'",
"in",
"regdict",
":",
"del",
"regdict",
"[",
"'version'",
"]"
] |
Enable/disable plugins depending on supported registry API versions
|
[
"Enable",
"/",
"disable",
"plugins",
"depending",
"on",
"supported",
"registry",
"API",
"versions"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L515-L544
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.adjust_for_triggers
|
def adjust_for_triggers(self):
"""Remove trigger-related plugins when needed
If there are no triggers defined, it's assumed the
feature is disabled and all trigger-related plugins
are removed.
If there are triggers defined, and this is a custom
base image, some trigger-related plugins do not apply.
Additionally, this method ensures that custom base
images never have triggers since triggering a base
image rebuild is not a valid scenario.
"""
triggers = self.template['spec'].get('triggers', [])
remove_plugins = [
("prebuild_plugins", "check_and_set_rebuild"),
("prebuild_plugins", "stop_autorebuild_if_disabled"),
]
should_remove = False
if triggers and (self.is_custom_base_image() or self.is_from_scratch_image()):
if self.is_custom_base_image():
msg = "removing %s from request because custom base image"
elif self.is_from_scratch_image():
msg = 'removing %s from request because FROM scratch image'
del self.template['spec']['triggers']
should_remove = True
elif not triggers:
msg = "removing %s from request because there are no triggers"
should_remove = True
if should_remove:
for when, which in remove_plugins:
logger.info(msg, which)
self.dj.remove_plugin(when, which)
|
python
|
def adjust_for_triggers(self):
"""Remove trigger-related plugins when needed
If there are no triggers defined, it's assumed the
feature is disabled and all trigger-related plugins
are removed.
If there are triggers defined, and this is a custom
base image, some trigger-related plugins do not apply.
Additionally, this method ensures that custom base
images never have triggers since triggering a base
image rebuild is not a valid scenario.
"""
triggers = self.template['spec'].get('triggers', [])
remove_plugins = [
("prebuild_plugins", "check_and_set_rebuild"),
("prebuild_plugins", "stop_autorebuild_if_disabled"),
]
should_remove = False
if triggers and (self.is_custom_base_image() or self.is_from_scratch_image()):
if self.is_custom_base_image():
msg = "removing %s from request because custom base image"
elif self.is_from_scratch_image():
msg = 'removing %s from request because FROM scratch image'
del self.template['spec']['triggers']
should_remove = True
elif not triggers:
msg = "removing %s from request because there are no triggers"
should_remove = True
if should_remove:
for when, which in remove_plugins:
logger.info(msg, which)
self.dj.remove_plugin(when, which)
|
[
"def",
"adjust_for_triggers",
"(",
"self",
")",
":",
"triggers",
"=",
"self",
".",
"template",
"[",
"'spec'",
"]",
".",
"get",
"(",
"'triggers'",
",",
"[",
"]",
")",
"remove_plugins",
"=",
"[",
"(",
"\"prebuild_plugins\"",
",",
"\"check_and_set_rebuild\"",
")",
",",
"(",
"\"prebuild_plugins\"",
",",
"\"stop_autorebuild_if_disabled\"",
")",
",",
"]",
"should_remove",
"=",
"False",
"if",
"triggers",
"and",
"(",
"self",
".",
"is_custom_base_image",
"(",
")",
"or",
"self",
".",
"is_from_scratch_image",
"(",
")",
")",
":",
"if",
"self",
".",
"is_custom_base_image",
"(",
")",
":",
"msg",
"=",
"\"removing %s from request because custom base image\"",
"elif",
"self",
".",
"is_from_scratch_image",
"(",
")",
":",
"msg",
"=",
"'removing %s from request because FROM scratch image'",
"del",
"self",
".",
"template",
"[",
"'spec'",
"]",
"[",
"'triggers'",
"]",
"should_remove",
"=",
"True",
"elif",
"not",
"triggers",
":",
"msg",
"=",
"\"removing %s from request because there are no triggers\"",
"should_remove",
"=",
"True",
"if",
"should_remove",
":",
"for",
"when",
",",
"which",
"in",
"remove_plugins",
":",
"logger",
".",
"info",
"(",
"msg",
",",
"which",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"when",
",",
"which",
")"
] |
Remove trigger-related plugins when needed
If there are no triggers defined, it's assumed the
feature is disabled and all trigger-related plugins
are removed.
If there are triggers defined, and this is a custom
base image, some trigger-related plugins do not apply.
Additionally, this method ensures that custom base
images never have triggers since triggering a base
image rebuild is not a valid scenario.
|
[
"Remove",
"trigger",
"-",
"related",
"plugins",
"when",
"needed"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L546-L583
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.adjust_for_scratch
|
def adjust_for_scratch(self):
"""
Remove certain plugins in order to handle the "scratch build"
scenario. Scratch builds must not affect subsequent builds,
and should not be imported into Koji.
"""
if self.scratch:
self.template['spec'].pop('triggers', None)
remove_plugins = [
("prebuild_plugins", "koji_parent"),
("postbuild_plugins", "compress"), # required only to make an archive for Koji
("postbuild_plugins", "pulp_pull"), # required only to make an archive for Koji
("postbuild_plugins", "koji_upload"),
("postbuild_plugins", "fetch_worker_metadata"),
("postbuild_plugins", "compare_components"),
("postbuild_plugins", "import_image"),
("exit_plugins", "koji_promote"),
("exit_plugins", "koji_import"),
("exit_plugins", "koji_tag_build"),
("exit_plugins", "remove_worker_metadata"),
("exit_plugins", "import_image"),
]
if not self.has_tag_suffixes_placeholder():
remove_plugins.append(("postbuild_plugins", "tag_from_config"))
for when, which in remove_plugins:
logger.info("removing %s from scratch build request",
which)
self.dj.remove_plugin(when, which)
if self.dj.dock_json_has_plugin_conf('postbuild_plugins',
'tag_by_labels'):
self.dj.dock_json_set_arg('postbuild_plugins', 'tag_by_labels',
'unique_tag_only', True)
self.set_label('scratch', 'true')
|
python
|
def adjust_for_scratch(self):
"""
Remove certain plugins in order to handle the "scratch build"
scenario. Scratch builds must not affect subsequent builds,
and should not be imported into Koji.
"""
if self.scratch:
self.template['spec'].pop('triggers', None)
remove_plugins = [
("prebuild_plugins", "koji_parent"),
("postbuild_plugins", "compress"), # required only to make an archive for Koji
("postbuild_plugins", "pulp_pull"), # required only to make an archive for Koji
("postbuild_plugins", "koji_upload"),
("postbuild_plugins", "fetch_worker_metadata"),
("postbuild_plugins", "compare_components"),
("postbuild_plugins", "import_image"),
("exit_plugins", "koji_promote"),
("exit_plugins", "koji_import"),
("exit_plugins", "koji_tag_build"),
("exit_plugins", "remove_worker_metadata"),
("exit_plugins", "import_image"),
]
if not self.has_tag_suffixes_placeholder():
remove_plugins.append(("postbuild_plugins", "tag_from_config"))
for when, which in remove_plugins:
logger.info("removing %s from scratch build request",
which)
self.dj.remove_plugin(when, which)
if self.dj.dock_json_has_plugin_conf('postbuild_plugins',
'tag_by_labels'):
self.dj.dock_json_set_arg('postbuild_plugins', 'tag_by_labels',
'unique_tag_only', True)
self.set_label('scratch', 'true')
|
[
"def",
"adjust_for_scratch",
"(",
"self",
")",
":",
"if",
"self",
".",
"scratch",
":",
"self",
".",
"template",
"[",
"'spec'",
"]",
".",
"pop",
"(",
"'triggers'",
",",
"None",
")",
"remove_plugins",
"=",
"[",
"(",
"\"prebuild_plugins\"",
",",
"\"koji_parent\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"compress\"",
")",
",",
"# required only to make an archive for Koji",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_pull\"",
")",
",",
"# required only to make an archive for Koji",
"(",
"\"postbuild_plugins\"",
",",
"\"koji_upload\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"fetch_worker_metadata\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"compare_components\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"import_image\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"koji_promote\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"koji_import\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"koji_tag_build\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"remove_worker_metadata\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"import_image\"",
")",
",",
"]",
"if",
"not",
"self",
".",
"has_tag_suffixes_placeholder",
"(",
")",
":",
"remove_plugins",
".",
"append",
"(",
"(",
"\"postbuild_plugins\"",
",",
"\"tag_from_config\"",
")",
")",
"for",
"when",
",",
"which",
"in",
"remove_plugins",
":",
"logger",
".",
"info",
"(",
"\"removing %s from scratch build request\"",
",",
"which",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"when",
",",
"which",
")",
"if",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'tag_by_labels'",
")",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'postbuild_plugins'",
",",
"'tag_by_labels'",
",",
"'unique_tag_only'",
",",
"True",
")",
"self",
".",
"set_label",
"(",
"'scratch'",
",",
"'true'",
")"
] |
Remove certain plugins in order to handle the "scratch build"
scenario. Scratch builds must not affect subsequent builds,
and should not be imported into Koji.
|
[
"Remove",
"certain",
"plugins",
"in",
"order",
"to",
"handle",
"the",
"scratch",
"build",
"scenario",
".",
"Scratch",
"builds",
"must",
"not",
"affect",
"subsequent",
"builds",
"and",
"should",
"not",
"be",
"imported",
"into",
"Koji",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L585-L622
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.adjust_for_custom_base_image
|
def adjust_for_custom_base_image(self):
"""
Disable plugins to handle builds depending on whether
or not this is a build from a custom base image.
"""
plugins = []
if self.is_custom_base_image():
# Plugins irrelevant to building base images.
plugins.append(("prebuild_plugins", "pull_base_image"))
plugins.append(("prebuild_plugins", "koji_parent"))
plugins.append(("prebuild_plugins", "inject_parent_image"))
msg = "removing %s from custom image build request"
else:
# Plugins not needed for building non base images.
plugins.append(("prebuild_plugins", "add_filesystem"))
msg = "removing %s from non custom image build request"
for when, which in plugins:
logger.info(msg, which)
self.dj.remove_plugin(when, which)
|
python
|
def adjust_for_custom_base_image(self):
"""
Disable plugins to handle builds depending on whether
or not this is a build from a custom base image.
"""
plugins = []
if self.is_custom_base_image():
# Plugins irrelevant to building base images.
plugins.append(("prebuild_plugins", "pull_base_image"))
plugins.append(("prebuild_plugins", "koji_parent"))
plugins.append(("prebuild_plugins", "inject_parent_image"))
msg = "removing %s from custom image build request"
else:
# Plugins not needed for building non base images.
plugins.append(("prebuild_plugins", "add_filesystem"))
msg = "removing %s from non custom image build request"
for when, which in plugins:
logger.info(msg, which)
self.dj.remove_plugin(when, which)
|
[
"def",
"adjust_for_custom_base_image",
"(",
"self",
")",
":",
"plugins",
"=",
"[",
"]",
"if",
"self",
".",
"is_custom_base_image",
"(",
")",
":",
"# Plugins irrelevant to building base images.",
"plugins",
".",
"append",
"(",
"(",
"\"prebuild_plugins\"",
",",
"\"pull_base_image\"",
")",
")",
"plugins",
".",
"append",
"(",
"(",
"\"prebuild_plugins\"",
",",
"\"koji_parent\"",
")",
")",
"plugins",
".",
"append",
"(",
"(",
"\"prebuild_plugins\"",
",",
"\"inject_parent_image\"",
")",
")",
"msg",
"=",
"\"removing %s from custom image build request\"",
"else",
":",
"# Plugins not needed for building non base images.",
"plugins",
".",
"append",
"(",
"(",
"\"prebuild_plugins\"",
",",
"\"add_filesystem\"",
")",
")",
"msg",
"=",
"\"removing %s from non custom image build request\"",
"for",
"when",
",",
"which",
"in",
"plugins",
":",
"logger",
".",
"info",
"(",
"msg",
",",
"which",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"when",
",",
"which",
")"
] |
Disable plugins to handle builds depending on whether
or not this is a build from a custom base image.
|
[
"Disable",
"plugins",
"to",
"handle",
"builds",
"depending",
"on",
"whether",
"or",
"not",
"this",
"is",
"a",
"build",
"from",
"a",
"custom",
"base",
"image",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L641-L660
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_koji
|
def render_koji(self):
"""
if there is yum repo specified, don't pick stuff from koji
"""
phase = 'prebuild_plugins'
plugin = 'koji'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
if self.spec.yum_repourls.value:
logger.info("removing koji from request "
"because there is yum repo specified")
self.dj.remove_plugin(phase, plugin)
elif not (self.spec.koji_target.value and
self.spec.kojiroot.value and
self.spec.kojihub.value):
logger.info("removing koji from request as not specified")
self.dj.remove_plugin(phase, plugin)
else:
self.dj.dock_json_set_arg(phase, plugin,
"target", self.spec.koji_target.value)
self.dj.dock_json_set_arg(phase, plugin,
"root", self.spec.kojiroot.value)
self.dj.dock_json_set_arg(phase, plugin,
"hub", self.spec.kojihub.value)
if self.spec.proxy.value:
self.dj.dock_json_set_arg(phase, plugin,
"proxy", self.spec.proxy.value)
|
python
|
def render_koji(self):
"""
if there is yum repo specified, don't pick stuff from koji
"""
phase = 'prebuild_plugins'
plugin = 'koji'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
if self.spec.yum_repourls.value:
logger.info("removing koji from request "
"because there is yum repo specified")
self.dj.remove_plugin(phase, plugin)
elif not (self.spec.koji_target.value and
self.spec.kojiroot.value and
self.spec.kojihub.value):
logger.info("removing koji from request as not specified")
self.dj.remove_plugin(phase, plugin)
else:
self.dj.dock_json_set_arg(phase, plugin,
"target", self.spec.koji_target.value)
self.dj.dock_json_set_arg(phase, plugin,
"root", self.spec.kojiroot.value)
self.dj.dock_json_set_arg(phase, plugin,
"hub", self.spec.kojihub.value)
if self.spec.proxy.value:
self.dj.dock_json_set_arg(phase, plugin,
"proxy", self.spec.proxy.value)
|
[
"def",
"render_koji",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'koji'",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"if",
"self",
".",
"spec",
".",
"yum_repourls",
".",
"value",
":",
"logger",
".",
"info",
"(",
"\"removing koji from request \"",
"\"because there is yum repo specified\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"elif",
"not",
"(",
"self",
".",
"spec",
".",
"koji_target",
".",
"value",
"and",
"self",
".",
"spec",
".",
"kojiroot",
".",
"value",
"and",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
")",
":",
"logger",
".",
"info",
"(",
"\"removing koji from request as not specified\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"else",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"\"target\"",
",",
"self",
".",
"spec",
".",
"koji_target",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"\"root\"",
",",
"self",
".",
"spec",
".",
"kojiroot",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"\"hub\"",
",",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
")",
"if",
"self",
".",
"spec",
".",
"proxy",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"\"proxy\"",
",",
"self",
".",
"spec",
".",
"proxy",
".",
"value",
")"
] |
if there is yum repo specified, don't pick stuff from koji
|
[
"if",
"there",
"is",
"yum",
"repo",
"specified",
"don",
"t",
"pick",
"stuff",
"from",
"koji"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L814-L841
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_bump_release
|
def render_bump_release(self):
"""
If the bump_release plugin is present, configure it
"""
phase = 'prebuild_plugins'
plugin = 'bump_release'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
if self.spec.release.value:
logger.info('removing %s from request as release already specified',
plugin)
self.dj.remove_plugin(phase, plugin)
return
hub = self.spec.kojihub.value
if not hub:
logger.info('removing %s from request as koji hub not specified',
plugin)
self.dj.remove_plugin(phase, plugin)
return
self.dj.dock_json_set_arg(phase, plugin, 'hub', hub)
# For flatpak, we want a name-version-release of
# <name>-<stream>-<module_build_version>.<n>, where the .<n> makes
# sure that the build is unique in Koji
if self.spec.flatpak.value:
self.dj.dock_json_set_arg(phase, plugin, 'append', True)
|
python
|
def render_bump_release(self):
"""
If the bump_release plugin is present, configure it
"""
phase = 'prebuild_plugins'
plugin = 'bump_release'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
if self.spec.release.value:
logger.info('removing %s from request as release already specified',
plugin)
self.dj.remove_plugin(phase, plugin)
return
hub = self.spec.kojihub.value
if not hub:
logger.info('removing %s from request as koji hub not specified',
plugin)
self.dj.remove_plugin(phase, plugin)
return
self.dj.dock_json_set_arg(phase, plugin, 'hub', hub)
# For flatpak, we want a name-version-release of
# <name>-<stream>-<module_build_version>.<n>, where the .<n> makes
# sure that the build is unique in Koji
if self.spec.flatpak.value:
self.dj.dock_json_set_arg(phase, plugin, 'append', True)
|
[
"def",
"render_bump_release",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'bump_release'",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"if",
"self",
".",
"spec",
".",
"release",
".",
"value",
":",
"logger",
".",
"info",
"(",
"'removing %s from request as release already specified'",
",",
"plugin",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"return",
"hub",
"=",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
"if",
"not",
"hub",
":",
"logger",
".",
"info",
"(",
"'removing %s from request as koji hub not specified'",
",",
"plugin",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"return",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'hub'",
",",
"hub",
")",
"# For flatpak, we want a name-version-release of",
"# <name>-<stream>-<module_build_version>.<n>, where the .<n> makes",
"# sure that the build is unique in Koji",
"if",
"self",
".",
"spec",
".",
"flatpak",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'append'",
",",
"True",
")"
] |
If the bump_release plugin is present, configure it
|
[
"If",
"the",
"bump_release",
"plugin",
"is",
"present",
"configure",
"it"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L843-L871
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_sendmail
|
def render_sendmail(self):
"""
if we have smtp_host and smtp_from, configure sendmail plugin,
else remove it
"""
phase = 'exit_plugins'
plugin = 'sendmail'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
if self.spec.smtp_host.value and self.spec.smtp_from.value:
self.dj.dock_json_set_arg(phase, plugin, 'url',
self.spec.builder_openshift_url.value)
self.dj.dock_json_set_arg(phase, plugin, 'smtp_host',
self.spec.smtp_host.value)
self.dj.dock_json_set_arg(phase, plugin, 'from_address',
self.spec.smtp_from.value)
else:
logger.info("removing sendmail from request, "
"requires smtp_host and smtp_from")
self.dj.remove_plugin(phase, plugin)
return
if self.spec.kojihub.value and self.spec.kojiroot.value:
self.dj.dock_json_set_arg(phase, plugin,
'koji_hub', self.spec.kojihub.value)
self.dj.dock_json_set_arg(phase, plugin,
"koji_root", self.spec.kojiroot.value)
if self.spec.smtp_to_submitter.value:
self.dj.dock_json_set_arg(phase, plugin, 'to_koji_submitter',
self.spec.smtp_to_submitter.value)
if self.spec.smtp_to_pkgowner.value:
self.dj.dock_json_set_arg(phase, plugin, 'to_koji_pkgowner',
self.spec.smtp_to_pkgowner.value)
if self.spec.smtp_additional_addresses.value:
self.dj.dock_json_set_arg(phase, plugin, 'additional_addresses',
self.spec.smtp_additional_addresses.value)
if self.spec.smtp_error_addresses.value:
self.dj.dock_json_set_arg(phase, plugin,
'error_addresses', self.spec.smtp_error_addresses.value)
if self.spec.smtp_email_domain.value:
self.dj.dock_json_set_arg(phase, plugin,
'email_domain', self.spec.smtp_email_domain.value)
|
python
|
def render_sendmail(self):
"""
if we have smtp_host and smtp_from, configure sendmail plugin,
else remove it
"""
phase = 'exit_plugins'
plugin = 'sendmail'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
if self.spec.smtp_host.value and self.spec.smtp_from.value:
self.dj.dock_json_set_arg(phase, plugin, 'url',
self.spec.builder_openshift_url.value)
self.dj.dock_json_set_arg(phase, plugin, 'smtp_host',
self.spec.smtp_host.value)
self.dj.dock_json_set_arg(phase, plugin, 'from_address',
self.spec.smtp_from.value)
else:
logger.info("removing sendmail from request, "
"requires smtp_host and smtp_from")
self.dj.remove_plugin(phase, plugin)
return
if self.spec.kojihub.value and self.spec.kojiroot.value:
self.dj.dock_json_set_arg(phase, plugin,
'koji_hub', self.spec.kojihub.value)
self.dj.dock_json_set_arg(phase, plugin,
"koji_root", self.spec.kojiroot.value)
if self.spec.smtp_to_submitter.value:
self.dj.dock_json_set_arg(phase, plugin, 'to_koji_submitter',
self.spec.smtp_to_submitter.value)
if self.spec.smtp_to_pkgowner.value:
self.dj.dock_json_set_arg(phase, plugin, 'to_koji_pkgowner',
self.spec.smtp_to_pkgowner.value)
if self.spec.smtp_additional_addresses.value:
self.dj.dock_json_set_arg(phase, plugin, 'additional_addresses',
self.spec.smtp_additional_addresses.value)
if self.spec.smtp_error_addresses.value:
self.dj.dock_json_set_arg(phase, plugin,
'error_addresses', self.spec.smtp_error_addresses.value)
if self.spec.smtp_email_domain.value:
self.dj.dock_json_set_arg(phase, plugin,
'email_domain', self.spec.smtp_email_domain.value)
|
[
"def",
"render_sendmail",
"(",
"self",
")",
":",
"phase",
"=",
"'exit_plugins'",
"plugin",
"=",
"'sendmail'",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"if",
"self",
".",
"spec",
".",
"smtp_host",
".",
"value",
"and",
"self",
".",
"spec",
".",
"smtp_from",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'url'",
",",
"self",
".",
"spec",
".",
"builder_openshift_url",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'smtp_host'",
",",
"self",
".",
"spec",
".",
"smtp_host",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'from_address'",
",",
"self",
".",
"spec",
".",
"smtp_from",
".",
"value",
")",
"else",
":",
"logger",
".",
"info",
"(",
"\"removing sendmail from request, \"",
"\"requires smtp_host and smtp_from\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"return",
"if",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
"and",
"self",
".",
"spec",
".",
"kojiroot",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'koji_hub'",
",",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"\"koji_root\"",
",",
"self",
".",
"spec",
".",
"kojiroot",
".",
"value",
")",
"if",
"self",
".",
"spec",
".",
"smtp_to_submitter",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'to_koji_submitter'",
",",
"self",
".",
"spec",
".",
"smtp_to_submitter",
".",
"value",
")",
"if",
"self",
".",
"spec",
".",
"smtp_to_pkgowner",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'to_koji_pkgowner'",
",",
"self",
".",
"spec",
".",
"smtp_to_pkgowner",
".",
"value",
")",
"if",
"self",
".",
"spec",
".",
"smtp_additional_addresses",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'additional_addresses'",
",",
"self",
".",
"spec",
".",
"smtp_additional_addresses",
".",
"value",
")",
"if",
"self",
".",
"spec",
".",
"smtp_error_addresses",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'error_addresses'",
",",
"self",
".",
"spec",
".",
"smtp_error_addresses",
".",
"value",
")",
"if",
"self",
".",
"spec",
".",
"smtp_email_domain",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'email_domain'",
",",
"self",
".",
"spec",
".",
"smtp_email_domain",
".",
"value",
")"
] |
if we have smtp_host and smtp_from, configure sendmail plugin,
else remove it
|
[
"if",
"we",
"have",
"smtp_host",
"and",
"smtp_from",
"configure",
"sendmail",
"plugin",
"else",
"remove",
"it"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L986-L1032
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_fetch_maven_artifacts
|
def render_fetch_maven_artifacts(self):
"""Configure fetch_maven_artifacts plugin"""
phase = 'prebuild_plugins'
plugin = 'fetch_maven_artifacts'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
koji_hub = self.spec.kojihub.value
koji_root = self.spec.kojiroot.value
if not koji_hub and not koji_root:
logger.info('Removing %s because kojihub and kojiroot were not specified', plugin)
self.dj.remove_plugin(phase, plugin)
return
self.dj.dock_json_set_arg(phase, plugin, 'koji_hub', koji_hub)
self.dj.dock_json_set_arg(phase, plugin, "koji_root", koji_root)
if self.spec.artifacts_allowed_domains.value:
self.dj.dock_json_set_arg(phase, plugin, 'allowed_domains',
self.spec.artifacts_allowed_domains.value)
|
python
|
def render_fetch_maven_artifacts(self):
"""Configure fetch_maven_artifacts plugin"""
phase = 'prebuild_plugins'
plugin = 'fetch_maven_artifacts'
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
return
koji_hub = self.spec.kojihub.value
koji_root = self.spec.kojiroot.value
if not koji_hub and not koji_root:
logger.info('Removing %s because kojihub and kojiroot were not specified', plugin)
self.dj.remove_plugin(phase, plugin)
return
self.dj.dock_json_set_arg(phase, plugin, 'koji_hub', koji_hub)
self.dj.dock_json_set_arg(phase, plugin, "koji_root", koji_root)
if self.spec.artifacts_allowed_domains.value:
self.dj.dock_json_set_arg(phase, plugin, 'allowed_domains',
self.spec.artifacts_allowed_domains.value)
|
[
"def",
"render_fetch_maven_artifacts",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'fetch_maven_artifacts'",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"koji_hub",
"=",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
"koji_root",
"=",
"self",
".",
"spec",
".",
"kojiroot",
".",
"value",
"if",
"not",
"koji_hub",
"and",
"not",
"koji_root",
":",
"logger",
".",
"info",
"(",
"'Removing %s because kojihub and kojiroot were not specified'",
",",
"plugin",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"return",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'koji_hub'",
",",
"koji_hub",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"\"koji_root\"",
",",
"koji_root",
")",
"if",
"self",
".",
"spec",
".",
"artifacts_allowed_domains",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'allowed_domains'",
",",
"self",
".",
"spec",
".",
"artifacts_allowed_domains",
".",
"value",
")"
] |
Configure fetch_maven_artifacts plugin
|
[
"Configure",
"fetch_maven_artifacts",
"plugin"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1034-L1054
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_tag_from_config
|
def render_tag_from_config(self):
"""Configure tag_from_config plugin"""
phase = 'postbuild_plugins'
plugin = 'tag_from_config'
if not self.has_tag_suffixes_placeholder():
return
unique_tag = self.spec.image_tag.value.split(':')[-1]
tag_suffixes = {'unique': [unique_tag], 'primary': []}
if self.spec.build_type.value == BUILD_TYPE_ORCHESTRATOR:
if self.scratch:
pass
elif self.isolated:
tag_suffixes['primary'].extend(['{version}-{release}'])
elif self._repo_info.additional_tags.from_container_yaml:
tag_suffixes['primary'].extend(['{version}-{release}'])
tag_suffixes['primary'].extend(self._repo_info.additional_tags.tags)
else:
tag_suffixes['primary'].extend(['latest', '{version}', '{version}-{release}'])
tag_suffixes['primary'].extend(self._repo_info.additional_tags.tags)
self.dj.dock_json_set_arg(phase, plugin, 'tag_suffixes', tag_suffixes)
|
python
|
def render_tag_from_config(self):
"""Configure tag_from_config plugin"""
phase = 'postbuild_plugins'
plugin = 'tag_from_config'
if not self.has_tag_suffixes_placeholder():
return
unique_tag = self.spec.image_tag.value.split(':')[-1]
tag_suffixes = {'unique': [unique_tag], 'primary': []}
if self.spec.build_type.value == BUILD_TYPE_ORCHESTRATOR:
if self.scratch:
pass
elif self.isolated:
tag_suffixes['primary'].extend(['{version}-{release}'])
elif self._repo_info.additional_tags.from_container_yaml:
tag_suffixes['primary'].extend(['{version}-{release}'])
tag_suffixes['primary'].extend(self._repo_info.additional_tags.tags)
else:
tag_suffixes['primary'].extend(['latest', '{version}', '{version}-{release}'])
tag_suffixes['primary'].extend(self._repo_info.additional_tags.tags)
self.dj.dock_json_set_arg(phase, plugin, 'tag_suffixes', tag_suffixes)
|
[
"def",
"render_tag_from_config",
"(",
"self",
")",
":",
"phase",
"=",
"'postbuild_plugins'",
"plugin",
"=",
"'tag_from_config'",
"if",
"not",
"self",
".",
"has_tag_suffixes_placeholder",
"(",
")",
":",
"return",
"unique_tag",
"=",
"self",
".",
"spec",
".",
"image_tag",
".",
"value",
".",
"split",
"(",
"':'",
")",
"[",
"-",
"1",
"]",
"tag_suffixes",
"=",
"{",
"'unique'",
":",
"[",
"unique_tag",
"]",
",",
"'primary'",
":",
"[",
"]",
"}",
"if",
"self",
".",
"spec",
".",
"build_type",
".",
"value",
"==",
"BUILD_TYPE_ORCHESTRATOR",
":",
"if",
"self",
".",
"scratch",
":",
"pass",
"elif",
"self",
".",
"isolated",
":",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"[",
"'{version}-{release}'",
"]",
")",
"elif",
"self",
".",
"_repo_info",
".",
"additional_tags",
".",
"from_container_yaml",
":",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"[",
"'{version}-{release}'",
"]",
")",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"self",
".",
"_repo_info",
".",
"additional_tags",
".",
"tags",
")",
"else",
":",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"[",
"'latest'",
",",
"'{version}'",
",",
"'{version}-{release}'",
"]",
")",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"self",
".",
"_repo_info",
".",
"additional_tags",
".",
"tags",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'tag_suffixes'",
",",
"tag_suffixes",
")"
] |
Configure tag_from_config plugin
|
[
"Configure",
"tag_from_config",
"plugin"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1056-L1078
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_pulp_pull
|
def render_pulp_pull(self):
"""
If a pulp registry is specified, use pulp_pull plugin
"""
# pulp_pull is a multi-phase plugin
phases = ('postbuild_plugins', 'exit_plugins')
plugin = 'pulp_pull'
for phase in phases:
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
continue
pulp_registry = self.spec.pulp_registry.value
if not pulp_registry:
logger.info("removing %s from request, requires pulp_registry", pulp_registry)
self.dj.remove_plugin(phase, plugin)
continue
if not self.spec.kojihub.value:
logger.info('Removing %s because no kojihub was specified', plugin)
self.dj.remove_plugin(phase, plugin)
continue
if self.spec.prefer_schema1_digest.value is not None:
self.dj.dock_json_set_arg(phase, 'pulp_pull',
'expect_v2schema2',
not self.spec.prefer_schema1_digest.value)
|
python
|
def render_pulp_pull(self):
"""
If a pulp registry is specified, use pulp_pull plugin
"""
# pulp_pull is a multi-phase plugin
phases = ('postbuild_plugins', 'exit_plugins')
plugin = 'pulp_pull'
for phase in phases:
if not self.dj.dock_json_has_plugin_conf(phase, plugin):
continue
pulp_registry = self.spec.pulp_registry.value
if not pulp_registry:
logger.info("removing %s from request, requires pulp_registry", pulp_registry)
self.dj.remove_plugin(phase, plugin)
continue
if not self.spec.kojihub.value:
logger.info('Removing %s because no kojihub was specified', plugin)
self.dj.remove_plugin(phase, plugin)
continue
if self.spec.prefer_schema1_digest.value is not None:
self.dj.dock_json_set_arg(phase, 'pulp_pull',
'expect_v2schema2',
not self.spec.prefer_schema1_digest.value)
|
[
"def",
"render_pulp_pull",
"(",
"self",
")",
":",
"# pulp_pull is a multi-phase plugin",
"phases",
"=",
"(",
"'postbuild_plugins'",
",",
"'exit_plugins'",
")",
"plugin",
"=",
"'pulp_pull'",
"for",
"phase",
"in",
"phases",
":",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"continue",
"pulp_registry",
"=",
"self",
".",
"spec",
".",
"pulp_registry",
".",
"value",
"if",
"not",
"pulp_registry",
":",
"logger",
".",
"info",
"(",
"\"removing %s from request, requires pulp_registry\"",
",",
"pulp_registry",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"continue",
"if",
"not",
"self",
".",
"spec",
".",
"kojihub",
".",
"value",
":",
"logger",
".",
"info",
"(",
"'Removing %s because no kojihub was specified'",
",",
"plugin",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"continue",
"if",
"self",
".",
"spec",
".",
"prefer_schema1_digest",
".",
"value",
"is",
"not",
"None",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"'pulp_pull'",
",",
"'expect_v2schema2'",
",",
"not",
"self",
".",
"spec",
".",
"prefer_schema1_digest",
".",
"value",
")"
] |
If a pulp registry is specified, use pulp_pull plugin
|
[
"If",
"a",
"pulp",
"registry",
"is",
"specified",
"use",
"pulp_pull",
"plugin"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1080-L1105
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_pulp_sync
|
def render_pulp_sync(self):
"""
If a pulp registry is specified, use the pulp plugin as well as the
delete_from_registry to delete the image after sync
"""
if not self.dj.dock_json_has_plugin_conf('postbuild_plugins',
'pulp_sync'):
return
pulp_registry = self.spec.pulp_registry.value
# Find which registry to use
docker_registry = None
registry_secret = None
registries = zip_longest(self.spec.registry_uris.value,
self.spec.registry_secrets.value)
for registry, secret in registries:
if registry.version == 'v2':
# First specified v2 registry is the one we'll tell pulp
# to sync from. Keep the http prefix -- pulp wants it.
docker_registry = registry.uri
registry_secret = secret
logger.info("using docker v2 registry %s for pulp_sync",
docker_registry)
break
if pulp_registry and docker_registry:
self.dj.dock_json_set_arg('postbuild_plugins', 'pulp_sync',
'pulp_registry_name', pulp_registry)
self.dj.dock_json_set_arg('postbuild_plugins', 'pulp_sync',
'docker_registry', docker_registry)
if registry_secret:
self.set_secret_for_plugin(registry_secret,
plugin=('postbuild_plugins',
'pulp_sync',
'registry_secret_path'))
# Verify we have a pulp secret
if self.spec.pulp_secret.value is None:
raise OsbsValidationException("Pulp registry specified "
"but no auth config")
source_registry = self.spec.source_registry_uri.value
perform_delete = (source_registry is None or
source_registry.docker_uri != registry.docker_uri)
if perform_delete:
push_conf = self.dj.dock_json_get_plugin_conf('exit_plugins',
'delete_from_registry')
args = push_conf.setdefault('args', {})
delete_registries = args.setdefault('registries', {})
placeholder = '{{REGISTRY_URI}}'
# use passed in params like 'insecure' if available
if placeholder in delete_registries:
regdict = delete_registries[placeholder].copy()
del delete_registries[placeholder]
else:
regdict = {}
if registry_secret:
regdict['secret'] = \
os.path.join(SECRETS_PATH, registry_secret)
# tag_and_push configured the registry secret, no neet to set it again
delete_registries[docker_registry] = regdict
self.dj.dock_json_set_arg('exit_plugins', 'delete_from_registry',
'registries', delete_registries)
else:
logger.info("removing delete_from_registry from request, "
"source and target registry are identical")
self.dj.remove_plugin("exit_plugins", "delete_from_registry")
else:
# If no pulp registry is specified, don't run the pulp plugin
logger.info("removing pulp_sync+delete_from_registry from request, "
"requires pulp_registry and a v2 registry")
self.dj.remove_plugin("postbuild_plugins", "pulp_sync")
self.dj.remove_plugin("exit_plugins", "delete_from_registry")
|
python
|
def render_pulp_sync(self):
"""
If a pulp registry is specified, use the pulp plugin as well as the
delete_from_registry to delete the image after sync
"""
if not self.dj.dock_json_has_plugin_conf('postbuild_plugins',
'pulp_sync'):
return
pulp_registry = self.spec.pulp_registry.value
# Find which registry to use
docker_registry = None
registry_secret = None
registries = zip_longest(self.spec.registry_uris.value,
self.spec.registry_secrets.value)
for registry, secret in registries:
if registry.version == 'v2':
# First specified v2 registry is the one we'll tell pulp
# to sync from. Keep the http prefix -- pulp wants it.
docker_registry = registry.uri
registry_secret = secret
logger.info("using docker v2 registry %s for pulp_sync",
docker_registry)
break
if pulp_registry and docker_registry:
self.dj.dock_json_set_arg('postbuild_plugins', 'pulp_sync',
'pulp_registry_name', pulp_registry)
self.dj.dock_json_set_arg('postbuild_plugins', 'pulp_sync',
'docker_registry', docker_registry)
if registry_secret:
self.set_secret_for_plugin(registry_secret,
plugin=('postbuild_plugins',
'pulp_sync',
'registry_secret_path'))
# Verify we have a pulp secret
if self.spec.pulp_secret.value is None:
raise OsbsValidationException("Pulp registry specified "
"but no auth config")
source_registry = self.spec.source_registry_uri.value
perform_delete = (source_registry is None or
source_registry.docker_uri != registry.docker_uri)
if perform_delete:
push_conf = self.dj.dock_json_get_plugin_conf('exit_plugins',
'delete_from_registry')
args = push_conf.setdefault('args', {})
delete_registries = args.setdefault('registries', {})
placeholder = '{{REGISTRY_URI}}'
# use passed in params like 'insecure' if available
if placeholder in delete_registries:
regdict = delete_registries[placeholder].copy()
del delete_registries[placeholder]
else:
regdict = {}
if registry_secret:
regdict['secret'] = \
os.path.join(SECRETS_PATH, registry_secret)
# tag_and_push configured the registry secret, no neet to set it again
delete_registries[docker_registry] = regdict
self.dj.dock_json_set_arg('exit_plugins', 'delete_from_registry',
'registries', delete_registries)
else:
logger.info("removing delete_from_registry from request, "
"source and target registry are identical")
self.dj.remove_plugin("exit_plugins", "delete_from_registry")
else:
# If no pulp registry is specified, don't run the pulp plugin
logger.info("removing pulp_sync+delete_from_registry from request, "
"requires pulp_registry and a v2 registry")
self.dj.remove_plugin("postbuild_plugins", "pulp_sync")
self.dj.remove_plugin("exit_plugins", "delete_from_registry")
|
[
"def",
"render_pulp_sync",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'pulp_sync'",
")",
":",
"return",
"pulp_registry",
"=",
"self",
".",
"spec",
".",
"pulp_registry",
".",
"value",
"# Find which registry to use",
"docker_registry",
"=",
"None",
"registry_secret",
"=",
"None",
"registries",
"=",
"zip_longest",
"(",
"self",
".",
"spec",
".",
"registry_uris",
".",
"value",
",",
"self",
".",
"spec",
".",
"registry_secrets",
".",
"value",
")",
"for",
"registry",
",",
"secret",
"in",
"registries",
":",
"if",
"registry",
".",
"version",
"==",
"'v2'",
":",
"# First specified v2 registry is the one we'll tell pulp",
"# to sync from. Keep the http prefix -- pulp wants it.",
"docker_registry",
"=",
"registry",
".",
"uri",
"registry_secret",
"=",
"secret",
"logger",
".",
"info",
"(",
"\"using docker v2 registry %s for pulp_sync\"",
",",
"docker_registry",
")",
"break",
"if",
"pulp_registry",
"and",
"docker_registry",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'postbuild_plugins'",
",",
"'pulp_sync'",
",",
"'pulp_registry_name'",
",",
"pulp_registry",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'postbuild_plugins'",
",",
"'pulp_sync'",
",",
"'docker_registry'",
",",
"docker_registry",
")",
"if",
"registry_secret",
":",
"self",
".",
"set_secret_for_plugin",
"(",
"registry_secret",
",",
"plugin",
"=",
"(",
"'postbuild_plugins'",
",",
"'pulp_sync'",
",",
"'registry_secret_path'",
")",
")",
"# Verify we have a pulp secret",
"if",
"self",
".",
"spec",
".",
"pulp_secret",
".",
"value",
"is",
"None",
":",
"raise",
"OsbsValidationException",
"(",
"\"Pulp registry specified \"",
"\"but no auth config\"",
")",
"source_registry",
"=",
"self",
".",
"spec",
".",
"source_registry_uri",
".",
"value",
"perform_delete",
"=",
"(",
"source_registry",
"is",
"None",
"or",
"source_registry",
".",
"docker_uri",
"!=",
"registry",
".",
"docker_uri",
")",
"if",
"perform_delete",
":",
"push_conf",
"=",
"self",
".",
"dj",
".",
"dock_json_get_plugin_conf",
"(",
"'exit_plugins'",
",",
"'delete_from_registry'",
")",
"args",
"=",
"push_conf",
".",
"setdefault",
"(",
"'args'",
",",
"{",
"}",
")",
"delete_registries",
"=",
"args",
".",
"setdefault",
"(",
"'registries'",
",",
"{",
"}",
")",
"placeholder",
"=",
"'{{REGISTRY_URI}}'",
"# use passed in params like 'insecure' if available",
"if",
"placeholder",
"in",
"delete_registries",
":",
"regdict",
"=",
"delete_registries",
"[",
"placeholder",
"]",
".",
"copy",
"(",
")",
"del",
"delete_registries",
"[",
"placeholder",
"]",
"else",
":",
"regdict",
"=",
"{",
"}",
"if",
"registry_secret",
":",
"regdict",
"[",
"'secret'",
"]",
"=",
"os",
".",
"path",
".",
"join",
"(",
"SECRETS_PATH",
",",
"registry_secret",
")",
"# tag_and_push configured the registry secret, no neet to set it again",
"delete_registries",
"[",
"docker_registry",
"]",
"=",
"regdict",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'exit_plugins'",
",",
"'delete_from_registry'",
",",
"'registries'",
",",
"delete_registries",
")",
"else",
":",
"logger",
".",
"info",
"(",
"\"removing delete_from_registry from request, \"",
"\"source and target registry are identical\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"\"exit_plugins\"",
",",
"\"delete_from_registry\"",
")",
"else",
":",
"# If no pulp registry is specified, don't run the pulp plugin",
"logger",
".",
"info",
"(",
"\"removing pulp_sync+delete_from_registry from request, \"",
"\"requires pulp_registry and a v2 registry\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_sync\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"\"exit_plugins\"",
",",
"\"delete_from_registry\"",
")"
] |
If a pulp registry is specified, use the pulp plugin as well as the
delete_from_registry to delete the image after sync
|
[
"If",
"a",
"pulp",
"registry",
"is",
"specified",
"use",
"the",
"pulp",
"plugin",
"as",
"well",
"as",
"the",
"delete_from_registry",
"to",
"delete",
"the",
"image",
"after",
"sync"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1134-L1213
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_pulp_tag
|
def render_pulp_tag(self):
"""
Configure the pulp_tag plugin.
"""
if not self.dj.dock_json_has_plugin_conf('postbuild_plugins',
'pulp_tag'):
return
pulp_registry = self.spec.pulp_registry.value
if pulp_registry:
self.dj.dock_json_set_arg('postbuild_plugins', 'pulp_tag',
'pulp_registry_name', pulp_registry)
# Verify we have either a secret or username/password
if self.spec.pulp_secret.value is None:
conf = self.dj.dock_json_get_plugin_conf('postbuild_plugins',
'pulp_tag')
args = conf.get('args', {})
if 'username' not in args:
raise OsbsValidationException("Pulp registry specified "
"but no auth config")
else:
# If no pulp registry is specified, don't run the pulp plugin
logger.info("removing pulp_tag from request, "
"requires pulp_registry")
self.dj.remove_plugin("postbuild_plugins", "pulp_tag")
|
python
|
def render_pulp_tag(self):
"""
Configure the pulp_tag plugin.
"""
if not self.dj.dock_json_has_plugin_conf('postbuild_plugins',
'pulp_tag'):
return
pulp_registry = self.spec.pulp_registry.value
if pulp_registry:
self.dj.dock_json_set_arg('postbuild_plugins', 'pulp_tag',
'pulp_registry_name', pulp_registry)
# Verify we have either a secret or username/password
if self.spec.pulp_secret.value is None:
conf = self.dj.dock_json_get_plugin_conf('postbuild_plugins',
'pulp_tag')
args = conf.get('args', {})
if 'username' not in args:
raise OsbsValidationException("Pulp registry specified "
"but no auth config")
else:
# If no pulp registry is specified, don't run the pulp plugin
logger.info("removing pulp_tag from request, "
"requires pulp_registry")
self.dj.remove_plugin("postbuild_plugins", "pulp_tag")
|
[
"def",
"render_pulp_tag",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'pulp_tag'",
")",
":",
"return",
"pulp_registry",
"=",
"self",
".",
"spec",
".",
"pulp_registry",
".",
"value",
"if",
"pulp_registry",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'postbuild_plugins'",
",",
"'pulp_tag'",
",",
"'pulp_registry_name'",
",",
"pulp_registry",
")",
"# Verify we have either a secret or username/password",
"if",
"self",
".",
"spec",
".",
"pulp_secret",
".",
"value",
"is",
"None",
":",
"conf",
"=",
"self",
".",
"dj",
".",
"dock_json_get_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'pulp_tag'",
")",
"args",
"=",
"conf",
".",
"get",
"(",
"'args'",
",",
"{",
"}",
")",
"if",
"'username'",
"not",
"in",
"args",
":",
"raise",
"OsbsValidationException",
"(",
"\"Pulp registry specified \"",
"\"but no auth config\"",
")",
"else",
":",
"# If no pulp registry is specified, don't run the pulp plugin",
"logger",
".",
"info",
"(",
"\"removing pulp_tag from request, \"",
"\"requires pulp_registry\"",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_tag\"",
")"
] |
Configure the pulp_tag plugin.
|
[
"Configure",
"the",
"pulp_tag",
"plugin",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1215-L1240
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_group_manifests
|
def render_group_manifests(self):
"""
Configure the group_manifests plugin. Group is always set to false for now.
"""
if not self.dj.dock_json_has_plugin_conf('postbuild_plugins', 'group_manifests'):
return
push_conf = self.dj.dock_json_get_plugin_conf('postbuild_plugins',
'group_manifests')
args = push_conf.setdefault('args', {})
# modify registries in place
registries = args.setdefault('registries', {})
placeholder = '{{REGISTRY_URI}}'
if placeholder in registries:
for registry, secret in zip_longest(self.spec.registry_uris.value,
self.spec.registry_secrets.value):
if not registry.uri:
continue
regdict = registries[placeholder].copy()
regdict['version'] = registry.version
if secret:
regdict['secret'] = os.path.join(SECRETS_PATH, secret)
registries[registry.docker_uri] = regdict
del registries[placeholder]
self.dj.dock_json_set_arg('postbuild_plugins', 'group_manifests',
'group', self.spec.group_manifests.value)
goarch = {}
for platform, architecture in self.platform_descriptors.items():
goarch[platform] = architecture['architecture']
self.dj.dock_json_set_arg('postbuild_plugins', 'group_manifests',
'goarch', goarch)
|
python
|
def render_group_manifests(self):
"""
Configure the group_manifests plugin. Group is always set to false for now.
"""
if not self.dj.dock_json_has_plugin_conf('postbuild_plugins', 'group_manifests'):
return
push_conf = self.dj.dock_json_get_plugin_conf('postbuild_plugins',
'group_manifests')
args = push_conf.setdefault('args', {})
# modify registries in place
registries = args.setdefault('registries', {})
placeholder = '{{REGISTRY_URI}}'
if placeholder in registries:
for registry, secret in zip_longest(self.spec.registry_uris.value,
self.spec.registry_secrets.value):
if not registry.uri:
continue
regdict = registries[placeholder].copy()
regdict['version'] = registry.version
if secret:
regdict['secret'] = os.path.join(SECRETS_PATH, secret)
registries[registry.docker_uri] = regdict
del registries[placeholder]
self.dj.dock_json_set_arg('postbuild_plugins', 'group_manifests',
'group', self.spec.group_manifests.value)
goarch = {}
for platform, architecture in self.platform_descriptors.items():
goarch[platform] = architecture['architecture']
self.dj.dock_json_set_arg('postbuild_plugins', 'group_manifests',
'goarch', goarch)
|
[
"def",
"render_group_manifests",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'group_manifests'",
")",
":",
"return",
"push_conf",
"=",
"self",
".",
"dj",
".",
"dock_json_get_plugin_conf",
"(",
"'postbuild_plugins'",
",",
"'group_manifests'",
")",
"args",
"=",
"push_conf",
".",
"setdefault",
"(",
"'args'",
",",
"{",
"}",
")",
"# modify registries in place",
"registries",
"=",
"args",
".",
"setdefault",
"(",
"'registries'",
",",
"{",
"}",
")",
"placeholder",
"=",
"'{{REGISTRY_URI}}'",
"if",
"placeholder",
"in",
"registries",
":",
"for",
"registry",
",",
"secret",
"in",
"zip_longest",
"(",
"self",
".",
"spec",
".",
"registry_uris",
".",
"value",
",",
"self",
".",
"spec",
".",
"registry_secrets",
".",
"value",
")",
":",
"if",
"not",
"registry",
".",
"uri",
":",
"continue",
"regdict",
"=",
"registries",
"[",
"placeholder",
"]",
".",
"copy",
"(",
")",
"regdict",
"[",
"'version'",
"]",
"=",
"registry",
".",
"version",
"if",
"secret",
":",
"regdict",
"[",
"'secret'",
"]",
"=",
"os",
".",
"path",
".",
"join",
"(",
"SECRETS_PATH",
",",
"secret",
")",
"registries",
"[",
"registry",
".",
"docker_uri",
"]",
"=",
"regdict",
"del",
"registries",
"[",
"placeholder",
"]",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'postbuild_plugins'",
",",
"'group_manifests'",
",",
"'group'",
",",
"self",
".",
"spec",
".",
"group_manifests",
".",
"value",
")",
"goarch",
"=",
"{",
"}",
"for",
"platform",
",",
"architecture",
"in",
"self",
".",
"platform_descriptors",
".",
"items",
"(",
")",
":",
"goarch",
"[",
"platform",
"]",
"=",
"architecture",
"[",
"'architecture'",
"]",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"'postbuild_plugins'",
",",
"'group_manifests'",
",",
"'goarch'",
",",
"goarch",
")"
] |
Configure the group_manifests plugin. Group is always set to false for now.
|
[
"Configure",
"the",
"group_manifests",
"plugin",
".",
"Group",
"is",
"always",
"set",
"to",
"false",
"for",
"now",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1269-L1301
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_import_image
|
def render_import_image(self, use_auth=None):
"""
Configure the import_image plugin
"""
# import_image is a multi-phase plugin
phases = ('postbuild_plugins', 'exit_plugins')
plugin = 'import_image'
for phase in phases:
if self.spec.imagestream_name.value is None or self.spec.imagestream_url.value is None:
logger.info("removing %s from request, "
"registry or repo url is not defined", plugin)
self.dj.remove_plugin(phase, plugin)
continue
if self.dj.dock_json_has_plugin_conf(phase, plugin):
self.dj.dock_json_set_arg(phase, plugin, 'imagestream',
self.spec.imagestream_name.value)
self.dj.dock_json_set_arg(phase, plugin, 'docker_image_repo',
self.spec.imagestream_url.value)
self.dj.dock_json_set_arg(phase, plugin, 'url',
self.spec.builder_openshift_url.value)
self.dj.dock_json_set_arg(phase, plugin, 'build_json_dir',
self.spec.builder_build_json_dir.value)
use_auth = self.spec.use_auth.value
if use_auth is not None:
self.dj.dock_json_set_arg(phase, plugin, 'use_auth', use_auth)
if self.spec.imagestream_insecure_registry.value:
self.dj.dock_json_set_arg(phase, plugin, 'insecure_registry', True)
|
python
|
def render_import_image(self, use_auth=None):
"""
Configure the import_image plugin
"""
# import_image is a multi-phase plugin
phases = ('postbuild_plugins', 'exit_plugins')
plugin = 'import_image'
for phase in phases:
if self.spec.imagestream_name.value is None or self.spec.imagestream_url.value is None:
logger.info("removing %s from request, "
"registry or repo url is not defined", plugin)
self.dj.remove_plugin(phase, plugin)
continue
if self.dj.dock_json_has_plugin_conf(phase, plugin):
self.dj.dock_json_set_arg(phase, plugin, 'imagestream',
self.spec.imagestream_name.value)
self.dj.dock_json_set_arg(phase, plugin, 'docker_image_repo',
self.spec.imagestream_url.value)
self.dj.dock_json_set_arg(phase, plugin, 'url',
self.spec.builder_openshift_url.value)
self.dj.dock_json_set_arg(phase, plugin, 'build_json_dir',
self.spec.builder_build_json_dir.value)
use_auth = self.spec.use_auth.value
if use_auth is not None:
self.dj.dock_json_set_arg(phase, plugin, 'use_auth', use_auth)
if self.spec.imagestream_insecure_registry.value:
self.dj.dock_json_set_arg(phase, plugin, 'insecure_registry', True)
|
[
"def",
"render_import_image",
"(",
"self",
",",
"use_auth",
"=",
"None",
")",
":",
"# import_image is a multi-phase plugin",
"phases",
"=",
"(",
"'postbuild_plugins'",
",",
"'exit_plugins'",
")",
"plugin",
"=",
"'import_image'",
"for",
"phase",
"in",
"phases",
":",
"if",
"self",
".",
"spec",
".",
"imagestream_name",
".",
"value",
"is",
"None",
"or",
"self",
".",
"spec",
".",
"imagestream_url",
".",
"value",
"is",
"None",
":",
"logger",
".",
"info",
"(",
"\"removing %s from request, \"",
"\"registry or repo url is not defined\"",
",",
"plugin",
")",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
")",
"continue",
"if",
"self",
".",
"dj",
".",
"dock_json_has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'imagestream'",
",",
"self",
".",
"spec",
".",
"imagestream_name",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'docker_image_repo'",
",",
"self",
".",
"spec",
".",
"imagestream_url",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'url'",
",",
"self",
".",
"spec",
".",
"builder_openshift_url",
".",
"value",
")",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'build_json_dir'",
",",
"self",
".",
"spec",
".",
"builder_build_json_dir",
".",
"value",
")",
"use_auth",
"=",
"self",
".",
"spec",
".",
"use_auth",
".",
"value",
"if",
"use_auth",
"is",
"not",
"None",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'use_auth'",
",",
"use_auth",
")",
"if",
"self",
".",
"spec",
".",
"imagestream_insecure_registry",
".",
"value",
":",
"self",
".",
"dj",
".",
"dock_json_set_arg",
"(",
"phase",
",",
"plugin",
",",
"'insecure_registry'",
",",
"True",
")"
] |
Configure the import_image plugin
|
[
"Configure",
"the",
"import_image",
"plugin"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1303-L1333
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_customizations
|
def render_customizations(self):
"""
Customize prod_inner for site specific customizations
"""
disable_plugins = self.customize_conf.get('disable_plugins', [])
if not disable_plugins:
logger.debug("No site-specific plugins to disable")
else:
for plugin_dict in disable_plugins:
try:
self.dj.remove_plugin(
plugin_dict['plugin_type'],
plugin_dict['plugin_name']
)
logger.debug(
"site-specific plugin disabled -> Type:{} Name:{}".format(
plugin_dict['plugin_type'],
plugin_dict['plugin_name']
)
)
except KeyError:
# Malformed config
logger.debug("Invalid custom configuration found for disable_plugins")
enable_plugins = self.customize_conf.get('enable_plugins', [])
if not enable_plugins:
logger.debug("No site-specific plugins to enable")
else:
for plugin_dict in enable_plugins:
try:
self.dj.add_plugin(
plugin_dict['plugin_type'],
plugin_dict['plugin_name'],
plugin_dict['plugin_args']
)
logger.debug(
"site-specific plugin enabled -> Type:{} Name:{} Args: {}".format(
plugin_dict['plugin_type'],
plugin_dict['plugin_name'],
plugin_dict['plugin_args']
)
)
except KeyError:
# Malformed config
logger.debug("Invalid custom configuration found for enable_plugins")
|
python
|
def render_customizations(self):
"""
Customize prod_inner for site specific customizations
"""
disable_plugins = self.customize_conf.get('disable_plugins', [])
if not disable_plugins:
logger.debug("No site-specific plugins to disable")
else:
for plugin_dict in disable_plugins:
try:
self.dj.remove_plugin(
plugin_dict['plugin_type'],
plugin_dict['plugin_name']
)
logger.debug(
"site-specific plugin disabled -> Type:{} Name:{}".format(
plugin_dict['plugin_type'],
plugin_dict['plugin_name']
)
)
except KeyError:
# Malformed config
logger.debug("Invalid custom configuration found for disable_plugins")
enable_plugins = self.customize_conf.get('enable_plugins', [])
if not enable_plugins:
logger.debug("No site-specific plugins to enable")
else:
for plugin_dict in enable_plugins:
try:
self.dj.add_plugin(
plugin_dict['plugin_type'],
plugin_dict['plugin_name'],
plugin_dict['plugin_args']
)
logger.debug(
"site-specific plugin enabled -> Type:{} Name:{} Args: {}".format(
plugin_dict['plugin_type'],
plugin_dict['plugin_name'],
plugin_dict['plugin_args']
)
)
except KeyError:
# Malformed config
logger.debug("Invalid custom configuration found for enable_plugins")
|
[
"def",
"render_customizations",
"(",
"self",
")",
":",
"disable_plugins",
"=",
"self",
".",
"customize_conf",
".",
"get",
"(",
"'disable_plugins'",
",",
"[",
"]",
")",
"if",
"not",
"disable_plugins",
":",
"logger",
".",
"debug",
"(",
"\"No site-specific plugins to disable\"",
")",
"else",
":",
"for",
"plugin_dict",
"in",
"disable_plugins",
":",
"try",
":",
"self",
".",
"dj",
".",
"remove_plugin",
"(",
"plugin_dict",
"[",
"'plugin_type'",
"]",
",",
"plugin_dict",
"[",
"'plugin_name'",
"]",
")",
"logger",
".",
"debug",
"(",
"\"site-specific plugin disabled -> Type:{} Name:{}\"",
".",
"format",
"(",
"plugin_dict",
"[",
"'plugin_type'",
"]",
",",
"plugin_dict",
"[",
"'plugin_name'",
"]",
")",
")",
"except",
"KeyError",
":",
"# Malformed config",
"logger",
".",
"debug",
"(",
"\"Invalid custom configuration found for disable_plugins\"",
")",
"enable_plugins",
"=",
"self",
".",
"customize_conf",
".",
"get",
"(",
"'enable_plugins'",
",",
"[",
"]",
")",
"if",
"not",
"enable_plugins",
":",
"logger",
".",
"debug",
"(",
"\"No site-specific plugins to enable\"",
")",
"else",
":",
"for",
"plugin_dict",
"in",
"enable_plugins",
":",
"try",
":",
"self",
".",
"dj",
".",
"add_plugin",
"(",
"plugin_dict",
"[",
"'plugin_type'",
"]",
",",
"plugin_dict",
"[",
"'plugin_name'",
"]",
",",
"plugin_dict",
"[",
"'plugin_args'",
"]",
")",
"logger",
".",
"debug",
"(",
"\"site-specific plugin enabled -> Type:{} Name:{} Args: {}\"",
".",
"format",
"(",
"plugin_dict",
"[",
"'plugin_type'",
"]",
",",
"plugin_dict",
"[",
"'plugin_name'",
"]",
",",
"plugin_dict",
"[",
"'plugin_args'",
"]",
")",
")",
"except",
"KeyError",
":",
"# Malformed config",
"logger",
".",
"debug",
"(",
"\"Invalid custom configuration found for enable_plugins\"",
")"
] |
Customize prod_inner for site specific customizations
|
[
"Customize",
"prod_inner",
"for",
"site",
"specific",
"customizations"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1375-L1420
|
train
|
projectatomic/osbs-client
|
osbs/build/build_request.py
|
BuildRequest.render_name
|
def render_name(self, name, image_tag, platform):
"""Sets the Build/BuildConfig object name"""
if self.scratch or self.isolated:
name = image_tag
# Platform name may contain characters not allowed by OpenShift.
if platform:
platform_suffix = '-{}'.format(platform)
if name.endswith(platform_suffix):
name = name[:-len(platform_suffix)]
_, salt, timestamp = name.rsplit('-', 2)
if self.scratch:
name = 'scratch-{}-{}'.format(salt, timestamp)
elif self.isolated:
name = 'isolated-{}-{}'.format(salt, timestamp)
# !IMPORTANT! can't be too long: https://github.com/openshift/origin/issues/733
self.template['metadata']['name'] = name
|
python
|
def render_name(self, name, image_tag, platform):
"""Sets the Build/BuildConfig object name"""
if self.scratch or self.isolated:
name = image_tag
# Platform name may contain characters not allowed by OpenShift.
if platform:
platform_suffix = '-{}'.format(platform)
if name.endswith(platform_suffix):
name = name[:-len(platform_suffix)]
_, salt, timestamp = name.rsplit('-', 2)
if self.scratch:
name = 'scratch-{}-{}'.format(salt, timestamp)
elif self.isolated:
name = 'isolated-{}-{}'.format(salt, timestamp)
# !IMPORTANT! can't be too long: https://github.com/openshift/origin/issues/733
self.template['metadata']['name'] = name
|
[
"def",
"render_name",
"(",
"self",
",",
"name",
",",
"image_tag",
",",
"platform",
")",
":",
"if",
"self",
".",
"scratch",
"or",
"self",
".",
"isolated",
":",
"name",
"=",
"image_tag",
"# Platform name may contain characters not allowed by OpenShift.",
"if",
"platform",
":",
"platform_suffix",
"=",
"'-{}'",
".",
"format",
"(",
"platform",
")",
"if",
"name",
".",
"endswith",
"(",
"platform_suffix",
")",
":",
"name",
"=",
"name",
"[",
":",
"-",
"len",
"(",
"platform_suffix",
")",
"]",
"_",
",",
"salt",
",",
"timestamp",
"=",
"name",
".",
"rsplit",
"(",
"'-'",
",",
"2",
")",
"if",
"self",
".",
"scratch",
":",
"name",
"=",
"'scratch-{}-{}'",
".",
"format",
"(",
"salt",
",",
"timestamp",
")",
"elif",
"self",
".",
"isolated",
":",
"name",
"=",
"'isolated-{}-{}'",
".",
"format",
"(",
"salt",
",",
"timestamp",
")",
"# !IMPORTANT! can't be too long: https://github.com/openshift/origin/issues/733",
"self",
".",
"template",
"[",
"'metadata'",
"]",
"[",
"'name'",
"]",
"=",
"name"
] |
Sets the Build/BuildConfig object name
|
[
"Sets",
"the",
"Build",
"/",
"BuildConfig",
"object",
"name"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_request.py#L1425-L1444
|
train
|
projectatomic/osbs-client
|
osbs/cli/capture.py
|
setup_json_capture
|
def setup_json_capture(osbs, os_conf, capture_dir):
"""
Only used for setting up the testing framework.
"""
try:
os.mkdir(capture_dir)
except OSError:
pass
finally:
osbs.os._con.request = ResponseSaver(capture_dir,
os_conf.get_openshift_api_uri(),
os_conf.get_k8s_api_uri(),
osbs.os._con.request).request
|
python
|
def setup_json_capture(osbs, os_conf, capture_dir):
"""
Only used for setting up the testing framework.
"""
try:
os.mkdir(capture_dir)
except OSError:
pass
finally:
osbs.os._con.request = ResponseSaver(capture_dir,
os_conf.get_openshift_api_uri(),
os_conf.get_k8s_api_uri(),
osbs.os._con.request).request
|
[
"def",
"setup_json_capture",
"(",
"osbs",
",",
"os_conf",
",",
"capture_dir",
")",
":",
"try",
":",
"os",
".",
"mkdir",
"(",
"capture_dir",
")",
"except",
"OSError",
":",
"pass",
"finally",
":",
"osbs",
".",
"os",
".",
"_con",
".",
"request",
"=",
"ResponseSaver",
"(",
"capture_dir",
",",
"os_conf",
".",
"get_openshift_api_uri",
"(",
")",
",",
"os_conf",
".",
"get_k8s_api_uri",
"(",
")",
",",
"osbs",
".",
"os",
".",
"_con",
".",
"request",
")",
".",
"request"
] |
Only used for setting up the testing framework.
|
[
"Only",
"used",
"for",
"setting",
"up",
"the",
"testing",
"framework",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/capture.py#L96-L109
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
get_terminal_size
|
def get_terminal_size():
"""
get size of console: rows x columns
:return: tuple, (int, int)
"""
try:
rows, columns = subprocess.check_output(['stty', 'size']).split()
except subprocess.CalledProcessError:
# not attached to terminal
logger.info("not attached to terminal")
return 0, 0
logger.debug("console size is %s %s", rows, columns)
return int(rows), int(columns)
|
python
|
def get_terminal_size():
"""
get size of console: rows x columns
:return: tuple, (int, int)
"""
try:
rows, columns = subprocess.check_output(['stty', 'size']).split()
except subprocess.CalledProcessError:
# not attached to terminal
logger.info("not attached to terminal")
return 0, 0
logger.debug("console size is %s %s", rows, columns)
return int(rows), int(columns)
|
[
"def",
"get_terminal_size",
"(",
")",
":",
"try",
":",
"rows",
",",
"columns",
"=",
"subprocess",
".",
"check_output",
"(",
"[",
"'stty'",
",",
"'size'",
"]",
")",
".",
"split",
"(",
")",
"except",
"subprocess",
".",
"CalledProcessError",
":",
"# not attached to terminal",
"logger",
".",
"info",
"(",
"\"not attached to terminal\"",
")",
"return",
"0",
",",
"0",
"logger",
".",
"debug",
"(",
"\"console size is %s %s\"",
",",
"rows",
",",
"columns",
")",
"return",
"int",
"(",
"rows",
")",
",",
"int",
"(",
"columns",
")"
] |
get size of console: rows x columns
:return: tuple, (int, int)
|
[
"get",
"size",
"of",
"console",
":",
"rows",
"x",
"columns"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L18-L31
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
TableFormatter._longest_val_in_column
|
def _longest_val_in_column(self, col):
"""
get size of longest value in specific column
:param col: str, column name
:return int
"""
try:
# +2 is for implicit separator
return max([len(x[col]) for x in self.table if x[col]]) + 2
except KeyError:
logger.error("there is no column %r", col)
raise
|
python
|
def _longest_val_in_column(self, col):
"""
get size of longest value in specific column
:param col: str, column name
:return int
"""
try:
# +2 is for implicit separator
return max([len(x[col]) for x in self.table if x[col]]) + 2
except KeyError:
logger.error("there is no column %r", col)
raise
|
[
"def",
"_longest_val_in_column",
"(",
"self",
",",
"col",
")",
":",
"try",
":",
"# +2 is for implicit separator",
"return",
"max",
"(",
"[",
"len",
"(",
"x",
"[",
"col",
"]",
")",
"for",
"x",
"in",
"self",
".",
"table",
"if",
"x",
"[",
"col",
"]",
"]",
")",
"+",
"2",
"except",
"KeyError",
":",
"logger",
".",
"error",
"(",
"\"there is no column %r\"",
",",
"col",
")",
"raise"
] |
get size of longest value in specific column
:param col: str, column name
:return int
|
[
"get",
"size",
"of",
"longest",
"value",
"in",
"specific",
"column"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L59-L71
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
TablePrinter._init
|
def _init(self):
"""
initialize all values based on provided input
:return: None
"""
self.col_count = len(self.col_list)
# list of lengths of longest entries in columns
self.col_longest = self.get_all_longest_col_lengths()
self.data_length = sum(self.col_longest.values())
if self.terminal_width > 0:
# free space is space which should be equeally distributed for all columns
# self.terminal_width -- terminal is our canvas
# - self.data_length -- substract length of content (the actual data)
# - self.col_count + 1 -- table lines are not part of free space, their width is
# (number of columns - 1)
self.total_free_space = (self.terminal_width - self.data_length) - self.col_count + 1
if self.total_free_space <= 0:
self.total_free_space = None
else:
self.default_column_space = self.total_free_space // self.col_count
self.default_column_space_remainder = self.total_free_space % self.col_count
logger.debug("total free space: %d, column space: %d, remainder: %d, columns: %d",
self.total_free_space, self.default_column_space,
self.default_column_space_remainder, self.col_count)
else:
self.total_free_space = None
|
python
|
def _init(self):
"""
initialize all values based on provided input
:return: None
"""
self.col_count = len(self.col_list)
# list of lengths of longest entries in columns
self.col_longest = self.get_all_longest_col_lengths()
self.data_length = sum(self.col_longest.values())
if self.terminal_width > 0:
# free space is space which should be equeally distributed for all columns
# self.terminal_width -- terminal is our canvas
# - self.data_length -- substract length of content (the actual data)
# - self.col_count + 1 -- table lines are not part of free space, their width is
# (number of columns - 1)
self.total_free_space = (self.terminal_width - self.data_length) - self.col_count + 1
if self.total_free_space <= 0:
self.total_free_space = None
else:
self.default_column_space = self.total_free_space // self.col_count
self.default_column_space_remainder = self.total_free_space % self.col_count
logger.debug("total free space: %d, column space: %d, remainder: %d, columns: %d",
self.total_free_space, self.default_column_space,
self.default_column_space_remainder, self.col_count)
else:
self.total_free_space = None
|
[
"def",
"_init",
"(",
"self",
")",
":",
"self",
".",
"col_count",
"=",
"len",
"(",
"self",
".",
"col_list",
")",
"# list of lengths of longest entries in columns",
"self",
".",
"col_longest",
"=",
"self",
".",
"get_all_longest_col_lengths",
"(",
")",
"self",
".",
"data_length",
"=",
"sum",
"(",
"self",
".",
"col_longest",
".",
"values",
"(",
")",
")",
"if",
"self",
".",
"terminal_width",
">",
"0",
":",
"# free space is space which should be equeally distributed for all columns",
"# self.terminal_width -- terminal is our canvas",
"# - self.data_length -- substract length of content (the actual data)",
"# - self.col_count + 1 -- table lines are not part of free space, their width is",
"# (number of columns - 1)",
"self",
".",
"total_free_space",
"=",
"(",
"self",
".",
"terminal_width",
"-",
"self",
".",
"data_length",
")",
"-",
"self",
".",
"col_count",
"+",
"1",
"if",
"self",
".",
"total_free_space",
"<=",
"0",
":",
"self",
".",
"total_free_space",
"=",
"None",
"else",
":",
"self",
".",
"default_column_space",
"=",
"self",
".",
"total_free_space",
"//",
"self",
".",
"col_count",
"self",
".",
"default_column_space_remainder",
"=",
"self",
".",
"total_free_space",
"%",
"self",
".",
"col_count",
"logger",
".",
"debug",
"(",
"\"total free space: %d, column space: %d, remainder: %d, columns: %d\"",
",",
"self",
".",
"total_free_space",
",",
"self",
".",
"default_column_space",
",",
"self",
".",
"default_column_space_remainder",
",",
"self",
".",
"col_count",
")",
"else",
":",
"self",
".",
"total_free_space",
"=",
"None"
] |
initialize all values based on provided input
:return: None
|
[
"initialize",
"all",
"values",
"based",
"on",
"provided",
"input"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L95-L122
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
TablePrinter._count_sizes
|
def _count_sizes(self):
"""
count all values needed to display whole table
<><---terminal-width-----------><>
<> HEADER | HEADER2 | HEADER3 <>
<>---------+----------+---------<>
kudos to PostgreSQL developers
:return: None
"""
format_list = []
header_sepa_format_list = []
# actual widths of columns
self.col_widths = {}
for col in self.col_list:
col_length = self.col_longest[col]
col_width = col_length + self._separate()
# -2 is for implicit separator -- spaces around
format_list.append(" {%s:%d} " % (col, col_width - 2))
header_sepa_format_list.append("{%s:%d}" % (col, col_width))
self.col_widths[col] = col_width
logger.debug("column widths %s", self.col_widths)
self.format_str = "|".join(format_list)
self.header_format_str = "+".join(header_sepa_format_list)
self.header_data = {}
for k in self.col_widths:
self.header_data[k] = "-" * self.col_widths[k]
|
python
|
def _count_sizes(self):
"""
count all values needed to display whole table
<><---terminal-width-----------><>
<> HEADER | HEADER2 | HEADER3 <>
<>---------+----------+---------<>
kudos to PostgreSQL developers
:return: None
"""
format_list = []
header_sepa_format_list = []
# actual widths of columns
self.col_widths = {}
for col in self.col_list:
col_length = self.col_longest[col]
col_width = col_length + self._separate()
# -2 is for implicit separator -- spaces around
format_list.append(" {%s:%d} " % (col, col_width - 2))
header_sepa_format_list.append("{%s:%d}" % (col, col_width))
self.col_widths[col] = col_width
logger.debug("column widths %s", self.col_widths)
self.format_str = "|".join(format_list)
self.header_format_str = "+".join(header_sepa_format_list)
self.header_data = {}
for k in self.col_widths:
self.header_data[k] = "-" * self.col_widths[k]
|
[
"def",
"_count_sizes",
"(",
"self",
")",
":",
"format_list",
"=",
"[",
"]",
"header_sepa_format_list",
"=",
"[",
"]",
"# actual widths of columns",
"self",
".",
"col_widths",
"=",
"{",
"}",
"for",
"col",
"in",
"self",
".",
"col_list",
":",
"col_length",
"=",
"self",
".",
"col_longest",
"[",
"col",
"]",
"col_width",
"=",
"col_length",
"+",
"self",
".",
"_separate",
"(",
")",
"# -2 is for implicit separator -- spaces around",
"format_list",
".",
"append",
"(",
"\" {%s:%d} \"",
"%",
"(",
"col",
",",
"col_width",
"-",
"2",
")",
")",
"header_sepa_format_list",
".",
"append",
"(",
"\"{%s:%d}\"",
"%",
"(",
"col",
",",
"col_width",
")",
")",
"self",
".",
"col_widths",
"[",
"col",
"]",
"=",
"col_width",
"logger",
".",
"debug",
"(",
"\"column widths %s\"",
",",
"self",
".",
"col_widths",
")",
"self",
".",
"format_str",
"=",
"\"|\"",
".",
"join",
"(",
"format_list",
")",
"self",
".",
"header_format_str",
"=",
"\"+\"",
".",
"join",
"(",
"header_sepa_format_list",
")",
"self",
".",
"header_data",
"=",
"{",
"}",
"for",
"k",
"in",
"self",
".",
"col_widths",
":",
"self",
".",
"header_data",
"[",
"k",
"]",
"=",
"\"-\"",
"*",
"self",
".",
"col_widths",
"[",
"k",
"]"
] |
count all values needed to display whole table
<><---terminal-width-----------><>
<> HEADER | HEADER2 | HEADER3 <>
<>---------+----------+---------<>
kudos to PostgreSQL developers
:return: None
|
[
"count",
"all",
"values",
"needed",
"to",
"display",
"whole",
"table"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L124-L157
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
TablePrinter.get_all_longest_col_lengths
|
def get_all_longest_col_lengths(self):
"""
iterate over all columns and get their longest values
:return: dict, {"column_name": 132}
"""
response = {}
for col in self.col_list:
response[col] = self._longest_val_in_column(col)
return response
|
python
|
def get_all_longest_col_lengths(self):
"""
iterate over all columns and get their longest values
:return: dict, {"column_name": 132}
"""
response = {}
for col in self.col_list:
response[col] = self._longest_val_in_column(col)
return response
|
[
"def",
"get_all_longest_col_lengths",
"(",
"self",
")",
":",
"response",
"=",
"{",
"}",
"for",
"col",
"in",
"self",
".",
"col_list",
":",
"response",
"[",
"col",
"]",
"=",
"self",
".",
"_longest_val_in_column",
"(",
"col",
")",
"return",
"response"
] |
iterate over all columns and get their longest values
:return: dict, {"column_name": 132}
|
[
"iterate",
"over",
"all",
"columns",
"and",
"get",
"their",
"longest",
"values"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L159-L168
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
TablePrinter._separate
|
def _separate(self):
"""
get a width of separator for current column
:return: int
"""
if self.total_free_space is None:
return 0
else:
sepa = self.default_column_space
# we need to distribute remainders
if self.default_column_space_remainder > 0:
sepa += 1
self.default_column_space_remainder -= 1
logger.debug("remainder: %d, separator: %d",
self.default_column_space_remainder, sepa)
return sepa
|
python
|
def _separate(self):
"""
get a width of separator for current column
:return: int
"""
if self.total_free_space is None:
return 0
else:
sepa = self.default_column_space
# we need to distribute remainders
if self.default_column_space_remainder > 0:
sepa += 1
self.default_column_space_remainder -= 1
logger.debug("remainder: %d, separator: %d",
self.default_column_space_remainder, sepa)
return sepa
|
[
"def",
"_separate",
"(",
"self",
")",
":",
"if",
"self",
".",
"total_free_space",
"is",
"None",
":",
"return",
"0",
"else",
":",
"sepa",
"=",
"self",
".",
"default_column_space",
"# we need to distribute remainders",
"if",
"self",
".",
"default_column_space_remainder",
">",
"0",
":",
"sepa",
"+=",
"1",
"self",
".",
"default_column_space_remainder",
"-=",
"1",
"logger",
".",
"debug",
"(",
"\"remainder: %d, separator: %d\"",
",",
"self",
".",
"default_column_space_remainder",
",",
"sepa",
")",
"return",
"sepa"
] |
get a width of separator for current column
:return: int
|
[
"get",
"a",
"width",
"of",
"separator",
"for",
"current",
"column"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L170-L186
|
train
|
projectatomic/osbs-client
|
osbs/cli/render.py
|
TablePrinter.render
|
def render(self):
"""
print provided table
:return: None
"""
print(self.format_str.format(**self.header), file=sys.stderr)
print(self.header_format_str.format(**self.header_data), file=sys.stderr)
for row in self.data:
print(self.format_str.format(**row))
|
python
|
def render(self):
"""
print provided table
:return: None
"""
print(self.format_str.format(**self.header), file=sys.stderr)
print(self.header_format_str.format(**self.header_data), file=sys.stderr)
for row in self.data:
print(self.format_str.format(**row))
|
[
"def",
"render",
"(",
"self",
")",
":",
"print",
"(",
"self",
".",
"format_str",
".",
"format",
"(",
"*",
"*",
"self",
".",
"header",
")",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"print",
"(",
"self",
".",
"header_format_str",
".",
"format",
"(",
"*",
"*",
"self",
".",
"header_data",
")",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"for",
"row",
"in",
"self",
".",
"data",
":",
"print",
"(",
"self",
".",
"format_str",
".",
"format",
"(",
"*",
"*",
"row",
")",
")"
] |
print provided table
:return: None
|
[
"print",
"provided",
"table"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/render.py#L188-L197
|
train
|
thombashi/pytablereader
|
pytablereader/json/formatter.py
|
JsonConverter._validate_source_data
|
def _validate_source_data(self):
"""
:raises ValidationError:
"""
try:
jsonschema.validate(self._buffer, self._schema)
except jsonschema.ValidationError as e:
raise ValidationError(e)
|
python
|
def _validate_source_data(self):
"""
:raises ValidationError:
"""
try:
jsonschema.validate(self._buffer, self._schema)
except jsonschema.ValidationError as e:
raise ValidationError(e)
|
[
"def",
"_validate_source_data",
"(",
"self",
")",
":",
"try",
":",
"jsonschema",
".",
"validate",
"(",
"self",
".",
"_buffer",
",",
"self",
".",
"_schema",
")",
"except",
"jsonschema",
".",
"ValidationError",
"as",
"e",
":",
"raise",
"ValidationError",
"(",
"e",
")"
] |
:raises ValidationError:
|
[
":",
"raises",
"ValidationError",
":"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/json/formatter.py#L38-L46
|
train
|
thombashi/pytablereader
|
pytablereader/json/formatter.py
|
SingleJsonTableConverterC.to_table_data
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
self._loader.inc_table_count()
yield TableData(
self._make_table_name(),
["key", "value"],
[record for record in self._buffer.items()],
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(),
)
|
python
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
self._loader.inc_table_count()
yield TableData(
self._make_table_name(),
["key", "value"],
[record for record in self._buffer.items()],
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(),
)
|
[
"def",
"to_table_data",
"(",
"self",
")",
":",
"self",
".",
"_validate_source_data",
"(",
")",
"self",
".",
"_loader",
".",
"inc_table_count",
"(",
")",
"yield",
"TableData",
"(",
"self",
".",
"_make_table_name",
"(",
")",
",",
"[",
"\"key\"",
",",
"\"value\"",
"]",
",",
"[",
"record",
"for",
"record",
"in",
"self",
".",
"_buffer",
".",
"items",
"(",
")",
"]",
",",
"dp_extractor",
"=",
"self",
".",
"_loader",
".",
"dp_extractor",
",",
"type_hints",
"=",
"self",
".",
"_extract_type_hints",
"(",
")",
",",
")"
] |
:raises ValueError:
:raises pytablereader.error.ValidationError:
|
[
":",
"raises",
"ValueError",
":",
":",
"raises",
"pytablereader",
".",
"error",
".",
"ValidationError",
":"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/json/formatter.py#L139-L154
|
train
|
thombashi/pytablereader
|
pytablereader/json/formatter.py
|
MultipleJsonTableConverterA.to_table_data
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
for table_key, json_records in six.iteritems(self._buffer):
attr_name_set = set()
for json_record in json_records:
attr_name_set = attr_name_set.union(six.viewkeys(json_record))
headers = sorted(attr_name_set)
self._loader.inc_table_count()
self._table_key = table_key
yield TableData(
self._make_table_name(),
headers,
json_records,
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(headers),
)
|
python
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
for table_key, json_records in six.iteritems(self._buffer):
attr_name_set = set()
for json_record in json_records:
attr_name_set = attr_name_set.union(six.viewkeys(json_record))
headers = sorted(attr_name_set)
self._loader.inc_table_count()
self._table_key = table_key
yield TableData(
self._make_table_name(),
headers,
json_records,
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(headers),
)
|
[
"def",
"to_table_data",
"(",
"self",
")",
":",
"self",
".",
"_validate_source_data",
"(",
")",
"for",
"table_key",
",",
"json_records",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_buffer",
")",
":",
"attr_name_set",
"=",
"set",
"(",
")",
"for",
"json_record",
"in",
"json_records",
":",
"attr_name_set",
"=",
"attr_name_set",
".",
"union",
"(",
"six",
".",
"viewkeys",
"(",
"json_record",
")",
")",
"headers",
"=",
"sorted",
"(",
"attr_name_set",
")",
"self",
".",
"_loader",
".",
"inc_table_count",
"(",
")",
"self",
".",
"_table_key",
"=",
"table_key",
"yield",
"TableData",
"(",
"self",
".",
"_make_table_name",
"(",
")",
",",
"headers",
",",
"json_records",
",",
"dp_extractor",
"=",
"self",
".",
"_loader",
".",
"dp_extractor",
",",
"type_hints",
"=",
"self",
".",
"_extract_type_hints",
"(",
"headers",
")",
",",
")"
] |
:raises ValueError:
:raises pytablereader.error.ValidationError:
|
[
":",
"raises",
"ValueError",
":",
":",
"raises",
"pytablereader",
".",
"error",
".",
"ValidationError",
":"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/json/formatter.py#L186-L209
|
train
|
thombashi/pytablereader
|
pytablereader/json/formatter.py
|
MultipleJsonTableConverterB.to_table_data
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
for table_key, json_records in six.iteritems(self._buffer):
headers = sorted(six.viewkeys(json_records))
self._loader.inc_table_count()
self._table_key = table_key
yield TableData(
self._make_table_name(),
headers,
zip(*[json_records.get(header) for header in headers]),
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(headers),
)
|
python
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
for table_key, json_records in six.iteritems(self._buffer):
headers = sorted(six.viewkeys(json_records))
self._loader.inc_table_count()
self._table_key = table_key
yield TableData(
self._make_table_name(),
headers,
zip(*[json_records.get(header) for header in headers]),
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(headers),
)
|
[
"def",
"to_table_data",
"(",
"self",
")",
":",
"self",
".",
"_validate_source_data",
"(",
")",
"for",
"table_key",
",",
"json_records",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_buffer",
")",
":",
"headers",
"=",
"sorted",
"(",
"six",
".",
"viewkeys",
"(",
"json_records",
")",
")",
"self",
".",
"_loader",
".",
"inc_table_count",
"(",
")",
"self",
".",
"_table_key",
"=",
"table_key",
"yield",
"TableData",
"(",
"self",
".",
"_make_table_name",
"(",
")",
",",
"headers",
",",
"zip",
"(",
"*",
"[",
"json_records",
".",
"get",
"(",
"header",
")",
"for",
"header",
"in",
"headers",
"]",
")",
",",
"dp_extractor",
"=",
"self",
".",
"_loader",
".",
"dp_extractor",
",",
"type_hints",
"=",
"self",
".",
"_extract_type_hints",
"(",
"headers",
")",
",",
")"
] |
:raises ValueError:
:raises pytablereader.error.ValidationError:
|
[
":",
"raises",
"ValueError",
":",
":",
"raises",
"pytablereader",
".",
"error",
".",
"ValidationError",
":"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/json/formatter.py#L227-L247
|
train
|
thombashi/pytablereader
|
pytablereader/json/formatter.py
|
MultipleJsonTableConverterC.to_table_data
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
for table_key, json_records in six.iteritems(self._buffer):
self._loader.inc_table_count()
self._table_key = table_key
yield TableData(
self._make_table_name(),
["key", "value"],
[record for record in json_records.items()],
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(),
)
|
python
|
def to_table_data(self):
"""
:raises ValueError:
:raises pytablereader.error.ValidationError:
"""
self._validate_source_data()
for table_key, json_records in six.iteritems(self._buffer):
self._loader.inc_table_count()
self._table_key = table_key
yield TableData(
self._make_table_name(),
["key", "value"],
[record for record in json_records.items()],
dp_extractor=self._loader.dp_extractor,
type_hints=self._extract_type_hints(),
)
|
[
"def",
"to_table_data",
"(",
"self",
")",
":",
"self",
".",
"_validate_source_data",
"(",
")",
"for",
"table_key",
",",
"json_records",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_buffer",
")",
":",
"self",
".",
"_loader",
".",
"inc_table_count",
"(",
")",
"self",
".",
"_table_key",
"=",
"table_key",
"yield",
"TableData",
"(",
"self",
".",
"_make_table_name",
"(",
")",
",",
"[",
"\"key\"",
",",
"\"value\"",
"]",
",",
"[",
"record",
"for",
"record",
"in",
"json_records",
".",
"items",
"(",
")",
"]",
",",
"dp_extractor",
"=",
"self",
".",
"_loader",
".",
"dp_extractor",
",",
"type_hints",
"=",
"self",
".",
"_extract_type_hints",
"(",
")",
",",
")"
] |
:raises ValueError:
:raises pytablereader.error.ValidationError:
|
[
":",
"raises",
"ValueError",
":",
":",
"raises",
"pytablereader",
".",
"error",
".",
"ValidationError",
":"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/json/formatter.py#L265-L283
|
train
|
projectatomic/osbs-client
|
osbs/build/build_requestv2.py
|
BuildRequestV2.set_params
|
def set_params(self, **kwargs):
"""
set parameters in the user parameters
these parameters are accepted:
:param git_uri: str, uri of the git repository for the source
:param git_ref: str, commit ID of the branch to be pulled
:param git_branch: str, branch name of the branch to be pulled
:param base_image: str, name of the parent image
:param name_label: str, label of the parent image
:param user: str, name of the user requesting the build
:param component: str, name of the component
:param release: str,
:param build_image: str,
:param build_imagestream: str,
:param build_from: str,
:param build_type: str, orchestrator or worker
:param platforms: list of str, platforms to build on
:param platform: str, platform
:param koji_target: str, koji tag with packages used to build the image
:param koji_task_id: str, koji ID
:param koji_parent_build: str,
:param koji_upload_dir: str, koji directory where the completed image will be uploaded
:param flatpak: if we should build a Flatpak OCI Image
:param flatpak_base_image: str, name of the Flatpack OCI Image
:param reactor_config_map: str, name of the config map containing the reactor environment
:param reactor_config_override: dict, data structure for reactor config to be injected as
an environment variable into a worker build;
when used, reactor_config_map is ignored.
:param yum_repourls: list of str, uris of the yum repos to pull from
:param signing_intent: bool, True to sign the resulting image
:param compose_ids: list of int, ODCS composes to use instead of generating new ones
:param filesystem_koji_task_id: int, Koji Task that created the base filesystem
:param platform_node_selector: dict, a nodeselector for a user_paramsific platform
:param scratch_build_node_selector: dict, a nodeselector for scratch builds
:param explicit_build_node_selector: dict, a nodeselector for explicit builds
:param auto_build_node_selector: dict, a nodeselector for auto builds
:param isolated_build_node_selector: dict, a nodeselector for isolated builds
:param operator_manifests_extract_platform: str, indicates which platform should upload
operator manifests to koji
:param parent_images_digests: dict, mapping image digests to names and platforms
"""
# Here we cater to the koji "scratch" build type, this will disable
# all plugins that might cause importing of data to koji
self.scratch = kwargs.get('scratch')
# When true, it indicates build was automatically started by
# OpenShift via a trigger, for instance ImageChangeTrigger
self.is_auto = kwargs.pop('is_auto', False)
# An isolated build is meant to patch a certain release and not
# update transient tags in container registry
self.isolated = kwargs.get('isolated')
self.osbs_api = kwargs.pop('osbs_api', None)
self.validate_build_variation()
self.base_image = kwargs.get('base_image')
self.platform_node_selector = kwargs.get('platform_node_selector', {})
self.scratch_build_node_selector = kwargs.get('scratch_build_node_selector', {})
self.explicit_build_node_selector = kwargs.get('explicit_build_node_selector', {})
self.auto_build_node_selector = kwargs.get('auto_build_node_selector', {})
self.isolated_build_node_selector = kwargs.get('isolated_build_node_selector', {})
logger.debug("now setting params '%s' for user_params", kwargs)
self.user_params.set_params(**kwargs)
self.source_registry = None
self.organization = None
|
python
|
def set_params(self, **kwargs):
"""
set parameters in the user parameters
these parameters are accepted:
:param git_uri: str, uri of the git repository for the source
:param git_ref: str, commit ID of the branch to be pulled
:param git_branch: str, branch name of the branch to be pulled
:param base_image: str, name of the parent image
:param name_label: str, label of the parent image
:param user: str, name of the user requesting the build
:param component: str, name of the component
:param release: str,
:param build_image: str,
:param build_imagestream: str,
:param build_from: str,
:param build_type: str, orchestrator or worker
:param platforms: list of str, platforms to build on
:param platform: str, platform
:param koji_target: str, koji tag with packages used to build the image
:param koji_task_id: str, koji ID
:param koji_parent_build: str,
:param koji_upload_dir: str, koji directory where the completed image will be uploaded
:param flatpak: if we should build a Flatpak OCI Image
:param flatpak_base_image: str, name of the Flatpack OCI Image
:param reactor_config_map: str, name of the config map containing the reactor environment
:param reactor_config_override: dict, data structure for reactor config to be injected as
an environment variable into a worker build;
when used, reactor_config_map is ignored.
:param yum_repourls: list of str, uris of the yum repos to pull from
:param signing_intent: bool, True to sign the resulting image
:param compose_ids: list of int, ODCS composes to use instead of generating new ones
:param filesystem_koji_task_id: int, Koji Task that created the base filesystem
:param platform_node_selector: dict, a nodeselector for a user_paramsific platform
:param scratch_build_node_selector: dict, a nodeselector for scratch builds
:param explicit_build_node_selector: dict, a nodeselector for explicit builds
:param auto_build_node_selector: dict, a nodeselector for auto builds
:param isolated_build_node_selector: dict, a nodeselector for isolated builds
:param operator_manifests_extract_platform: str, indicates which platform should upload
operator manifests to koji
:param parent_images_digests: dict, mapping image digests to names and platforms
"""
# Here we cater to the koji "scratch" build type, this will disable
# all plugins that might cause importing of data to koji
self.scratch = kwargs.get('scratch')
# When true, it indicates build was automatically started by
# OpenShift via a trigger, for instance ImageChangeTrigger
self.is_auto = kwargs.pop('is_auto', False)
# An isolated build is meant to patch a certain release and not
# update transient tags in container registry
self.isolated = kwargs.get('isolated')
self.osbs_api = kwargs.pop('osbs_api', None)
self.validate_build_variation()
self.base_image = kwargs.get('base_image')
self.platform_node_selector = kwargs.get('platform_node_selector', {})
self.scratch_build_node_selector = kwargs.get('scratch_build_node_selector', {})
self.explicit_build_node_selector = kwargs.get('explicit_build_node_selector', {})
self.auto_build_node_selector = kwargs.get('auto_build_node_selector', {})
self.isolated_build_node_selector = kwargs.get('isolated_build_node_selector', {})
logger.debug("now setting params '%s' for user_params", kwargs)
self.user_params.set_params(**kwargs)
self.source_registry = None
self.organization = None
|
[
"def",
"set_params",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"# Here we cater to the koji \"scratch\" build type, this will disable",
"# all plugins that might cause importing of data to koji",
"self",
".",
"scratch",
"=",
"kwargs",
".",
"get",
"(",
"'scratch'",
")",
"# When true, it indicates build was automatically started by",
"# OpenShift via a trigger, for instance ImageChangeTrigger",
"self",
".",
"is_auto",
"=",
"kwargs",
".",
"pop",
"(",
"'is_auto'",
",",
"False",
")",
"# An isolated build is meant to patch a certain release and not",
"# update transient tags in container registry",
"self",
".",
"isolated",
"=",
"kwargs",
".",
"get",
"(",
"'isolated'",
")",
"self",
".",
"osbs_api",
"=",
"kwargs",
".",
"pop",
"(",
"'osbs_api'",
",",
"None",
")",
"self",
".",
"validate_build_variation",
"(",
")",
"self",
".",
"base_image",
"=",
"kwargs",
".",
"get",
"(",
"'base_image'",
")",
"self",
".",
"platform_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'platform_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"scratch_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'scratch_build_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"explicit_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'explicit_build_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"auto_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'auto_build_node_selector'",
",",
"{",
"}",
")",
"self",
".",
"isolated_build_node_selector",
"=",
"kwargs",
".",
"get",
"(",
"'isolated_build_node_selector'",
",",
"{",
"}",
")",
"logger",
".",
"debug",
"(",
"\"now setting params '%s' for user_params\"",
",",
"kwargs",
")",
"self",
".",
"user_params",
".",
"set_params",
"(",
"*",
"*",
"kwargs",
")",
"self",
".",
"source_registry",
"=",
"None",
"self",
".",
"organization",
"=",
"None"
] |
set parameters in the user parameters
these parameters are accepted:
:param git_uri: str, uri of the git repository for the source
:param git_ref: str, commit ID of the branch to be pulled
:param git_branch: str, branch name of the branch to be pulled
:param base_image: str, name of the parent image
:param name_label: str, label of the parent image
:param user: str, name of the user requesting the build
:param component: str, name of the component
:param release: str,
:param build_image: str,
:param build_imagestream: str,
:param build_from: str,
:param build_type: str, orchestrator or worker
:param platforms: list of str, platforms to build on
:param platform: str, platform
:param koji_target: str, koji tag with packages used to build the image
:param koji_task_id: str, koji ID
:param koji_parent_build: str,
:param koji_upload_dir: str, koji directory where the completed image will be uploaded
:param flatpak: if we should build a Flatpak OCI Image
:param flatpak_base_image: str, name of the Flatpack OCI Image
:param reactor_config_map: str, name of the config map containing the reactor environment
:param reactor_config_override: dict, data structure for reactor config to be injected as
an environment variable into a worker build;
when used, reactor_config_map is ignored.
:param yum_repourls: list of str, uris of the yum repos to pull from
:param signing_intent: bool, True to sign the resulting image
:param compose_ids: list of int, ODCS composes to use instead of generating new ones
:param filesystem_koji_task_id: int, Koji Task that created the base filesystem
:param platform_node_selector: dict, a nodeselector for a user_paramsific platform
:param scratch_build_node_selector: dict, a nodeselector for scratch builds
:param explicit_build_node_selector: dict, a nodeselector for explicit builds
:param auto_build_node_selector: dict, a nodeselector for auto builds
:param isolated_build_node_selector: dict, a nodeselector for isolated builds
:param operator_manifests_extract_platform: str, indicates which platform should upload
operator manifests to koji
:param parent_images_digests: dict, mapping image digests to names and platforms
|
[
"set",
"parameters",
"in",
"the",
"user",
"parameters"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_requestv2.py#L42-L109
|
train
|
projectatomic/osbs-client
|
osbs/build/build_requestv2.py
|
BuildRequestV2.set_data_from_reactor_config
|
def set_data_from_reactor_config(self):
"""
Sets data from reactor config
"""
reactor_config_override = self.user_params.reactor_config_override.value
reactor_config_map = self.user_params.reactor_config_map.value
data = None
if reactor_config_override:
data = reactor_config_override
elif reactor_config_map:
config_map = self.osbs_api.get_config_map(reactor_config_map)
data = config_map.get_data_by_key('config.yaml')
if not data:
if self.user_params.flatpak.value:
raise OsbsValidationException("flatpak_base_image must be provided")
else:
return
source_registry_key = 'source_registry'
registry_organization_key = 'registries_organization'
req_secrets_key = 'required_secrets'
token_secrets_key = 'worker_token_secrets'
flatpak_key = 'flatpak'
flatpak_base_image_key = 'base_image'
if source_registry_key in data:
self.source_registry = data[source_registry_key]
if registry_organization_key in data:
self.organization = data[registry_organization_key]
if self.user_params.flatpak.value:
flatpack_base_image = data.get(flatpak_key, {}).get(flatpak_base_image_key, None)
if flatpack_base_image:
self.base_image = flatpack_base_image
self.user_params.base_image.value = flatpack_base_image
else:
raise OsbsValidationException("flatpak_base_image must be provided")
required_secrets = data.get(req_secrets_key, [])
token_secrets = data.get(token_secrets_key, [])
self._set_required_secrets(required_secrets, token_secrets)
|
python
|
def set_data_from_reactor_config(self):
"""
Sets data from reactor config
"""
reactor_config_override = self.user_params.reactor_config_override.value
reactor_config_map = self.user_params.reactor_config_map.value
data = None
if reactor_config_override:
data = reactor_config_override
elif reactor_config_map:
config_map = self.osbs_api.get_config_map(reactor_config_map)
data = config_map.get_data_by_key('config.yaml')
if not data:
if self.user_params.flatpak.value:
raise OsbsValidationException("flatpak_base_image must be provided")
else:
return
source_registry_key = 'source_registry'
registry_organization_key = 'registries_organization'
req_secrets_key = 'required_secrets'
token_secrets_key = 'worker_token_secrets'
flatpak_key = 'flatpak'
flatpak_base_image_key = 'base_image'
if source_registry_key in data:
self.source_registry = data[source_registry_key]
if registry_organization_key in data:
self.organization = data[registry_organization_key]
if self.user_params.flatpak.value:
flatpack_base_image = data.get(flatpak_key, {}).get(flatpak_base_image_key, None)
if flatpack_base_image:
self.base_image = flatpack_base_image
self.user_params.base_image.value = flatpack_base_image
else:
raise OsbsValidationException("flatpak_base_image must be provided")
required_secrets = data.get(req_secrets_key, [])
token_secrets = data.get(token_secrets_key, [])
self._set_required_secrets(required_secrets, token_secrets)
|
[
"def",
"set_data_from_reactor_config",
"(",
"self",
")",
":",
"reactor_config_override",
"=",
"self",
".",
"user_params",
".",
"reactor_config_override",
".",
"value",
"reactor_config_map",
"=",
"self",
".",
"user_params",
".",
"reactor_config_map",
".",
"value",
"data",
"=",
"None",
"if",
"reactor_config_override",
":",
"data",
"=",
"reactor_config_override",
"elif",
"reactor_config_map",
":",
"config_map",
"=",
"self",
".",
"osbs_api",
".",
"get_config_map",
"(",
"reactor_config_map",
")",
"data",
"=",
"config_map",
".",
"get_data_by_key",
"(",
"'config.yaml'",
")",
"if",
"not",
"data",
":",
"if",
"self",
".",
"user_params",
".",
"flatpak",
".",
"value",
":",
"raise",
"OsbsValidationException",
"(",
"\"flatpak_base_image must be provided\"",
")",
"else",
":",
"return",
"source_registry_key",
"=",
"'source_registry'",
"registry_organization_key",
"=",
"'registries_organization'",
"req_secrets_key",
"=",
"'required_secrets'",
"token_secrets_key",
"=",
"'worker_token_secrets'",
"flatpak_key",
"=",
"'flatpak'",
"flatpak_base_image_key",
"=",
"'base_image'",
"if",
"source_registry_key",
"in",
"data",
":",
"self",
".",
"source_registry",
"=",
"data",
"[",
"source_registry_key",
"]",
"if",
"registry_organization_key",
"in",
"data",
":",
"self",
".",
"organization",
"=",
"data",
"[",
"registry_organization_key",
"]",
"if",
"self",
".",
"user_params",
".",
"flatpak",
".",
"value",
":",
"flatpack_base_image",
"=",
"data",
".",
"get",
"(",
"flatpak_key",
",",
"{",
"}",
")",
".",
"get",
"(",
"flatpak_base_image_key",
",",
"None",
")",
"if",
"flatpack_base_image",
":",
"self",
".",
"base_image",
"=",
"flatpack_base_image",
"self",
".",
"user_params",
".",
"base_image",
".",
"value",
"=",
"flatpack_base_image",
"else",
":",
"raise",
"OsbsValidationException",
"(",
"\"flatpak_base_image must be provided\"",
")",
"required_secrets",
"=",
"data",
".",
"get",
"(",
"req_secrets_key",
",",
"[",
"]",
")",
"token_secrets",
"=",
"data",
".",
"get",
"(",
"token_secrets_key",
",",
"[",
"]",
")",
"self",
".",
"_set_required_secrets",
"(",
"required_secrets",
",",
"token_secrets",
")"
] |
Sets data from reactor config
|
[
"Sets",
"data",
"from",
"reactor",
"config"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_requestv2.py#L166-L208
|
train
|
projectatomic/osbs-client
|
osbs/build/build_requestv2.py
|
BuildRequestV2._set_required_secrets
|
def _set_required_secrets(self, required_secrets, token_secrets):
"""
Sets required secrets
"""
if self.user_params.build_type.value == BUILD_TYPE_ORCHESTRATOR:
required_secrets += token_secrets
if not required_secrets:
return
secrets = self.template['spec']['strategy']['customStrategy'].setdefault('secrets', [])
existing = set(secret_mount['secretSource']['name'] for secret_mount in secrets)
required_secrets = set(required_secrets)
already_set = required_secrets.intersection(existing)
if already_set:
logger.debug("secrets %s are already set", already_set)
for secret in required_secrets - existing:
secret_path = os.path.join(SECRETS_PATH, secret)
logger.info("Configuring %s secret at %s", secret, secret_path)
secrets.append({
'secretSource': {
'name': secret,
},
'mountPath': secret_path,
})
|
python
|
def _set_required_secrets(self, required_secrets, token_secrets):
"""
Sets required secrets
"""
if self.user_params.build_type.value == BUILD_TYPE_ORCHESTRATOR:
required_secrets += token_secrets
if not required_secrets:
return
secrets = self.template['spec']['strategy']['customStrategy'].setdefault('secrets', [])
existing = set(secret_mount['secretSource']['name'] for secret_mount in secrets)
required_secrets = set(required_secrets)
already_set = required_secrets.intersection(existing)
if already_set:
logger.debug("secrets %s are already set", already_set)
for secret in required_secrets - existing:
secret_path = os.path.join(SECRETS_PATH, secret)
logger.info("Configuring %s secret at %s", secret, secret_path)
secrets.append({
'secretSource': {
'name': secret,
},
'mountPath': secret_path,
})
|
[
"def",
"_set_required_secrets",
"(",
"self",
",",
"required_secrets",
",",
"token_secrets",
")",
":",
"if",
"self",
".",
"user_params",
".",
"build_type",
".",
"value",
"==",
"BUILD_TYPE_ORCHESTRATOR",
":",
"required_secrets",
"+=",
"token_secrets",
"if",
"not",
"required_secrets",
":",
"return",
"secrets",
"=",
"self",
".",
"template",
"[",
"'spec'",
"]",
"[",
"'strategy'",
"]",
"[",
"'customStrategy'",
"]",
".",
"setdefault",
"(",
"'secrets'",
",",
"[",
"]",
")",
"existing",
"=",
"set",
"(",
"secret_mount",
"[",
"'secretSource'",
"]",
"[",
"'name'",
"]",
"for",
"secret_mount",
"in",
"secrets",
")",
"required_secrets",
"=",
"set",
"(",
"required_secrets",
")",
"already_set",
"=",
"required_secrets",
".",
"intersection",
"(",
"existing",
")",
"if",
"already_set",
":",
"logger",
".",
"debug",
"(",
"\"secrets %s are already set\"",
",",
"already_set",
")",
"for",
"secret",
"in",
"required_secrets",
"-",
"existing",
":",
"secret_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"SECRETS_PATH",
",",
"secret",
")",
"logger",
".",
"info",
"(",
"\"Configuring %s secret at %s\"",
",",
"secret",
",",
"secret_path",
")",
"secrets",
".",
"append",
"(",
"{",
"'secretSource'",
":",
"{",
"'name'",
":",
"secret",
",",
"}",
",",
"'mountPath'",
":",
"secret_path",
",",
"}",
")"
] |
Sets required secrets
|
[
"Sets",
"required",
"secrets"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/build_requestv2.py#L210-L237
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsTemplate.remove_plugin
|
def remove_plugin(self, phase, name, reason=None):
"""
if config contains plugin, remove it
"""
for p in self.template[phase]:
if p.get('name') == name:
self.template[phase].remove(p)
if reason:
logger.info('Removing {}:{}, {}'.format(phase, name, reason))
break
|
python
|
def remove_plugin(self, phase, name, reason=None):
"""
if config contains plugin, remove it
"""
for p in self.template[phase]:
if p.get('name') == name:
self.template[phase].remove(p)
if reason:
logger.info('Removing {}:{}, {}'.format(phase, name, reason))
break
|
[
"def",
"remove_plugin",
"(",
"self",
",",
"phase",
",",
"name",
",",
"reason",
"=",
"None",
")",
":",
"for",
"p",
"in",
"self",
".",
"template",
"[",
"phase",
"]",
":",
"if",
"p",
".",
"get",
"(",
"'name'",
")",
"==",
"name",
":",
"self",
".",
"template",
"[",
"phase",
"]",
".",
"remove",
"(",
"p",
")",
"if",
"reason",
":",
"logger",
".",
"info",
"(",
"'Removing {}:{}, {}'",
".",
"format",
"(",
"phase",
",",
"name",
",",
"reason",
")",
")",
"break"
] |
if config contains plugin, remove it
|
[
"if",
"config",
"contains",
"plugin",
"remove",
"it"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L55-L64
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsTemplate.add_plugin
|
def add_plugin(self, phase, name, args, reason=None):
"""
if config has plugin, override it, else add it
"""
plugin_modified = False
for plugin in self.template[phase]:
if plugin['name'] == name:
plugin['args'] = args
plugin_modified = True
if not plugin_modified:
self.template[phase].append({"name": name, "args": args})
if reason:
logger.info('{}:{} with args {}, {}'.format(phase, name, args, reason))
|
python
|
def add_plugin(self, phase, name, args, reason=None):
"""
if config has plugin, override it, else add it
"""
plugin_modified = False
for plugin in self.template[phase]:
if plugin['name'] == name:
plugin['args'] = args
plugin_modified = True
if not plugin_modified:
self.template[phase].append({"name": name, "args": args})
if reason:
logger.info('{}:{} with args {}, {}'.format(phase, name, args, reason))
|
[
"def",
"add_plugin",
"(",
"self",
",",
"phase",
",",
"name",
",",
"args",
",",
"reason",
"=",
"None",
")",
":",
"plugin_modified",
"=",
"False",
"for",
"plugin",
"in",
"self",
".",
"template",
"[",
"phase",
"]",
":",
"if",
"plugin",
"[",
"'name'",
"]",
"==",
"name",
":",
"plugin",
"[",
"'args'",
"]",
"=",
"args",
"plugin_modified",
"=",
"True",
"if",
"not",
"plugin_modified",
":",
"self",
".",
"template",
"[",
"phase",
"]",
".",
"append",
"(",
"{",
"\"name\"",
":",
"name",
",",
"\"args\"",
":",
"args",
"}",
")",
"if",
"reason",
":",
"logger",
".",
"info",
"(",
"'{}:{} with args {}, {}'",
".",
"format",
"(",
"phase",
",",
"name",
",",
"args",
",",
"reason",
")",
")"
] |
if config has plugin, override it, else add it
|
[
"if",
"config",
"has",
"plugin",
"override",
"it",
"else",
"add",
"it"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L66-L80
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsTemplate.get_plugin_conf
|
def get_plugin_conf(self, phase, name):
"""
Return the configuration for a plugin.
Raises KeyError if there are no plugins of that type.
Raises IndexError if the named plugin is not listed.
"""
match = [x for x in self.template[phase] if x.get('name') == name]
return match[0]
|
python
|
def get_plugin_conf(self, phase, name):
"""
Return the configuration for a plugin.
Raises KeyError if there are no plugins of that type.
Raises IndexError if the named plugin is not listed.
"""
match = [x for x in self.template[phase] if x.get('name') == name]
return match[0]
|
[
"def",
"get_plugin_conf",
"(",
"self",
",",
"phase",
",",
"name",
")",
":",
"match",
"=",
"[",
"x",
"for",
"x",
"in",
"self",
".",
"template",
"[",
"phase",
"]",
"if",
"x",
".",
"get",
"(",
"'name'",
")",
"==",
"name",
"]",
"return",
"match",
"[",
"0",
"]"
] |
Return the configuration for a plugin.
Raises KeyError if there are no plugins of that type.
Raises IndexError if the named plugin is not listed.
|
[
"Return",
"the",
"configuration",
"for",
"a",
"plugin",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L82-L90
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsTemplate.has_plugin_conf
|
def has_plugin_conf(self, phase, name):
"""
Check whether a plugin is configured.
"""
try:
self.get_plugin_conf(phase, name)
return True
except (KeyError, IndexError):
return False
|
python
|
def has_plugin_conf(self, phase, name):
"""
Check whether a plugin is configured.
"""
try:
self.get_plugin_conf(phase, name)
return True
except (KeyError, IndexError):
return False
|
[
"def",
"has_plugin_conf",
"(",
"self",
",",
"phase",
",",
"name",
")",
":",
"try",
":",
"self",
".",
"get_plugin_conf",
"(",
"phase",
",",
"name",
")",
"return",
"True",
"except",
"(",
"KeyError",
",",
"IndexError",
")",
":",
"return",
"False"
] |
Check whether a plugin is configured.
|
[
"Check",
"whether",
"a",
"plugin",
"is",
"configured",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L92-L100
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.adjust_for_scratch
|
def adjust_for_scratch(self):
"""
Remove certain plugins in order to handle the "scratch build"
scenario. Scratch builds must not affect subsequent builds,
and should not be imported into Koji.
"""
if self.user_params.scratch.value:
remove_plugins = [
("prebuild_plugins", "koji_parent"),
("postbuild_plugins", "compress"), # required only to make an archive for Koji
("postbuild_plugins", "pulp_pull"), # required only to make an archive for Koji
("postbuild_plugins", "compare_components"),
("postbuild_plugins", "import_image"),
("exit_plugins", "koji_promote"),
("exit_plugins", "koji_tag_build"),
("exit_plugins", "import_image"),
("prebuild_plugins", "check_and_set_rebuild"),
("prebuild_plugins", "stop_autorebuild_if_disabled")
]
if not self.has_tag_suffixes_placeholder():
remove_plugins.append(("postbuild_plugins", "tag_from_config"))
for when, which in remove_plugins:
self.pt.remove_plugin(when, which, 'removed from scratch build request')
|
python
|
def adjust_for_scratch(self):
"""
Remove certain plugins in order to handle the "scratch build"
scenario. Scratch builds must not affect subsequent builds,
and should not be imported into Koji.
"""
if self.user_params.scratch.value:
remove_plugins = [
("prebuild_plugins", "koji_parent"),
("postbuild_plugins", "compress"), # required only to make an archive for Koji
("postbuild_plugins", "pulp_pull"), # required only to make an archive for Koji
("postbuild_plugins", "compare_components"),
("postbuild_plugins", "import_image"),
("exit_plugins", "koji_promote"),
("exit_plugins", "koji_tag_build"),
("exit_plugins", "import_image"),
("prebuild_plugins", "check_and_set_rebuild"),
("prebuild_plugins", "stop_autorebuild_if_disabled")
]
if not self.has_tag_suffixes_placeholder():
remove_plugins.append(("postbuild_plugins", "tag_from_config"))
for when, which in remove_plugins:
self.pt.remove_plugin(when, which, 'removed from scratch build request')
|
[
"def",
"adjust_for_scratch",
"(",
"self",
")",
":",
"if",
"self",
".",
"user_params",
".",
"scratch",
".",
"value",
":",
"remove_plugins",
"=",
"[",
"(",
"\"prebuild_plugins\"",
",",
"\"koji_parent\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"compress\"",
")",
",",
"# required only to make an archive for Koji",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_pull\"",
")",
",",
"# required only to make an archive for Koji",
"(",
"\"postbuild_plugins\"",
",",
"\"compare_components\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"import_image\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"koji_promote\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"koji_tag_build\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"import_image\"",
")",
",",
"(",
"\"prebuild_plugins\"",
",",
"\"check_and_set_rebuild\"",
")",
",",
"(",
"\"prebuild_plugins\"",
",",
"\"stop_autorebuild_if_disabled\"",
")",
"]",
"if",
"not",
"self",
".",
"has_tag_suffixes_placeholder",
"(",
")",
":",
"remove_plugins",
".",
"append",
"(",
"(",
"\"postbuild_plugins\"",
",",
"\"tag_from_config\"",
")",
")",
"for",
"when",
",",
"which",
"in",
"remove_plugins",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"when",
",",
"which",
",",
"'removed from scratch build request'",
")"
] |
Remove certain plugins in order to handle the "scratch build"
scenario. Scratch builds must not affect subsequent builds,
and should not be imported into Koji.
|
[
"Remove",
"certain",
"plugins",
"in",
"order",
"to",
"handle",
"the",
"scratch",
"build",
"scenario",
".",
"Scratch",
"builds",
"must",
"not",
"affect",
"subsequent",
"builds",
"and",
"should",
"not",
"be",
"imported",
"into",
"Koji",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L149-L173
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.adjust_for_isolated
|
def adjust_for_isolated(self):
"""
Remove certain plugins in order to handle the "isolated build"
scenario.
"""
if self.user_params.isolated.value:
remove_plugins = [
("prebuild_plugins", "check_and_set_rebuild"),
("prebuild_plugins", "stop_autorebuild_if_disabled")
]
for when, which in remove_plugins:
self.pt.remove_plugin(when, which, 'removed from isolated build request')
|
python
|
def adjust_for_isolated(self):
"""
Remove certain plugins in order to handle the "isolated build"
scenario.
"""
if self.user_params.isolated.value:
remove_plugins = [
("prebuild_plugins", "check_and_set_rebuild"),
("prebuild_plugins", "stop_autorebuild_if_disabled")
]
for when, which in remove_plugins:
self.pt.remove_plugin(when, which, 'removed from isolated build request')
|
[
"def",
"adjust_for_isolated",
"(",
"self",
")",
":",
"if",
"self",
".",
"user_params",
".",
"isolated",
".",
"value",
":",
"remove_plugins",
"=",
"[",
"(",
"\"prebuild_plugins\"",
",",
"\"check_and_set_rebuild\"",
")",
",",
"(",
"\"prebuild_plugins\"",
",",
"\"stop_autorebuild_if_disabled\"",
")",
"]",
"for",
"when",
",",
"which",
"in",
"remove_plugins",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"when",
",",
"which",
",",
"'removed from isolated build request'",
")"
] |
Remove certain plugins in order to handle the "isolated build"
scenario.
|
[
"Remove",
"certain",
"plugins",
"in",
"order",
"to",
"handle",
"the",
"isolated",
"build",
"scenario",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L175-L187
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.adjust_for_flatpak
|
def adjust_for_flatpak(self):
"""
Remove plugins that don't work when building Flatpaks
"""
if self.user_params.flatpak.value:
remove_plugins = [
("prebuild_plugins", "resolve_composes"),
# We'll extract the filesystem anyways for a Flatpak instead of exporting
# the docker image directly, so squash just slows things down.
("prepublish_plugins", "squash"),
# Pulp can't currently handle Flatpaks, which are OCI images
("postbuild_plugins", "pulp_push"),
("postbuild_plugins", "pulp_tag"),
("postbuild_plugins", "pulp_sync"),
("exit_plugins", "pulp_publish"),
("exit_plugins", "pulp_pull"),
# delete_from_registry is used for deleting builds from the temporary registry
# that pulp_sync mirrors from.
("exit_plugins", "delete_from_registry"),
]
for when, which in remove_plugins:
self.pt.remove_plugin(when, which, 'not needed for flatpak build')
|
python
|
def adjust_for_flatpak(self):
"""
Remove plugins that don't work when building Flatpaks
"""
if self.user_params.flatpak.value:
remove_plugins = [
("prebuild_plugins", "resolve_composes"),
# We'll extract the filesystem anyways for a Flatpak instead of exporting
# the docker image directly, so squash just slows things down.
("prepublish_plugins", "squash"),
# Pulp can't currently handle Flatpaks, which are OCI images
("postbuild_plugins", "pulp_push"),
("postbuild_plugins", "pulp_tag"),
("postbuild_plugins", "pulp_sync"),
("exit_plugins", "pulp_publish"),
("exit_plugins", "pulp_pull"),
# delete_from_registry is used for deleting builds from the temporary registry
# that pulp_sync mirrors from.
("exit_plugins", "delete_from_registry"),
]
for when, which in remove_plugins:
self.pt.remove_plugin(when, which, 'not needed for flatpak build')
|
[
"def",
"adjust_for_flatpak",
"(",
"self",
")",
":",
"if",
"self",
".",
"user_params",
".",
"flatpak",
".",
"value",
":",
"remove_plugins",
"=",
"[",
"(",
"\"prebuild_plugins\"",
",",
"\"resolve_composes\"",
")",
",",
"# We'll extract the filesystem anyways for a Flatpak instead of exporting",
"# the docker image directly, so squash just slows things down.",
"(",
"\"prepublish_plugins\"",
",",
"\"squash\"",
")",
",",
"# Pulp can't currently handle Flatpaks, which are OCI images",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_push\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_tag\"",
")",
",",
"(",
"\"postbuild_plugins\"",
",",
"\"pulp_sync\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"pulp_publish\"",
")",
",",
"(",
"\"exit_plugins\"",
",",
"\"pulp_pull\"",
")",
",",
"# delete_from_registry is used for deleting builds from the temporary registry",
"# that pulp_sync mirrors from.",
"(",
"\"exit_plugins\"",
",",
"\"delete_from_registry\"",
")",
",",
"]",
"for",
"when",
",",
"which",
"in",
"remove_plugins",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"when",
",",
"which",
",",
"'not needed for flatpak build'",
")"
] |
Remove plugins that don't work when building Flatpaks
|
[
"Remove",
"plugins",
"that",
"don",
"t",
"work",
"when",
"building",
"Flatpaks"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L189-L210
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_customizations
|
def render_customizations(self):
"""
Customize template for site user specified customizations
"""
disable_plugins = self.pt.customize_conf.get('disable_plugins', [])
if not disable_plugins:
logger.debug('No site-user specified plugins to disable')
else:
for plugin in disable_plugins:
try:
self.pt.remove_plugin(plugin['plugin_type'], plugin['plugin_name'],
'disabled at user request')
except KeyError:
# Malformed config
logger.info('Invalid custom configuration found for disable_plugins')
enable_plugins = self.pt.customize_conf.get('enable_plugins', [])
if not enable_plugins:
logger.debug('No site-user specified plugins to enable"')
else:
for plugin in enable_plugins:
try:
msg = 'enabled at user request'
self.pt.add_plugin(plugin['plugin_type'], plugin['plugin_name'],
plugin['plugin_args'], msg)
except KeyError:
# Malformed config
logger.info('Invalid custom configuration found for enable_plugins')
|
python
|
def render_customizations(self):
"""
Customize template for site user specified customizations
"""
disable_plugins = self.pt.customize_conf.get('disable_plugins', [])
if not disable_plugins:
logger.debug('No site-user specified plugins to disable')
else:
for plugin in disable_plugins:
try:
self.pt.remove_plugin(plugin['plugin_type'], plugin['plugin_name'],
'disabled at user request')
except KeyError:
# Malformed config
logger.info('Invalid custom configuration found for disable_plugins')
enable_plugins = self.pt.customize_conf.get('enable_plugins', [])
if not enable_plugins:
logger.debug('No site-user specified plugins to enable"')
else:
for plugin in enable_plugins:
try:
msg = 'enabled at user request'
self.pt.add_plugin(plugin['plugin_type'], plugin['plugin_name'],
plugin['plugin_args'], msg)
except KeyError:
# Malformed config
logger.info('Invalid custom configuration found for enable_plugins')
|
[
"def",
"render_customizations",
"(",
"self",
")",
":",
"disable_plugins",
"=",
"self",
".",
"pt",
".",
"customize_conf",
".",
"get",
"(",
"'disable_plugins'",
",",
"[",
"]",
")",
"if",
"not",
"disable_plugins",
":",
"logger",
".",
"debug",
"(",
"'No site-user specified plugins to disable'",
")",
"else",
":",
"for",
"plugin",
"in",
"disable_plugins",
":",
"try",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"plugin",
"[",
"'plugin_type'",
"]",
",",
"plugin",
"[",
"'plugin_name'",
"]",
",",
"'disabled at user request'",
")",
"except",
"KeyError",
":",
"# Malformed config",
"logger",
".",
"info",
"(",
"'Invalid custom configuration found for disable_plugins'",
")",
"enable_plugins",
"=",
"self",
".",
"pt",
".",
"customize_conf",
".",
"get",
"(",
"'enable_plugins'",
",",
"[",
"]",
")",
"if",
"not",
"enable_plugins",
":",
"logger",
".",
"debug",
"(",
"'No site-user specified plugins to enable\"'",
")",
"else",
":",
"for",
"plugin",
"in",
"enable_plugins",
":",
"try",
":",
"msg",
"=",
"'enabled at user request'",
"self",
".",
"pt",
".",
"add_plugin",
"(",
"plugin",
"[",
"'plugin_type'",
"]",
",",
"plugin",
"[",
"'plugin_name'",
"]",
",",
"plugin",
"[",
"'plugin_args'",
"]",
",",
"msg",
")",
"except",
"KeyError",
":",
"# Malformed config",
"logger",
".",
"info",
"(",
"'Invalid custom configuration found for enable_plugins'",
")"
] |
Customize template for site user specified customizations
|
[
"Customize",
"template",
"for",
"site",
"user",
"specified",
"customizations"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L239-L266
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_koji
|
def render_koji(self):
"""
if there is yum repo in user params, don't pick stuff from koji
"""
phase = 'prebuild_plugins'
plugin = 'koji'
if not self.pt.has_plugin_conf(phase, plugin):
return
if self.user_params.yum_repourls.value:
self.pt.remove_plugin(phase, plugin, 'there is a yum repo user parameter')
elif not self.pt.set_plugin_arg_valid(phase, plugin, "target",
self.user_params.koji_target.value):
self.pt.remove_plugin(phase, plugin, 'no koji target supplied in user parameters')
|
python
|
def render_koji(self):
"""
if there is yum repo in user params, don't pick stuff from koji
"""
phase = 'prebuild_plugins'
plugin = 'koji'
if not self.pt.has_plugin_conf(phase, plugin):
return
if self.user_params.yum_repourls.value:
self.pt.remove_plugin(phase, plugin, 'there is a yum repo user parameter')
elif not self.pt.set_plugin_arg_valid(phase, plugin, "target",
self.user_params.koji_target.value):
self.pt.remove_plugin(phase, plugin, 'no koji target supplied in user parameters')
|
[
"def",
"render_koji",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'koji'",
"if",
"not",
"self",
".",
"pt",
".",
"has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"if",
"self",
".",
"user_params",
".",
"yum_repourls",
".",
"value",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
",",
"'there is a yum repo user parameter'",
")",
"elif",
"not",
"self",
".",
"pt",
".",
"set_plugin_arg_valid",
"(",
"phase",
",",
"plugin",
",",
"\"target\"",
",",
"self",
".",
"user_params",
".",
"koji_target",
".",
"value",
")",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
",",
"'no koji target supplied in user parameters'",
")"
] |
if there is yum repo in user params, don't pick stuff from koji
|
[
"if",
"there",
"is",
"yum",
"repo",
"in",
"user",
"params",
"don",
"t",
"pick",
"stuff",
"from",
"koji"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L285-L298
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_bump_release
|
def render_bump_release(self):
"""
If the bump_release plugin is present, configure it
"""
phase = 'prebuild_plugins'
plugin = 'bump_release'
if not self.pt.has_plugin_conf(phase, plugin):
return
if self.user_params.release.value:
self.pt.remove_plugin(phase, plugin, 'release value supplied as user parameter')
return
# For flatpak, we want a name-version-release of
# <name>-<stream>-<module_build_version>.<n>, where the .<n> makes
# sure that the build is unique in Koji
if self.user_params.flatpak.value:
self.pt.set_plugin_arg(phase, plugin, 'append', True)
|
python
|
def render_bump_release(self):
"""
If the bump_release plugin is present, configure it
"""
phase = 'prebuild_plugins'
plugin = 'bump_release'
if not self.pt.has_plugin_conf(phase, plugin):
return
if self.user_params.release.value:
self.pt.remove_plugin(phase, plugin, 'release value supplied as user parameter')
return
# For flatpak, we want a name-version-release of
# <name>-<stream>-<module_build_version>.<n>, where the .<n> makes
# sure that the build is unique in Koji
if self.user_params.flatpak.value:
self.pt.set_plugin_arg(phase, plugin, 'append', True)
|
[
"def",
"render_bump_release",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'bump_release'",
"if",
"not",
"self",
".",
"pt",
".",
"has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"if",
"self",
".",
"user_params",
".",
"release",
".",
"value",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"phase",
",",
"plugin",
",",
"'release value supplied as user parameter'",
")",
"return",
"# For flatpak, we want a name-version-release of",
"# <name>-<stream>-<module_build_version>.<n>, where the .<n> makes",
"# sure that the build is unique in Koji",
"if",
"self",
".",
"user_params",
".",
"flatpak",
".",
"value",
":",
"self",
".",
"pt",
".",
"set_plugin_arg",
"(",
"phase",
",",
"plugin",
",",
"'append'",
",",
"True",
")"
] |
If the bump_release plugin is present, configure it
|
[
"If",
"the",
"bump_release",
"plugin",
"is",
"present",
"configure",
"it"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L300-L317
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_check_and_set_platforms
|
def render_check_and_set_platforms(self):
"""
If the check_and_set_platforms plugin is present, configure it
"""
phase = 'prebuild_plugins'
plugin = 'check_and_set_platforms'
if not self.pt.has_plugin_conf(phase, plugin):
return
if self.user_params.koji_target.value:
self.pt.set_plugin_arg(phase, plugin, "koji_target",
self.user_params.koji_target.value)
|
python
|
def render_check_and_set_platforms(self):
"""
If the check_and_set_platforms plugin is present, configure it
"""
phase = 'prebuild_plugins'
plugin = 'check_and_set_platforms'
if not self.pt.has_plugin_conf(phase, plugin):
return
if self.user_params.koji_target.value:
self.pt.set_plugin_arg(phase, plugin, "koji_target",
self.user_params.koji_target.value)
|
[
"def",
"render_check_and_set_platforms",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'check_and_set_platforms'",
"if",
"not",
"self",
".",
"pt",
".",
"has_plugin_conf",
"(",
"phase",
",",
"plugin",
")",
":",
"return",
"if",
"self",
".",
"user_params",
".",
"koji_target",
".",
"value",
":",
"self",
".",
"pt",
".",
"set_plugin_arg",
"(",
"phase",
",",
"plugin",
",",
"\"koji_target\"",
",",
"self",
".",
"user_params",
".",
"koji_target",
".",
"value",
")"
] |
If the check_and_set_platforms plugin is present, configure it
|
[
"If",
"the",
"check_and_set_platforms",
"plugin",
"is",
"present",
"configure",
"it"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L319-L330
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_import_image
|
def render_import_image(self, use_auth=None):
"""
Configure the import_image plugin
"""
# import_image is a multi-phase plugin
if self.user_params.imagestream_name.value is None:
self.pt.remove_plugin('exit_plugins', 'import_image',
'imagestream not in user parameters')
elif self.pt.has_plugin_conf('exit_plugins', 'import_image'):
self.pt.set_plugin_arg('exit_plugins', 'import_image', 'imagestream',
self.user_params.imagestream_name.value)
|
python
|
def render_import_image(self, use_auth=None):
"""
Configure the import_image plugin
"""
# import_image is a multi-phase plugin
if self.user_params.imagestream_name.value is None:
self.pt.remove_plugin('exit_plugins', 'import_image',
'imagestream not in user parameters')
elif self.pt.has_plugin_conf('exit_plugins', 'import_image'):
self.pt.set_plugin_arg('exit_plugins', 'import_image', 'imagestream',
self.user_params.imagestream_name.value)
|
[
"def",
"render_import_image",
"(",
"self",
",",
"use_auth",
"=",
"None",
")",
":",
"# import_image is a multi-phase plugin",
"if",
"self",
".",
"user_params",
".",
"imagestream_name",
".",
"value",
"is",
"None",
":",
"self",
".",
"pt",
".",
"remove_plugin",
"(",
"'exit_plugins'",
",",
"'import_image'",
",",
"'imagestream not in user parameters'",
")",
"elif",
"self",
".",
"pt",
".",
"has_plugin_conf",
"(",
"'exit_plugins'",
",",
"'import_image'",
")",
":",
"self",
".",
"pt",
".",
"set_plugin_arg",
"(",
"'exit_plugins'",
",",
"'import_image'",
",",
"'imagestream'",
",",
"self",
".",
"user_params",
".",
"imagestream_name",
".",
"value",
")"
] |
Configure the import_image plugin
|
[
"Configure",
"the",
"import_image",
"plugin"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L332-L342
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_tag_from_config
|
def render_tag_from_config(self):
"""Configure tag_from_config plugin"""
phase = 'postbuild_plugins'
plugin = 'tag_from_config'
if not self.has_tag_suffixes_placeholder():
return
unique_tag = self.user_params.image_tag.value.split(':')[-1]
tag_suffixes = {'unique': [unique_tag], 'primary': []}
if self.user_params.build_type.value == BUILD_TYPE_ORCHESTRATOR:
additional_tags = self.user_params.additional_tags.value or set()
if self.user_params.scratch.value:
pass
elif self.user_params.isolated.value:
tag_suffixes['primary'].extend(['{version}-{release}'])
elif self.user_params.tags_from_yaml.value:
tag_suffixes['primary'].extend(['{version}-{release}'])
tag_suffixes['primary'].extend(additional_tags)
else:
tag_suffixes['primary'].extend(['latest', '{version}', '{version}-{release}'])
tag_suffixes['primary'].extend(additional_tags)
self.pt.set_plugin_arg(phase, plugin, 'tag_suffixes', tag_suffixes)
|
python
|
def render_tag_from_config(self):
"""Configure tag_from_config plugin"""
phase = 'postbuild_plugins'
plugin = 'tag_from_config'
if not self.has_tag_suffixes_placeholder():
return
unique_tag = self.user_params.image_tag.value.split(':')[-1]
tag_suffixes = {'unique': [unique_tag], 'primary': []}
if self.user_params.build_type.value == BUILD_TYPE_ORCHESTRATOR:
additional_tags = self.user_params.additional_tags.value or set()
if self.user_params.scratch.value:
pass
elif self.user_params.isolated.value:
tag_suffixes['primary'].extend(['{version}-{release}'])
elif self.user_params.tags_from_yaml.value:
tag_suffixes['primary'].extend(['{version}-{release}'])
tag_suffixes['primary'].extend(additional_tags)
else:
tag_suffixes['primary'].extend(['latest', '{version}', '{version}-{release}'])
tag_suffixes['primary'].extend(additional_tags)
self.pt.set_plugin_arg(phase, plugin, 'tag_suffixes', tag_suffixes)
|
[
"def",
"render_tag_from_config",
"(",
"self",
")",
":",
"phase",
"=",
"'postbuild_plugins'",
"plugin",
"=",
"'tag_from_config'",
"if",
"not",
"self",
".",
"has_tag_suffixes_placeholder",
"(",
")",
":",
"return",
"unique_tag",
"=",
"self",
".",
"user_params",
".",
"image_tag",
".",
"value",
".",
"split",
"(",
"':'",
")",
"[",
"-",
"1",
"]",
"tag_suffixes",
"=",
"{",
"'unique'",
":",
"[",
"unique_tag",
"]",
",",
"'primary'",
":",
"[",
"]",
"}",
"if",
"self",
".",
"user_params",
".",
"build_type",
".",
"value",
"==",
"BUILD_TYPE_ORCHESTRATOR",
":",
"additional_tags",
"=",
"self",
".",
"user_params",
".",
"additional_tags",
".",
"value",
"or",
"set",
"(",
")",
"if",
"self",
".",
"user_params",
".",
"scratch",
".",
"value",
":",
"pass",
"elif",
"self",
".",
"user_params",
".",
"isolated",
".",
"value",
":",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"[",
"'{version}-{release}'",
"]",
")",
"elif",
"self",
".",
"user_params",
".",
"tags_from_yaml",
".",
"value",
":",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"[",
"'{version}-{release}'",
"]",
")",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"additional_tags",
")",
"else",
":",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"[",
"'latest'",
",",
"'{version}'",
",",
"'{version}-{release}'",
"]",
")",
"tag_suffixes",
"[",
"'primary'",
"]",
".",
"extend",
"(",
"additional_tags",
")",
"self",
".",
"pt",
".",
"set_plugin_arg",
"(",
"phase",
",",
"plugin",
",",
"'tag_suffixes'",
",",
"tag_suffixes",
")"
] |
Configure tag_from_config plugin
|
[
"Configure",
"tag_from_config",
"plugin"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L459-L483
|
train
|
projectatomic/osbs-client
|
osbs/build/plugins_configuration.py
|
PluginsConfiguration.render_pull_base_image
|
def render_pull_base_image(self):
"""Configure pull_base_image"""
phase = 'prebuild_plugins'
plugin = 'pull_base_image'
if self.user_params.parent_images_digests.value:
self.pt.set_plugin_arg(phase, plugin, 'parent_images_digests',
self.user_params.parent_images_digests.value)
|
python
|
def render_pull_base_image(self):
"""Configure pull_base_image"""
phase = 'prebuild_plugins'
plugin = 'pull_base_image'
if self.user_params.parent_images_digests.value:
self.pt.set_plugin_arg(phase, plugin, 'parent_images_digests',
self.user_params.parent_images_digests.value)
|
[
"def",
"render_pull_base_image",
"(",
"self",
")",
":",
"phase",
"=",
"'prebuild_plugins'",
"plugin",
"=",
"'pull_base_image'",
"if",
"self",
".",
"user_params",
".",
"parent_images_digests",
".",
"value",
":",
"self",
".",
"pt",
".",
"set_plugin_arg",
"(",
"phase",
",",
"plugin",
",",
"'parent_images_digests'",
",",
"self",
".",
"user_params",
".",
"parent_images_digests",
".",
"value",
")"
] |
Configure pull_base_image
|
[
"Configure",
"pull_base_image"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/build/plugins_configuration.py#L485-L492
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.get_user
|
def get_user(self, username="~"):
"""
get info about user (if no user specified, use the one initiating request)
:param username: str, name of user to get info about, default="~"
:return: dict
"""
url = self._build_url("users/%s/" % username, _prepend_namespace=False)
response = self._get(url)
check_response(response)
return response
|
python
|
def get_user(self, username="~"):
"""
get info about user (if no user specified, use the one initiating request)
:param username: str, name of user to get info about, default="~"
:return: dict
"""
url = self._build_url("users/%s/" % username, _prepend_namespace=False)
response = self._get(url)
check_response(response)
return response
|
[
"def",
"get_user",
"(",
"self",
",",
"username",
"=",
"\"~\"",
")",
":",
"url",
"=",
"self",
".",
"_build_url",
"(",
"\"users/%s/\"",
"%",
"username",
",",
"_prepend_namespace",
"=",
"False",
")",
"response",
"=",
"self",
".",
"_get",
"(",
"url",
")",
"check_response",
"(",
"response",
")",
"return",
"response"
] |
get info about user (if no user specified, use the one initiating request)
:param username: str, name of user to get info about, default="~"
:return: dict
|
[
"get",
"info",
"about",
"user",
"(",
"if",
"no",
"user",
"specified",
"use",
"the",
"one",
"initiating",
"request",
")"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L233-L243
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.create_build
|
def create_build(self, build_json):
"""
:return:
"""
url = self._build_url("builds/")
logger.debug(build_json)
return self._post(url, data=json.dumps(build_json),
headers={"Content-Type": "application/json"})
|
python
|
def create_build(self, build_json):
"""
:return:
"""
url = self._build_url("builds/")
logger.debug(build_json)
return self._post(url, data=json.dumps(build_json),
headers={"Content-Type": "application/json"})
|
[
"def",
"create_build",
"(",
"self",
",",
"build_json",
")",
":",
"url",
"=",
"self",
".",
"_build_url",
"(",
"\"builds/\"",
")",
"logger",
".",
"debug",
"(",
"build_json",
")",
"return",
"self",
".",
"_post",
"(",
"url",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"build_json",
")",
",",
"headers",
"=",
"{",
"\"Content-Type\"",
":",
"\"application/json\"",
"}",
")"
] |
:return:
|
[
":",
"return",
":"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L296-L303
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.get_all_build_configs_by_labels
|
def get_all_build_configs_by_labels(self, label_selectors):
"""
Returns all builds matching a given set of label selectors. It is up to the
calling function to filter the results.
"""
labels = ['%s=%s' % (field, value) for field, value in label_selectors]
labels = ','.join(labels)
url = self._build_url("buildconfigs/", labelSelector=labels)
return self._get(url).json()['items']
|
python
|
def get_all_build_configs_by_labels(self, label_selectors):
"""
Returns all builds matching a given set of label selectors. It is up to the
calling function to filter the results.
"""
labels = ['%s=%s' % (field, value) for field, value in label_selectors]
labels = ','.join(labels)
url = self._build_url("buildconfigs/", labelSelector=labels)
return self._get(url).json()['items']
|
[
"def",
"get_all_build_configs_by_labels",
"(",
"self",
",",
"label_selectors",
")",
":",
"labels",
"=",
"[",
"'%s=%s'",
"%",
"(",
"field",
",",
"value",
")",
"for",
"field",
",",
"value",
"in",
"label_selectors",
"]",
"labels",
"=",
"','",
".",
"join",
"(",
"labels",
")",
"url",
"=",
"self",
".",
"_build_url",
"(",
"\"buildconfigs/\"",
",",
"labelSelector",
"=",
"labels",
")",
"return",
"self",
".",
"_get",
"(",
"url",
")",
".",
"json",
"(",
")",
"[",
"'items'",
"]"
] |
Returns all builds matching a given set of label selectors. It is up to the
calling function to filter the results.
|
[
"Returns",
"all",
"builds",
"matching",
"a",
"given",
"set",
"of",
"label",
"selectors",
".",
"It",
"is",
"up",
"to",
"the",
"calling",
"function",
"to",
"filter",
"the",
"results",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L326-L334
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.get_build_config_by_labels
|
def get_build_config_by_labels(self, label_selectors):
"""
Returns a build config matching the given label
selectors. This method will raise OsbsException
if not exactly one build config is found.
"""
items = self.get_all_build_configs_by_labels(label_selectors)
if not items:
raise OsbsException(
"Build config not found for labels: %r" %
(label_selectors, ))
if len(items) > 1:
raise OsbsException(
"More than one build config found for labels: %r" %
(label_selectors, ))
return items[0]
|
python
|
def get_build_config_by_labels(self, label_selectors):
"""
Returns a build config matching the given label
selectors. This method will raise OsbsException
if not exactly one build config is found.
"""
items = self.get_all_build_configs_by_labels(label_selectors)
if not items:
raise OsbsException(
"Build config not found for labels: %r" %
(label_selectors, ))
if len(items) > 1:
raise OsbsException(
"More than one build config found for labels: %r" %
(label_selectors, ))
return items[0]
|
[
"def",
"get_build_config_by_labels",
"(",
"self",
",",
"label_selectors",
")",
":",
"items",
"=",
"self",
".",
"get_all_build_configs_by_labels",
"(",
"label_selectors",
")",
"if",
"not",
"items",
":",
"raise",
"OsbsException",
"(",
"\"Build config not found for labels: %r\"",
"%",
"(",
"label_selectors",
",",
")",
")",
"if",
"len",
"(",
"items",
")",
">",
"1",
":",
"raise",
"OsbsException",
"(",
"\"More than one build config found for labels: %r\"",
"%",
"(",
"label_selectors",
",",
")",
")",
"return",
"items",
"[",
"0",
"]"
] |
Returns a build config matching the given label
selectors. This method will raise OsbsException
if not exactly one build config is found.
|
[
"Returns",
"a",
"build",
"config",
"matching",
"the",
"given",
"label",
"selectors",
".",
"This",
"method",
"will",
"raise",
"OsbsException",
"if",
"not",
"exactly",
"one",
"build",
"config",
"is",
"found",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L336-L353
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.get_build_config_by_labels_filtered
|
def get_build_config_by_labels_filtered(self, label_selectors, filter_key, filter_value):
"""
Returns a build config matching the given label selectors, filtering against
another predetermined value. This method will raise OsbsException
if not exactly one build config is found after filtering.
"""
items = self.get_all_build_configs_by_labels(label_selectors)
if filter_value is not None:
build_configs = []
for build_config in items:
match_value = graceful_chain_get(build_config, *filter_key.split('.'))
if filter_value == match_value:
build_configs.append(build_config)
items = build_configs
if not items:
raise OsbsException(
"Build config not found for labels: %r" %
(label_selectors, ))
if len(items) > 1:
raise OsbsException(
"More than one build config found for labels: %r" %
(label_selectors, ))
return items[0]
|
python
|
def get_build_config_by_labels_filtered(self, label_selectors, filter_key, filter_value):
"""
Returns a build config matching the given label selectors, filtering against
another predetermined value. This method will raise OsbsException
if not exactly one build config is found after filtering.
"""
items = self.get_all_build_configs_by_labels(label_selectors)
if filter_value is not None:
build_configs = []
for build_config in items:
match_value = graceful_chain_get(build_config, *filter_key.split('.'))
if filter_value == match_value:
build_configs.append(build_config)
items = build_configs
if not items:
raise OsbsException(
"Build config not found for labels: %r" %
(label_selectors, ))
if len(items) > 1:
raise OsbsException(
"More than one build config found for labels: %r" %
(label_selectors, ))
return items[0]
|
[
"def",
"get_build_config_by_labels_filtered",
"(",
"self",
",",
"label_selectors",
",",
"filter_key",
",",
"filter_value",
")",
":",
"items",
"=",
"self",
".",
"get_all_build_configs_by_labels",
"(",
"label_selectors",
")",
"if",
"filter_value",
"is",
"not",
"None",
":",
"build_configs",
"=",
"[",
"]",
"for",
"build_config",
"in",
"items",
":",
"match_value",
"=",
"graceful_chain_get",
"(",
"build_config",
",",
"*",
"filter_key",
".",
"split",
"(",
"'.'",
")",
")",
"if",
"filter_value",
"==",
"match_value",
":",
"build_configs",
".",
"append",
"(",
"build_config",
")",
"items",
"=",
"build_configs",
"if",
"not",
"items",
":",
"raise",
"OsbsException",
"(",
"\"Build config not found for labels: %r\"",
"%",
"(",
"label_selectors",
",",
")",
")",
"if",
"len",
"(",
"items",
")",
">",
"1",
":",
"raise",
"OsbsException",
"(",
"\"More than one build config found for labels: %r\"",
"%",
"(",
"label_selectors",
",",
")",
")",
"return",
"items",
"[",
"0",
"]"
] |
Returns a build config matching the given label selectors, filtering against
another predetermined value. This method will raise OsbsException
if not exactly one build config is found after filtering.
|
[
"Returns",
"a",
"build",
"config",
"matching",
"the",
"given",
"label",
"selectors",
"filtering",
"against",
"another",
"predetermined",
"value",
".",
"This",
"method",
"will",
"raise",
"OsbsException",
"if",
"not",
"exactly",
"one",
"build",
"config",
"is",
"found",
"after",
"filtering",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L355-L379
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.create_build_config
|
def create_build_config(self, build_config_json):
"""
:return:
"""
url = self._build_url("buildconfigs/")
return self._post(url, data=build_config_json,
headers={"Content-Type": "application/json"})
|
python
|
def create_build_config(self, build_config_json):
"""
:return:
"""
url = self._build_url("buildconfigs/")
return self._post(url, data=build_config_json,
headers={"Content-Type": "application/json"})
|
[
"def",
"create_build_config",
"(",
"self",
",",
"build_config_json",
")",
":",
"url",
"=",
"self",
".",
"_build_url",
"(",
"\"buildconfigs/\"",
")",
"return",
"self",
".",
"_post",
"(",
"url",
",",
"data",
"=",
"build_config_json",
",",
"headers",
"=",
"{",
"\"Content-Type\"",
":",
"\"application/json\"",
"}",
")"
] |
:return:
|
[
":",
"return",
":"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L381-L387
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.stream_logs
|
def stream_logs(self, build_id):
"""
stream logs from build
:param build_id: str
:return: iterator
"""
kwargs = {'follow': 1}
# If connection is closed within this many seconds, give up:
min_idle_timeout = 60
# Stream logs, but be careful of the connection closing
# due to idle timeout. In that case, try again until the
# call returns more quickly than a reasonable timeout
# would be set to.
last_activity = time.time()
while True:
buildlogs_url = self._build_url("builds/%s/log/" % build_id,
**kwargs)
try:
response = self._get(buildlogs_url, stream=1,
headers={'Connection': 'close'})
check_response(response)
for line in response.iter_lines():
last_activity = time.time()
yield line
# NOTE1: If self._get causes ChunkedEncodingError, ConnectionError,
# or IncompleteRead to be raised, they'll be wrapped in
# OsbsNetworkException or OsbsException
# NOTE2: If iter_lines causes ChunkedEncodingError
# or IncompleteRead to be raised, it'll simply be silenced.
# NOTE3: An exception may be raised from
# check_response(). In this case, exception will be
# wrapped in OsbsException or OsbsNetworkException,
# inspect cause to detect ConnectionError.
except OsbsException as exc:
if not isinstance(exc.cause, ConnectionError):
raise
idle = time.time() - last_activity
logger.debug("connection closed after %ds", idle)
if idle < min_idle_timeout:
# Finish output
return
since = int(idle - 1)
logger.debug("fetching logs starting from %ds ago", since)
kwargs['sinceSeconds'] = since
|
python
|
def stream_logs(self, build_id):
"""
stream logs from build
:param build_id: str
:return: iterator
"""
kwargs = {'follow': 1}
# If connection is closed within this many seconds, give up:
min_idle_timeout = 60
# Stream logs, but be careful of the connection closing
# due to idle timeout. In that case, try again until the
# call returns more quickly than a reasonable timeout
# would be set to.
last_activity = time.time()
while True:
buildlogs_url = self._build_url("builds/%s/log/" % build_id,
**kwargs)
try:
response = self._get(buildlogs_url, stream=1,
headers={'Connection': 'close'})
check_response(response)
for line in response.iter_lines():
last_activity = time.time()
yield line
# NOTE1: If self._get causes ChunkedEncodingError, ConnectionError,
# or IncompleteRead to be raised, they'll be wrapped in
# OsbsNetworkException or OsbsException
# NOTE2: If iter_lines causes ChunkedEncodingError
# or IncompleteRead to be raised, it'll simply be silenced.
# NOTE3: An exception may be raised from
# check_response(). In this case, exception will be
# wrapped in OsbsException or OsbsNetworkException,
# inspect cause to detect ConnectionError.
except OsbsException as exc:
if not isinstance(exc.cause, ConnectionError):
raise
idle = time.time() - last_activity
logger.debug("connection closed after %ds", idle)
if idle < min_idle_timeout:
# Finish output
return
since = int(idle - 1)
logger.debug("fetching logs starting from %ds ago", since)
kwargs['sinceSeconds'] = since
|
[
"def",
"stream_logs",
"(",
"self",
",",
"build_id",
")",
":",
"kwargs",
"=",
"{",
"'follow'",
":",
"1",
"}",
"# If connection is closed within this many seconds, give up:",
"min_idle_timeout",
"=",
"60",
"# Stream logs, but be careful of the connection closing",
"# due to idle timeout. In that case, try again until the",
"# call returns more quickly than a reasonable timeout",
"# would be set to.",
"last_activity",
"=",
"time",
".",
"time",
"(",
")",
"while",
"True",
":",
"buildlogs_url",
"=",
"self",
".",
"_build_url",
"(",
"\"builds/%s/log/\"",
"%",
"build_id",
",",
"*",
"*",
"kwargs",
")",
"try",
":",
"response",
"=",
"self",
".",
"_get",
"(",
"buildlogs_url",
",",
"stream",
"=",
"1",
",",
"headers",
"=",
"{",
"'Connection'",
":",
"'close'",
"}",
")",
"check_response",
"(",
"response",
")",
"for",
"line",
"in",
"response",
".",
"iter_lines",
"(",
")",
":",
"last_activity",
"=",
"time",
".",
"time",
"(",
")",
"yield",
"line",
"# NOTE1: If self._get causes ChunkedEncodingError, ConnectionError,",
"# or IncompleteRead to be raised, they'll be wrapped in",
"# OsbsNetworkException or OsbsException",
"# NOTE2: If iter_lines causes ChunkedEncodingError",
"# or IncompleteRead to be raised, it'll simply be silenced.",
"# NOTE3: An exception may be raised from",
"# check_response(). In this case, exception will be",
"# wrapped in OsbsException or OsbsNetworkException,",
"# inspect cause to detect ConnectionError.",
"except",
"OsbsException",
"as",
"exc",
":",
"if",
"not",
"isinstance",
"(",
"exc",
".",
"cause",
",",
"ConnectionError",
")",
":",
"raise",
"idle",
"=",
"time",
".",
"time",
"(",
")",
"-",
"last_activity",
"logger",
".",
"debug",
"(",
"\"connection closed after %ds\"",
",",
"idle",
")",
"if",
"idle",
"<",
"min_idle_timeout",
":",
"# Finish output",
"return",
"since",
"=",
"int",
"(",
"idle",
"-",
"1",
")",
"logger",
".",
"debug",
"(",
"\"fetching logs starting from %ds ago\"",
",",
"since",
")",
"kwargs",
"[",
"'sinceSeconds'",
"]",
"=",
"since"
] |
stream logs from build
:param build_id: str
:return: iterator
|
[
"stream",
"logs",
"from",
"build"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L434-L483
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.logs
|
def logs(self, build_id, follow=False, build_json=None, wait_if_missing=False):
"""
provide logs from build
:param build_id: str
:param follow: bool, fetch logs as they come?
:param build_json: dict, to save one get-build query
:param wait_if_missing: bool, if build doesn't exist, wait
:return: None, str or iterator
"""
# does build exist?
try:
build_json = build_json or self.get_build(build_id).json()
except OsbsResponseException as ex:
if ex.status_code == 404:
if not wait_if_missing:
raise OsbsException("Build '%s' doesn't exist." % build_id)
else:
raise
if follow or wait_if_missing:
build_json = self.wait_for_build_to_get_scheduled(build_id)
br = BuildResponse(build_json)
# When build is in new or pending state, openshift responds with 500
if br.is_pending():
return
if follow:
return self.stream_logs(build_id)
buildlogs_url = self._build_url("builds/%s/log/" % build_id)
response = self._get(buildlogs_url, headers={'Connection': 'close'})
check_response(response)
return response.content
|
python
|
def logs(self, build_id, follow=False, build_json=None, wait_if_missing=False):
"""
provide logs from build
:param build_id: str
:param follow: bool, fetch logs as they come?
:param build_json: dict, to save one get-build query
:param wait_if_missing: bool, if build doesn't exist, wait
:return: None, str or iterator
"""
# does build exist?
try:
build_json = build_json or self.get_build(build_id).json()
except OsbsResponseException as ex:
if ex.status_code == 404:
if not wait_if_missing:
raise OsbsException("Build '%s' doesn't exist." % build_id)
else:
raise
if follow or wait_if_missing:
build_json = self.wait_for_build_to_get_scheduled(build_id)
br = BuildResponse(build_json)
# When build is in new or pending state, openshift responds with 500
if br.is_pending():
return
if follow:
return self.stream_logs(build_id)
buildlogs_url = self._build_url("builds/%s/log/" % build_id)
response = self._get(buildlogs_url, headers={'Connection': 'close'})
check_response(response)
return response.content
|
[
"def",
"logs",
"(",
"self",
",",
"build_id",
",",
"follow",
"=",
"False",
",",
"build_json",
"=",
"None",
",",
"wait_if_missing",
"=",
"False",
")",
":",
"# does build exist?",
"try",
":",
"build_json",
"=",
"build_json",
"or",
"self",
".",
"get_build",
"(",
"build_id",
")",
".",
"json",
"(",
")",
"except",
"OsbsResponseException",
"as",
"ex",
":",
"if",
"ex",
".",
"status_code",
"==",
"404",
":",
"if",
"not",
"wait_if_missing",
":",
"raise",
"OsbsException",
"(",
"\"Build '%s' doesn't exist.\"",
"%",
"build_id",
")",
"else",
":",
"raise",
"if",
"follow",
"or",
"wait_if_missing",
":",
"build_json",
"=",
"self",
".",
"wait_for_build_to_get_scheduled",
"(",
"build_id",
")",
"br",
"=",
"BuildResponse",
"(",
"build_json",
")",
"# When build is in new or pending state, openshift responds with 500",
"if",
"br",
".",
"is_pending",
"(",
")",
":",
"return",
"if",
"follow",
":",
"return",
"self",
".",
"stream_logs",
"(",
"build_id",
")",
"buildlogs_url",
"=",
"self",
".",
"_build_url",
"(",
"\"builds/%s/log/\"",
"%",
"build_id",
")",
"response",
"=",
"self",
".",
"_get",
"(",
"buildlogs_url",
",",
"headers",
"=",
"{",
"'Connection'",
":",
"'close'",
"}",
")",
"check_response",
"(",
"response",
")",
"return",
"response",
".",
"content"
] |
provide logs from build
:param build_id: str
:param follow: bool, fetch logs as they come?
:param build_json: dict, to save one get-build query
:param wait_if_missing: bool, if build doesn't exist, wait
:return: None, str or iterator
|
[
"provide",
"logs",
"from",
"build"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L485-L520
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.list_builds
|
def list_builds(self, build_config_id=None, koji_task_id=None,
field_selector=None, labels=None):
"""
List builds matching criteria
:param build_config_id: str, only list builds created from BuildConfig
:param koji_task_id: str, only list builds for Koji Task ID
:param field_selector: str, field selector for query
:return: HttpResponse
"""
query = {}
selector = '{key}={value}'
label = {}
if labels is not None:
label.update(labels)
if build_config_id is not None:
label['buildconfig'] = build_config_id
if koji_task_id is not None:
label['koji-task-id'] = str(koji_task_id)
if label:
query['labelSelector'] = ','.join([selector.format(key=key,
value=value)
for key, value in label.items()])
if field_selector is not None:
query['fieldSelector'] = field_selector
url = self._build_url("builds/", **query)
return self._get(url)
|
python
|
def list_builds(self, build_config_id=None, koji_task_id=None,
field_selector=None, labels=None):
"""
List builds matching criteria
:param build_config_id: str, only list builds created from BuildConfig
:param koji_task_id: str, only list builds for Koji Task ID
:param field_selector: str, field selector for query
:return: HttpResponse
"""
query = {}
selector = '{key}={value}'
label = {}
if labels is not None:
label.update(labels)
if build_config_id is not None:
label['buildconfig'] = build_config_id
if koji_task_id is not None:
label['koji-task-id'] = str(koji_task_id)
if label:
query['labelSelector'] = ','.join([selector.format(key=key,
value=value)
for key, value in label.items()])
if field_selector is not None:
query['fieldSelector'] = field_selector
url = self._build_url("builds/", **query)
return self._get(url)
|
[
"def",
"list_builds",
"(",
"self",
",",
"build_config_id",
"=",
"None",
",",
"koji_task_id",
"=",
"None",
",",
"field_selector",
"=",
"None",
",",
"labels",
"=",
"None",
")",
":",
"query",
"=",
"{",
"}",
"selector",
"=",
"'{key}={value}'",
"label",
"=",
"{",
"}",
"if",
"labels",
"is",
"not",
"None",
":",
"label",
".",
"update",
"(",
"labels",
")",
"if",
"build_config_id",
"is",
"not",
"None",
":",
"label",
"[",
"'buildconfig'",
"]",
"=",
"build_config_id",
"if",
"koji_task_id",
"is",
"not",
"None",
":",
"label",
"[",
"'koji-task-id'",
"]",
"=",
"str",
"(",
"koji_task_id",
")",
"if",
"label",
":",
"query",
"[",
"'labelSelector'",
"]",
"=",
"','",
".",
"join",
"(",
"[",
"selector",
".",
"format",
"(",
"key",
"=",
"key",
",",
"value",
"=",
"value",
")",
"for",
"key",
",",
"value",
"in",
"label",
".",
"items",
"(",
")",
"]",
")",
"if",
"field_selector",
"is",
"not",
"None",
":",
"query",
"[",
"'fieldSelector'",
"]",
"=",
"field_selector",
"url",
"=",
"self",
".",
"_build_url",
"(",
"\"builds/\"",
",",
"*",
"*",
"query",
")",
"return",
"self",
".",
"_get",
"(",
"url",
")"
] |
List builds matching criteria
:param build_config_id: str, only list builds created from BuildConfig
:param koji_task_id: str, only list builds for Koji Task ID
:param field_selector: str, field selector for query
:return: HttpResponse
|
[
"List",
"builds",
"matching",
"criteria"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L522-L553
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.create_resource_quota
|
def create_resource_quota(self, name, quota_json):
"""
Prevent builds being scheduled and wait for running builds to finish.
:return:
"""
url = self._build_k8s_url("resourcequotas/")
response = self._post(url, data=json.dumps(quota_json),
headers={"Content-Type": "application/json"})
if response.status_code == http_client.CONFLICT:
url = self._build_k8s_url("resourcequotas/%s" % name)
response = self._put(url, data=json.dumps(quota_json),
headers={"Content-Type": "application/json"})
check_response(response)
return response
|
python
|
def create_resource_quota(self, name, quota_json):
"""
Prevent builds being scheduled and wait for running builds to finish.
:return:
"""
url = self._build_k8s_url("resourcequotas/")
response = self._post(url, data=json.dumps(quota_json),
headers={"Content-Type": "application/json"})
if response.status_code == http_client.CONFLICT:
url = self._build_k8s_url("resourcequotas/%s" % name)
response = self._put(url, data=json.dumps(quota_json),
headers={"Content-Type": "application/json"})
check_response(response)
return response
|
[
"def",
"create_resource_quota",
"(",
"self",
",",
"name",
",",
"quota_json",
")",
":",
"url",
"=",
"self",
".",
"_build_k8s_url",
"(",
"\"resourcequotas/\"",
")",
"response",
"=",
"self",
".",
"_post",
"(",
"url",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"quota_json",
")",
",",
"headers",
"=",
"{",
"\"Content-Type\"",
":",
"\"application/json\"",
"}",
")",
"if",
"response",
".",
"status_code",
"==",
"http_client",
".",
"CONFLICT",
":",
"url",
"=",
"self",
".",
"_build_k8s_url",
"(",
"\"resourcequotas/%s\"",
"%",
"name",
")",
"response",
"=",
"self",
".",
"_put",
"(",
"url",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"quota_json",
")",
",",
"headers",
"=",
"{",
"\"Content-Type\"",
":",
"\"application/json\"",
"}",
")",
"check_response",
"(",
"response",
")",
"return",
"response"
] |
Prevent builds being scheduled and wait for running builds to finish.
:return:
|
[
"Prevent",
"builds",
"being",
"scheduled",
"and",
"wait",
"for",
"running",
"builds",
"to",
"finish",
"."
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L577-L593
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.wait
|
def wait(self, build_id, states):
"""
:param build_id: wait for build to finish
:return:
"""
logger.info("watching build '%s'", build_id)
for changetype, obj in self.watch_resource("builds", build_id):
try:
obj_name = obj["metadata"]["name"]
except KeyError:
logger.error("'object' doesn't have any name")
continue
try:
obj_status = obj["status"]["phase"]
except KeyError:
logger.error("'object' doesn't have any status")
continue
else:
obj_status_lower = obj_status.lower()
logger.info("object has changed: '%s', status: '%s', name: '%s'",
changetype, obj_status, obj_name)
if obj_name == build_id:
logger.info("matching build found")
logger.debug("is %s in %s?", repr(obj_status_lower), states)
if obj_status_lower in states:
logger.debug("Yes, build is in the state I'm waiting for.")
return obj
else:
logger.debug("No, build is not in the state I'm "
"waiting for.")
else:
logger.info("The build %r isn't me %r", obj_name, build_id)
# I'm not sure how we can end up here since there are two possible scenarios:
# 1. our object was found and we are returning in the loop
# 2. our object was not found and we keep waiting (in the loop)
# Therefore, let's raise here
logger.warning("build '%s' was not found during wait", build_id)
raise OsbsWatchBuildNotFound("build '%s' was not found and response stream ended" %
build_id)
|
python
|
def wait(self, build_id, states):
"""
:param build_id: wait for build to finish
:return:
"""
logger.info("watching build '%s'", build_id)
for changetype, obj in self.watch_resource("builds", build_id):
try:
obj_name = obj["metadata"]["name"]
except KeyError:
logger.error("'object' doesn't have any name")
continue
try:
obj_status = obj["status"]["phase"]
except KeyError:
logger.error("'object' doesn't have any status")
continue
else:
obj_status_lower = obj_status.lower()
logger.info("object has changed: '%s', status: '%s', name: '%s'",
changetype, obj_status, obj_name)
if obj_name == build_id:
logger.info("matching build found")
logger.debug("is %s in %s?", repr(obj_status_lower), states)
if obj_status_lower in states:
logger.debug("Yes, build is in the state I'm waiting for.")
return obj
else:
logger.debug("No, build is not in the state I'm "
"waiting for.")
else:
logger.info("The build %r isn't me %r", obj_name, build_id)
# I'm not sure how we can end up here since there are two possible scenarios:
# 1. our object was found and we are returning in the loop
# 2. our object was not found and we keep waiting (in the loop)
# Therefore, let's raise here
logger.warning("build '%s' was not found during wait", build_id)
raise OsbsWatchBuildNotFound("build '%s' was not found and response stream ended" %
build_id)
|
[
"def",
"wait",
"(",
"self",
",",
"build_id",
",",
"states",
")",
":",
"logger",
".",
"info",
"(",
"\"watching build '%s'\"",
",",
"build_id",
")",
"for",
"changetype",
",",
"obj",
"in",
"self",
".",
"watch_resource",
"(",
"\"builds\"",
",",
"build_id",
")",
":",
"try",
":",
"obj_name",
"=",
"obj",
"[",
"\"metadata\"",
"]",
"[",
"\"name\"",
"]",
"except",
"KeyError",
":",
"logger",
".",
"error",
"(",
"\"'object' doesn't have any name\"",
")",
"continue",
"try",
":",
"obj_status",
"=",
"obj",
"[",
"\"status\"",
"]",
"[",
"\"phase\"",
"]",
"except",
"KeyError",
":",
"logger",
".",
"error",
"(",
"\"'object' doesn't have any status\"",
")",
"continue",
"else",
":",
"obj_status_lower",
"=",
"obj_status",
".",
"lower",
"(",
")",
"logger",
".",
"info",
"(",
"\"object has changed: '%s', status: '%s', name: '%s'\"",
",",
"changetype",
",",
"obj_status",
",",
"obj_name",
")",
"if",
"obj_name",
"==",
"build_id",
":",
"logger",
".",
"info",
"(",
"\"matching build found\"",
")",
"logger",
".",
"debug",
"(",
"\"is %s in %s?\"",
",",
"repr",
"(",
"obj_status_lower",
")",
",",
"states",
")",
"if",
"obj_status_lower",
"in",
"states",
":",
"logger",
".",
"debug",
"(",
"\"Yes, build is in the state I'm waiting for.\"",
")",
"return",
"obj",
"else",
":",
"logger",
".",
"debug",
"(",
"\"No, build is not in the state I'm \"",
"\"waiting for.\"",
")",
"else",
":",
"logger",
".",
"info",
"(",
"\"The build %r isn't me %r\"",
",",
"obj_name",
",",
"build_id",
")",
"# I'm not sure how we can end up here since there are two possible scenarios:",
"# 1. our object was found and we are returning in the loop",
"# 2. our object was not found and we keep waiting (in the loop)",
"# Therefore, let's raise here",
"logger",
".",
"warning",
"(",
"\"build '%s' was not found during wait\"",
",",
"build_id",
")",
"raise",
"OsbsWatchBuildNotFound",
"(",
"\"build '%s' was not found and response stream ended\"",
"%",
"build_id",
")"
] |
:param build_id: wait for build to finish
:return:
|
[
":",
"param",
"build_id",
":",
"wait",
"for",
"build",
"to",
"finish"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L654-L694
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.adjust_attributes_on_object
|
def adjust_attributes_on_object(self, collection, name, things, values, how):
"""
adjust labels or annotations on object
labels have to match RE: (([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])? and
have at most 63 chars
:param collection: str, object collection e.g. 'builds'
:param name: str, name of object
:param things: str, 'labels' or 'annotations'
:param values: dict, values to set
:param how: callable, how to adjust the values e.g.
self._replace_metadata_things
:return:
"""
url = self._build_url("%s/%s" % (collection, name))
response = self._get(url)
logger.debug("before modification: %s", response.content)
build_json = response.json()
how(build_json['metadata'], things, values)
response = self._put(url, data=json.dumps(build_json), use_json=True)
check_response(response)
return response
|
python
|
def adjust_attributes_on_object(self, collection, name, things, values, how):
"""
adjust labels or annotations on object
labels have to match RE: (([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])? and
have at most 63 chars
:param collection: str, object collection e.g. 'builds'
:param name: str, name of object
:param things: str, 'labels' or 'annotations'
:param values: dict, values to set
:param how: callable, how to adjust the values e.g.
self._replace_metadata_things
:return:
"""
url = self._build_url("%s/%s" % (collection, name))
response = self._get(url)
logger.debug("before modification: %s", response.content)
build_json = response.json()
how(build_json['metadata'], things, values)
response = self._put(url, data=json.dumps(build_json), use_json=True)
check_response(response)
return response
|
[
"def",
"adjust_attributes_on_object",
"(",
"self",
",",
"collection",
",",
"name",
",",
"things",
",",
"values",
",",
"how",
")",
":",
"url",
"=",
"self",
".",
"_build_url",
"(",
"\"%s/%s\"",
"%",
"(",
"collection",
",",
"name",
")",
")",
"response",
"=",
"self",
".",
"_get",
"(",
"url",
")",
"logger",
".",
"debug",
"(",
"\"before modification: %s\"",
",",
"response",
".",
"content",
")",
"build_json",
"=",
"response",
".",
"json",
"(",
")",
"how",
"(",
"build_json",
"[",
"'metadata'",
"]",
",",
"things",
",",
"values",
")",
"response",
"=",
"self",
".",
"_put",
"(",
"url",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"build_json",
")",
",",
"use_json",
"=",
"True",
")",
"check_response",
"(",
"response",
")",
"return",
"response"
] |
adjust labels or annotations on object
labels have to match RE: (([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])? and
have at most 63 chars
:param collection: str, object collection e.g. 'builds'
:param name: str, name of object
:param things: str, 'labels' or 'annotations'
:param values: dict, values to set
:param how: callable, how to adjust the values e.g.
self._replace_metadata_things
:return:
|
[
"adjust",
"labels",
"or",
"annotations",
"on",
"object"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L727-L749
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.update_annotations_on_build
|
def update_annotations_on_build(self, build_id, annotations):
"""
set annotations on build object
:param build_id: str, id of build
:param annotations: dict, annotations to set
:return:
"""
return self.adjust_attributes_on_object('builds', build_id,
'annotations', annotations,
self._update_metadata_things)
|
python
|
def update_annotations_on_build(self, build_id, annotations):
"""
set annotations on build object
:param build_id: str, id of build
:param annotations: dict, annotations to set
:return:
"""
return self.adjust_attributes_on_object('builds', build_id,
'annotations', annotations,
self._update_metadata_things)
|
[
"def",
"update_annotations_on_build",
"(",
"self",
",",
"build_id",
",",
"annotations",
")",
":",
"return",
"self",
".",
"adjust_attributes_on_object",
"(",
"'builds'",
",",
"build_id",
",",
"'annotations'",
",",
"annotations",
",",
"self",
".",
"_update_metadata_things",
")"
] |
set annotations on build object
:param build_id: str, id of build
:param annotations: dict, annotations to set
:return:
|
[
"set",
"annotations",
"on",
"build",
"object"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L771-L781
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.import_image
|
def import_image(self, name, stream_import, tags=None):
"""
Import image tags from a Docker registry into an ImageStream
:return: bool, whether tags were imported
"""
# Get the JSON for the ImageStream
imagestream_json = self.get_image_stream(name).json()
logger.debug("imagestream: %r", imagestream_json)
if 'dockerImageRepository' in imagestream_json.get('spec', {}):
logger.debug("Removing 'dockerImageRepository' from ImageStream %s", name)
source_repo = imagestream_json['spec'].pop('dockerImageRepository')
imagestream_json['metadata']['annotations'][ANNOTATION_SOURCE_REPO] = source_repo
imagestream_json = self.update_image_stream(name, imagestream_json).json()
# Note the tags before import
oldtags = imagestream_json.get('status', {}).get('tags', [])
logger.debug("tags before import: %r", oldtags)
stream_import['metadata']['name'] = name
stream_import['spec']['images'] = []
tags_set = set(tags) if tags else set()
for tag in imagestream_json.get('spec', {}).get('tags', []):
if tags_set and tag['name'] not in tags_set:
continue
image_import = {
'from': tag['from'],
'to': {'name': tag['name']},
'importPolicy': tag.get('importPolicy'),
'referencePolicy': tag.get('referencePolicy'),
}
stream_import['spec']['images'].append(image_import)
if not stream_import['spec']['images']:
logger.debug('No tags to import')
return False
import_url = self._build_url("imagestreamimports/")
import_response = self._post(import_url, data=json.dumps(stream_import),
use_json=True)
self._check_import_image_response(import_response)
new_tags = [
image['tag']
for image in import_response.json().get('status', {}).get('images', [])]
logger.debug("tags after import: %r", new_tags)
return True
|
python
|
def import_image(self, name, stream_import, tags=None):
"""
Import image tags from a Docker registry into an ImageStream
:return: bool, whether tags were imported
"""
# Get the JSON for the ImageStream
imagestream_json = self.get_image_stream(name).json()
logger.debug("imagestream: %r", imagestream_json)
if 'dockerImageRepository' in imagestream_json.get('spec', {}):
logger.debug("Removing 'dockerImageRepository' from ImageStream %s", name)
source_repo = imagestream_json['spec'].pop('dockerImageRepository')
imagestream_json['metadata']['annotations'][ANNOTATION_SOURCE_REPO] = source_repo
imagestream_json = self.update_image_stream(name, imagestream_json).json()
# Note the tags before import
oldtags = imagestream_json.get('status', {}).get('tags', [])
logger.debug("tags before import: %r", oldtags)
stream_import['metadata']['name'] = name
stream_import['spec']['images'] = []
tags_set = set(tags) if tags else set()
for tag in imagestream_json.get('spec', {}).get('tags', []):
if tags_set and tag['name'] not in tags_set:
continue
image_import = {
'from': tag['from'],
'to': {'name': tag['name']},
'importPolicy': tag.get('importPolicy'),
'referencePolicy': tag.get('referencePolicy'),
}
stream_import['spec']['images'].append(image_import)
if not stream_import['spec']['images']:
logger.debug('No tags to import')
return False
import_url = self._build_url("imagestreamimports/")
import_response = self._post(import_url, data=json.dumps(stream_import),
use_json=True)
self._check_import_image_response(import_response)
new_tags = [
image['tag']
for image in import_response.json().get('status', {}).get('images', [])]
logger.debug("tags after import: %r", new_tags)
return True
|
[
"def",
"import_image",
"(",
"self",
",",
"name",
",",
"stream_import",
",",
"tags",
"=",
"None",
")",
":",
"# Get the JSON for the ImageStream",
"imagestream_json",
"=",
"self",
".",
"get_image_stream",
"(",
"name",
")",
".",
"json",
"(",
")",
"logger",
".",
"debug",
"(",
"\"imagestream: %r\"",
",",
"imagestream_json",
")",
"if",
"'dockerImageRepository'",
"in",
"imagestream_json",
".",
"get",
"(",
"'spec'",
",",
"{",
"}",
")",
":",
"logger",
".",
"debug",
"(",
"\"Removing 'dockerImageRepository' from ImageStream %s\"",
",",
"name",
")",
"source_repo",
"=",
"imagestream_json",
"[",
"'spec'",
"]",
".",
"pop",
"(",
"'dockerImageRepository'",
")",
"imagestream_json",
"[",
"'metadata'",
"]",
"[",
"'annotations'",
"]",
"[",
"ANNOTATION_SOURCE_REPO",
"]",
"=",
"source_repo",
"imagestream_json",
"=",
"self",
".",
"update_image_stream",
"(",
"name",
",",
"imagestream_json",
")",
".",
"json",
"(",
")",
"# Note the tags before import",
"oldtags",
"=",
"imagestream_json",
".",
"get",
"(",
"'status'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'tags'",
",",
"[",
"]",
")",
"logger",
".",
"debug",
"(",
"\"tags before import: %r\"",
",",
"oldtags",
")",
"stream_import",
"[",
"'metadata'",
"]",
"[",
"'name'",
"]",
"=",
"name",
"stream_import",
"[",
"'spec'",
"]",
"[",
"'images'",
"]",
"=",
"[",
"]",
"tags_set",
"=",
"set",
"(",
"tags",
")",
"if",
"tags",
"else",
"set",
"(",
")",
"for",
"tag",
"in",
"imagestream_json",
".",
"get",
"(",
"'spec'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'tags'",
",",
"[",
"]",
")",
":",
"if",
"tags_set",
"and",
"tag",
"[",
"'name'",
"]",
"not",
"in",
"tags_set",
":",
"continue",
"image_import",
"=",
"{",
"'from'",
":",
"tag",
"[",
"'from'",
"]",
",",
"'to'",
":",
"{",
"'name'",
":",
"tag",
"[",
"'name'",
"]",
"}",
",",
"'importPolicy'",
":",
"tag",
".",
"get",
"(",
"'importPolicy'",
")",
",",
"'referencePolicy'",
":",
"tag",
".",
"get",
"(",
"'referencePolicy'",
")",
",",
"}",
"stream_import",
"[",
"'spec'",
"]",
"[",
"'images'",
"]",
".",
"append",
"(",
"image_import",
")",
"if",
"not",
"stream_import",
"[",
"'spec'",
"]",
"[",
"'images'",
"]",
":",
"logger",
".",
"debug",
"(",
"'No tags to import'",
")",
"return",
"False",
"import_url",
"=",
"self",
".",
"_build_url",
"(",
"\"imagestreamimports/\"",
")",
"import_response",
"=",
"self",
".",
"_post",
"(",
"import_url",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"stream_import",
")",
",",
"use_json",
"=",
"True",
")",
"self",
".",
"_check_import_image_response",
"(",
"import_response",
")",
"new_tags",
"=",
"[",
"image",
"[",
"'tag'",
"]",
"for",
"image",
"in",
"import_response",
".",
"json",
"(",
")",
".",
"get",
"(",
"'status'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'images'",
",",
"[",
"]",
")",
"]",
"logger",
".",
"debug",
"(",
"\"tags after import: %r\"",
",",
"new_tags",
")",
"return",
"True"
] |
Import image tags from a Docker registry into an ImageStream
:return: bool, whether tags were imported
|
[
"Import",
"image",
"tags",
"from",
"a",
"Docker",
"registry",
"into",
"an",
"ImageStream"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L872-L922
|
train
|
projectatomic/osbs-client
|
osbs/core.py
|
Openshift.import_image_tags
|
def import_image_tags(self, name, stream_import, tags, repository, insecure):
"""
Import image tags from a Docker registry into an ImageStream
:return: bool, whether tags were imported
"""
# Get the JSON for the ImageStream
imagestream_json = self.get_image_stream(name).json()
logger.debug("imagestream: %r", imagestream_json)
changed = False
# existence of dockerImageRepository is limiting how many tags are updated
if 'dockerImageRepository' in imagestream_json.get('spec', {}):
logger.debug("Removing 'dockerImageRepository' from ImageStream %s", name)
imagestream_json['spec'].pop('dockerImageRepository')
changed = True
all_annotations = imagestream_json.get('metadata', {}).get('annotations', {})
# remove annotations about registry, since method will get them as arguments
for annotation in ANNOTATION_SOURCE_REPO, ANNOTATION_INSECURE_REPO:
if annotation in all_annotations:
imagestream_json['metadata']['annotations'].pop(annotation)
changed = True
if changed:
imagestream_json = self.update_image_stream(name, imagestream_json).json()
# Note the tags before import
oldtags = imagestream_json.get('status', {}).get('tags', [])
logger.debug("tags before import: %r", oldtags)
stream_import['metadata']['name'] = name
stream_import['spec']['images'] = []
tags_set = set(tags) if tags else set()
if not tags_set:
logger.debug('No tags to import')
return False
for tag in tags_set:
image_import = {
'from': {"kind": "DockerImage",
"name": '{}:{}'.format(repository, tag)},
'to': {'name': tag},
'importPolicy': {'insecure': insecure},
# referencePolicy will default to "type: source"
# so we don't have to explicitly set it
}
stream_import['spec']['images'].append(image_import)
import_url = self._build_url("imagestreamimports/")
import_response = self._post(import_url, data=json.dumps(stream_import),
use_json=True)
self._check_import_image_response(import_response)
new_tags = [
image['tag']
for image in import_response.json().get('status', {}).get('images', [])]
logger.debug("tags after import: %r", new_tags)
return True
|
python
|
def import_image_tags(self, name, stream_import, tags, repository, insecure):
"""
Import image tags from a Docker registry into an ImageStream
:return: bool, whether tags were imported
"""
# Get the JSON for the ImageStream
imagestream_json = self.get_image_stream(name).json()
logger.debug("imagestream: %r", imagestream_json)
changed = False
# existence of dockerImageRepository is limiting how many tags are updated
if 'dockerImageRepository' in imagestream_json.get('spec', {}):
logger.debug("Removing 'dockerImageRepository' from ImageStream %s", name)
imagestream_json['spec'].pop('dockerImageRepository')
changed = True
all_annotations = imagestream_json.get('metadata', {}).get('annotations', {})
# remove annotations about registry, since method will get them as arguments
for annotation in ANNOTATION_SOURCE_REPO, ANNOTATION_INSECURE_REPO:
if annotation in all_annotations:
imagestream_json['metadata']['annotations'].pop(annotation)
changed = True
if changed:
imagestream_json = self.update_image_stream(name, imagestream_json).json()
# Note the tags before import
oldtags = imagestream_json.get('status', {}).get('tags', [])
logger.debug("tags before import: %r", oldtags)
stream_import['metadata']['name'] = name
stream_import['spec']['images'] = []
tags_set = set(tags) if tags else set()
if not tags_set:
logger.debug('No tags to import')
return False
for tag in tags_set:
image_import = {
'from': {"kind": "DockerImage",
"name": '{}:{}'.format(repository, tag)},
'to': {'name': tag},
'importPolicy': {'insecure': insecure},
# referencePolicy will default to "type: source"
# so we don't have to explicitly set it
}
stream_import['spec']['images'].append(image_import)
import_url = self._build_url("imagestreamimports/")
import_response = self._post(import_url, data=json.dumps(stream_import),
use_json=True)
self._check_import_image_response(import_response)
new_tags = [
image['tag']
for image in import_response.json().get('status', {}).get('images', [])]
logger.debug("tags after import: %r", new_tags)
return True
|
[
"def",
"import_image_tags",
"(",
"self",
",",
"name",
",",
"stream_import",
",",
"tags",
",",
"repository",
",",
"insecure",
")",
":",
"# Get the JSON for the ImageStream",
"imagestream_json",
"=",
"self",
".",
"get_image_stream",
"(",
"name",
")",
".",
"json",
"(",
")",
"logger",
".",
"debug",
"(",
"\"imagestream: %r\"",
",",
"imagestream_json",
")",
"changed",
"=",
"False",
"# existence of dockerImageRepository is limiting how many tags are updated",
"if",
"'dockerImageRepository'",
"in",
"imagestream_json",
".",
"get",
"(",
"'spec'",
",",
"{",
"}",
")",
":",
"logger",
".",
"debug",
"(",
"\"Removing 'dockerImageRepository' from ImageStream %s\"",
",",
"name",
")",
"imagestream_json",
"[",
"'spec'",
"]",
".",
"pop",
"(",
"'dockerImageRepository'",
")",
"changed",
"=",
"True",
"all_annotations",
"=",
"imagestream_json",
".",
"get",
"(",
"'metadata'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'annotations'",
",",
"{",
"}",
")",
"# remove annotations about registry, since method will get them as arguments",
"for",
"annotation",
"in",
"ANNOTATION_SOURCE_REPO",
",",
"ANNOTATION_INSECURE_REPO",
":",
"if",
"annotation",
"in",
"all_annotations",
":",
"imagestream_json",
"[",
"'metadata'",
"]",
"[",
"'annotations'",
"]",
".",
"pop",
"(",
"annotation",
")",
"changed",
"=",
"True",
"if",
"changed",
":",
"imagestream_json",
"=",
"self",
".",
"update_image_stream",
"(",
"name",
",",
"imagestream_json",
")",
".",
"json",
"(",
")",
"# Note the tags before import",
"oldtags",
"=",
"imagestream_json",
".",
"get",
"(",
"'status'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'tags'",
",",
"[",
"]",
")",
"logger",
".",
"debug",
"(",
"\"tags before import: %r\"",
",",
"oldtags",
")",
"stream_import",
"[",
"'metadata'",
"]",
"[",
"'name'",
"]",
"=",
"name",
"stream_import",
"[",
"'spec'",
"]",
"[",
"'images'",
"]",
"=",
"[",
"]",
"tags_set",
"=",
"set",
"(",
"tags",
")",
"if",
"tags",
"else",
"set",
"(",
")",
"if",
"not",
"tags_set",
":",
"logger",
".",
"debug",
"(",
"'No tags to import'",
")",
"return",
"False",
"for",
"tag",
"in",
"tags_set",
":",
"image_import",
"=",
"{",
"'from'",
":",
"{",
"\"kind\"",
":",
"\"DockerImage\"",
",",
"\"name\"",
":",
"'{}:{}'",
".",
"format",
"(",
"repository",
",",
"tag",
")",
"}",
",",
"'to'",
":",
"{",
"'name'",
":",
"tag",
"}",
",",
"'importPolicy'",
":",
"{",
"'insecure'",
":",
"insecure",
"}",
",",
"# referencePolicy will default to \"type: source\"",
"# so we don't have to explicitly set it",
"}",
"stream_import",
"[",
"'spec'",
"]",
"[",
"'images'",
"]",
".",
"append",
"(",
"image_import",
")",
"import_url",
"=",
"self",
".",
"_build_url",
"(",
"\"imagestreamimports/\"",
")",
"import_response",
"=",
"self",
".",
"_post",
"(",
"import_url",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"stream_import",
")",
",",
"use_json",
"=",
"True",
")",
"self",
".",
"_check_import_image_response",
"(",
"import_response",
")",
"new_tags",
"=",
"[",
"image",
"[",
"'tag'",
"]",
"for",
"image",
"in",
"import_response",
".",
"json",
"(",
")",
".",
"get",
"(",
"'status'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'images'",
",",
"[",
"]",
")",
"]",
"logger",
".",
"debug",
"(",
"\"tags after import: %r\"",
",",
"new_tags",
")",
"return",
"True"
] |
Import image tags from a Docker registry into an ImageStream
:return: bool, whether tags were imported
|
[
"Import",
"image",
"tags",
"from",
"a",
"Docker",
"registry",
"into",
"an",
"ImageStream"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/core.py#L926-L986
|
train
|
projectatomic/osbs-client
|
osbs/cli/main.py
|
str_on_2_unicode_on_3
|
def str_on_2_unicode_on_3(s):
"""
argparse is way too awesome when doing repr() on choices when printing usage
:param s: str or unicode
:return: str on 2, unicode on 3
"""
if not PY3:
return str(s)
else: # 3+
if not isinstance(s, str):
return str(s, encoding="utf-8")
return s
|
python
|
def str_on_2_unicode_on_3(s):
"""
argparse is way too awesome when doing repr() on choices when printing usage
:param s: str or unicode
:return: str on 2, unicode on 3
"""
if not PY3:
return str(s)
else: # 3+
if not isinstance(s, str):
return str(s, encoding="utf-8")
return s
|
[
"def",
"str_on_2_unicode_on_3",
"(",
"s",
")",
":",
"if",
"not",
"PY3",
":",
"return",
"str",
"(",
"s",
")",
"else",
":",
"# 3+",
"if",
"not",
"isinstance",
"(",
"s",
",",
"str",
")",
":",
"return",
"str",
"(",
"s",
",",
"encoding",
"=",
"\"utf-8\"",
")",
"return",
"s"
] |
argparse is way too awesome when doing repr() on choices when printing usage
:param s: str or unicode
:return: str on 2, unicode on 3
|
[
"argparse",
"is",
"way",
"too",
"awesome",
"when",
"doing",
"repr",
"()",
"on",
"choices",
"when",
"printing",
"usage"
] |
571fe035dab3a7c02e1dccd5d65ffd75be750458
|
https://github.com/projectatomic/osbs-client/blob/571fe035dab3a7c02e1dccd5d65ffd75be750458/osbs/cli/main.py#L534-L547
|
train
|
thombashi/pytablereader
|
pytablereader/jsonlines/core.py
|
JsonLinesTableFileLoader.load
|
def load(self):
"""
Extract tabular data as |TableData| instances from a Line-delimited JSON file.
|load_source_desc_file|
:return:
Loaded table data iterator.
|load_table_name_desc|
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the data is invalid Line-delimited JSON.
:raises pytablereader.error.ValidationError:
If the data is not acceptable Line-delimited JSON format.
"""
formatter = JsonLinesTableFormatter(self.load_dict())
formatter.accept(self)
return formatter.to_table_data()
|
python
|
def load(self):
"""
Extract tabular data as |TableData| instances from a Line-delimited JSON file.
|load_source_desc_file|
:return:
Loaded table data iterator.
|load_table_name_desc|
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the data is invalid Line-delimited JSON.
:raises pytablereader.error.ValidationError:
If the data is not acceptable Line-delimited JSON format.
"""
formatter = JsonLinesTableFormatter(self.load_dict())
formatter.accept(self)
return formatter.to_table_data()
|
[
"def",
"load",
"(",
"self",
")",
":",
"formatter",
"=",
"JsonLinesTableFormatter",
"(",
"self",
".",
"load_dict",
"(",
")",
")",
"formatter",
".",
"accept",
"(",
"self",
")",
"return",
"formatter",
".",
"to_table_data",
"(",
")"
] |
Extract tabular data as |TableData| instances from a Line-delimited JSON file.
|load_source_desc_file|
:return:
Loaded table data iterator.
|load_table_name_desc|
:rtype: |TableData| iterator
:raises pytablereader.DataError:
If the data is invalid Line-delimited JSON.
:raises pytablereader.error.ValidationError:
If the data is not acceptable Line-delimited JSON format.
|
[
"Extract",
"tabular",
"data",
"as",
"|TableData|",
"instances",
"from",
"a",
"Line",
"-",
"delimited",
"JSON",
"file",
".",
"|load_source_desc_file|"
] |
bc3c057a2cc775bcce690e0e9019c2907b638101
|
https://github.com/thombashi/pytablereader/blob/bc3c057a2cc775bcce690e0e9019c2907b638101/pytablereader/jsonlines/core.py#L75-L94
|
train
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.