repo
stringlengths 7
55
| path
stringlengths 4
223
| func_name
stringlengths 1
134
| original_string
stringlengths 75
104k
| language
stringclasses 1
value | code
stringlengths 75
104k
| code_tokens
listlengths 19
28.4k
| docstring
stringlengths 1
46.9k
| docstring_tokens
listlengths 1
1.97k
| sha
stringlengths 40
40
| url
stringlengths 87
315
| partition
stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
cimm-kzn/CGRtools
|
CGRtools/reactor.py
|
Reactor.__get_mapping
|
def __get_mapping(self, structures):
"""
match each pattern to each molecule.
if all patterns matches with all molecules
return generator of all possible mapping.
:param structures: disjoint molecules
:return: mapping generator
"""
for c in permutations(structures, len(self.__patterns)):
for m in product(*(x.get_substructure_mapping(y, limit=0) for x, y in zip(self.__patterns, c))):
mapping = {}
for i in m:
mapping.update(i)
if mapping:
yield mapping
|
python
|
def __get_mapping(self, structures):
"""
match each pattern to each molecule.
if all patterns matches with all molecules
return generator of all possible mapping.
:param structures: disjoint molecules
:return: mapping generator
"""
for c in permutations(structures, len(self.__patterns)):
for m in product(*(x.get_substructure_mapping(y, limit=0) for x, y in zip(self.__patterns, c))):
mapping = {}
for i in m:
mapping.update(i)
if mapping:
yield mapping
|
[
"def",
"__get_mapping",
"(",
"self",
",",
"structures",
")",
":",
"for",
"c",
"in",
"permutations",
"(",
"structures",
",",
"len",
"(",
"self",
".",
"__patterns",
")",
")",
":",
"for",
"m",
"in",
"product",
"(",
"*",
"(",
"x",
".",
"get_substructure_mapping",
"(",
"y",
",",
"limit",
"=",
"0",
")",
"for",
"x",
",",
"y",
"in",
"zip",
"(",
"self",
".",
"__patterns",
",",
"c",
")",
")",
")",
":",
"mapping",
"=",
"{",
"}",
"for",
"i",
"in",
"m",
":",
"mapping",
".",
"update",
"(",
"i",
")",
"if",
"mapping",
":",
"yield",
"mapping"
] |
match each pattern to each molecule.
if all patterns matches with all molecules
return generator of all possible mapping.
:param structures: disjoint molecules
:return: mapping generator
|
[
"match",
"each",
"pattern",
"to",
"each",
"molecule",
".",
"if",
"all",
"patterns",
"matches",
"with",
"all",
"molecules",
"return",
"generator",
"of",
"all",
"possible",
"mapping",
"."
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/reactor.py#L300-L315
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
Cache.get
|
def get(self, default=None):
"""
return the cached value or default if it can't be found
:param default: default value
:return: cached value
"""
d = cache.get(self.key)
return ((json.loads(d.decode('utf-8')) if self.serialize else d)
if d is not None
else default)
|
python
|
def get(self, default=None):
"""
return the cached value or default if it can't be found
:param default: default value
:return: cached value
"""
d = cache.get(self.key)
return ((json.loads(d.decode('utf-8')) if self.serialize else d)
if d is not None
else default)
|
[
"def",
"get",
"(",
"self",
",",
"default",
"=",
"None",
")",
":",
"d",
"=",
"cache",
".",
"get",
"(",
"self",
".",
"key",
")",
"return",
"(",
"(",
"json",
".",
"loads",
"(",
"d",
".",
"decode",
"(",
"'utf-8'",
")",
")",
"if",
"self",
".",
"serialize",
"else",
"d",
")",
"if",
"d",
"is",
"not",
"None",
"else",
"default",
")"
] |
return the cached value or default if it can't be found
:param default: default value
:return: cached value
|
[
"return",
"the",
"cached",
"value",
"or",
"default",
"if",
"it",
"can",
"t",
"be",
"found"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L79-L89
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
Cache.set
|
def set(self, val, lifetime=None):
"""
set cache value
:param val: any picklable object
:param lifetime: exprition time in sec
:return: val
"""
cache.set(self.key,
(json.dumps(val) if self.serialize else val),
lifetime or settings.DEFAULT_CACHE_EXPIRE_TIME)
return val
|
python
|
def set(self, val, lifetime=None):
"""
set cache value
:param val: any picklable object
:param lifetime: exprition time in sec
:return: val
"""
cache.set(self.key,
(json.dumps(val) if self.serialize else val),
lifetime or settings.DEFAULT_CACHE_EXPIRE_TIME)
return val
|
[
"def",
"set",
"(",
"self",
",",
"val",
",",
"lifetime",
"=",
"None",
")",
":",
"cache",
".",
"set",
"(",
"self",
".",
"key",
",",
"(",
"json",
".",
"dumps",
"(",
"val",
")",
"if",
"self",
".",
"serialize",
"else",
"val",
")",
",",
"lifetime",
"or",
"settings",
".",
"DEFAULT_CACHE_EXPIRE_TIME",
")",
"return",
"val"
] |
set cache value
:param val: any picklable object
:param lifetime: exprition time in sec
:return: val
|
[
"set",
"cache",
"value"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L91-L102
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
Cache.add
|
def add(self, val):
"""
Add given value to item (list)
Args:
val: A JSON serializable object.
Returns:
Cache backend response.
"""
return cache.lpush(self.key, json.dumps(val) if self.serialize else val)
|
python
|
def add(self, val):
"""
Add given value to item (list)
Args:
val: A JSON serializable object.
Returns:
Cache backend response.
"""
return cache.lpush(self.key, json.dumps(val) if self.serialize else val)
|
[
"def",
"add",
"(",
"self",
",",
"val",
")",
":",
"return",
"cache",
".",
"lpush",
"(",
"self",
".",
"key",
",",
"json",
".",
"dumps",
"(",
"val",
")",
"if",
"self",
".",
"serialize",
"else",
"val",
")"
] |
Add given value to item (list)
Args:
val: A JSON serializable object.
Returns:
Cache backend response.
|
[
"Add",
"given",
"value",
"to",
"item",
"(",
"list",
")"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L143-L153
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
Cache.get_all
|
def get_all(self):
"""
Get all list items.
Returns:
Cache backend response.
"""
result = cache.lrange(self.key, 0, -1)
return (json.loads(item.decode('utf-8')) for item in result if
item) if self.serialize else result
|
python
|
def get_all(self):
"""
Get all list items.
Returns:
Cache backend response.
"""
result = cache.lrange(self.key, 0, -1)
return (json.loads(item.decode('utf-8')) for item in result if
item) if self.serialize else result
|
[
"def",
"get_all",
"(",
"self",
")",
":",
"result",
"=",
"cache",
".",
"lrange",
"(",
"self",
".",
"key",
",",
"0",
",",
"-",
"1",
")",
"return",
"(",
"json",
".",
"loads",
"(",
"item",
".",
"decode",
"(",
"'utf-8'",
")",
")",
"for",
"item",
"in",
"result",
"if",
"item",
")",
"if",
"self",
".",
"serialize",
"else",
"result"
] |
Get all list items.
Returns:
Cache backend response.
|
[
"Get",
"all",
"list",
"items",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L155-L164
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
Cache.remove_item
|
def remove_item(self, val):
"""
Removes given item from the list.
Args:
val: Item
Returns:
Cache backend response.
"""
return cache.lrem(self.key, json.dumps(val))
|
python
|
def remove_item(self, val):
"""
Removes given item from the list.
Args:
val: Item
Returns:
Cache backend response.
"""
return cache.lrem(self.key, json.dumps(val))
|
[
"def",
"remove_item",
"(",
"self",
",",
"val",
")",
":",
"return",
"cache",
".",
"lrem",
"(",
"self",
".",
"key",
",",
"json",
".",
"dumps",
"(",
"val",
")",
")"
] |
Removes given item from the list.
Args:
val: Item
Returns:
Cache backend response.
|
[
"Removes",
"given",
"item",
"from",
"the",
"list",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L175-L185
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
Cache.flush
|
def flush(cls, *args):
"""
Removes all keys of this namespace
Without args, clears all keys starting with cls.PREFIX
if called with args, clears keys starting with given cls.PREFIX + args
Args:
*args: Arbitrary number of arguments.
Returns:
List of removed keys.
"""
return _remove_keys([], [(cls._make_key(args) if args else cls.PREFIX) + '*'])
|
python
|
def flush(cls, *args):
"""
Removes all keys of this namespace
Without args, clears all keys starting with cls.PREFIX
if called with args, clears keys starting with given cls.PREFIX + args
Args:
*args: Arbitrary number of arguments.
Returns:
List of removed keys.
"""
return _remove_keys([], [(cls._make_key(args) if args else cls.PREFIX) + '*'])
|
[
"def",
"flush",
"(",
"cls",
",",
"*",
"args",
")",
":",
"return",
"_remove_keys",
"(",
"[",
"]",
",",
"[",
"(",
"cls",
".",
"_make_key",
"(",
"args",
")",
"if",
"args",
"else",
"cls",
".",
"PREFIX",
")",
"+",
"'*'",
"]",
")"
] |
Removes all keys of this namespace
Without args, clears all keys starting with cls.PREFIX
if called with args, clears keys starting with given cls.PREFIX + args
Args:
*args: Arbitrary number of arguments.
Returns:
List of removed keys.
|
[
"Removes",
"all",
"keys",
"of",
"this",
"namespace",
"Without",
"args",
"clears",
"all",
"keys",
"starting",
"with",
"cls",
".",
"PREFIX",
"if",
"called",
"with",
"args",
"clears",
"keys",
"starting",
"with",
"given",
"cls",
".",
"PREFIX",
"+",
"args"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L188-L200
|
train
|
zetaops/zengine
|
zengine/lib/cache.py
|
KeepAlive.update_or_expire_session
|
def update_or_expire_session(self):
"""
Deletes session if keepalive request expired
otherwise updates the keepalive timestamp value
"""
if not hasattr(self, 'key'):
return
now = time.time()
timestamp = float(self.get() or 0) or now
sess_id = self.sess_id or UserSessionID(self.user_id).get()
if sess_id and now - timestamp > self.SESSION_EXPIRE_TIME:
Session(sess_id).delete()
return False
else:
self.set(now)
return True
|
python
|
def update_or_expire_session(self):
"""
Deletes session if keepalive request expired
otherwise updates the keepalive timestamp value
"""
if not hasattr(self, 'key'):
return
now = time.time()
timestamp = float(self.get() or 0) or now
sess_id = self.sess_id or UserSessionID(self.user_id).get()
if sess_id and now - timestamp > self.SESSION_EXPIRE_TIME:
Session(sess_id).delete()
return False
else:
self.set(now)
return True
|
[
"def",
"update_or_expire_session",
"(",
"self",
")",
":",
"if",
"not",
"hasattr",
"(",
"self",
",",
"'key'",
")",
":",
"return",
"now",
"=",
"time",
".",
"time",
"(",
")",
"timestamp",
"=",
"float",
"(",
"self",
".",
"get",
"(",
")",
"or",
"0",
")",
"or",
"now",
"sess_id",
"=",
"self",
".",
"sess_id",
"or",
"UserSessionID",
"(",
"self",
".",
"user_id",
")",
".",
"get",
"(",
")",
"if",
"sess_id",
"and",
"now",
"-",
"timestamp",
">",
"self",
".",
"SESSION_EXPIRE_TIME",
":",
"Session",
"(",
"sess_id",
")",
".",
"delete",
"(",
")",
"return",
"False",
"else",
":",
"self",
".",
"set",
"(",
"now",
")",
"return",
"True"
] |
Deletes session if keepalive request expired
otherwise updates the keepalive timestamp value
|
[
"Deletes",
"session",
"if",
"keepalive",
"request",
"expired",
"otherwise",
"updates",
"the",
"keepalive",
"timestamp",
"value"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/cache.py#L250-L265
|
train
|
zetaops/zengine
|
zengine/receivers.py
|
send_message_for_lane_change
|
def send_message_for_lane_change(sender, **kwargs):
"""
Sends a message to possible owners of the current workflows
next lane.
Args:
**kwargs: ``current`` and ``possible_owners`` are required.
sender (User): User object
"""
current = kwargs['current']
owners = kwargs['possible_owners']
if 'lane_change_invite' in current.task_data:
msg_context = current.task_data.pop('lane_change_invite')
else:
msg_context = DEFAULT_LANE_CHANGE_INVITE_MSG
wfi = WFCache(current).get_instance()
# Deletion of used passive task invitation which belongs to previous lane.
TaskInvitation.objects.filter(instance=wfi, role=current.role, wf_name=wfi.wf.name).delete()
today = datetime.today()
for recipient in owners:
inv = TaskInvitation(
instance=wfi,
role=recipient,
wf_name=wfi.wf.name,
progress=30,
start_date=today,
finish_date=today + timedelta(15)
)
inv.title = current.task_data.get('INVITATION_TITLE') or wfi.wf.title
inv.save()
# try to send notification, if it fails go on
try:
recipient.send_notification(title=msg_context['title'],
message="%s %s" % (wfi.wf.title, msg_context['body']),
typ=1, # info
url='',
sender=sender
)
except: # todo: specify which exception
pass
|
python
|
def send_message_for_lane_change(sender, **kwargs):
"""
Sends a message to possible owners of the current workflows
next lane.
Args:
**kwargs: ``current`` and ``possible_owners`` are required.
sender (User): User object
"""
current = kwargs['current']
owners = kwargs['possible_owners']
if 'lane_change_invite' in current.task_data:
msg_context = current.task_data.pop('lane_change_invite')
else:
msg_context = DEFAULT_LANE_CHANGE_INVITE_MSG
wfi = WFCache(current).get_instance()
# Deletion of used passive task invitation which belongs to previous lane.
TaskInvitation.objects.filter(instance=wfi, role=current.role, wf_name=wfi.wf.name).delete()
today = datetime.today()
for recipient in owners:
inv = TaskInvitation(
instance=wfi,
role=recipient,
wf_name=wfi.wf.name,
progress=30,
start_date=today,
finish_date=today + timedelta(15)
)
inv.title = current.task_data.get('INVITATION_TITLE') or wfi.wf.title
inv.save()
# try to send notification, if it fails go on
try:
recipient.send_notification(title=msg_context['title'],
message="%s %s" % (wfi.wf.title, msg_context['body']),
typ=1, # info
url='',
sender=sender
)
except: # todo: specify which exception
pass
|
[
"def",
"send_message_for_lane_change",
"(",
"sender",
",",
"*",
"*",
"kwargs",
")",
":",
"current",
"=",
"kwargs",
"[",
"'current'",
"]",
"owners",
"=",
"kwargs",
"[",
"'possible_owners'",
"]",
"if",
"'lane_change_invite'",
"in",
"current",
".",
"task_data",
":",
"msg_context",
"=",
"current",
".",
"task_data",
".",
"pop",
"(",
"'lane_change_invite'",
")",
"else",
":",
"msg_context",
"=",
"DEFAULT_LANE_CHANGE_INVITE_MSG",
"wfi",
"=",
"WFCache",
"(",
"current",
")",
".",
"get_instance",
"(",
")",
"# Deletion of used passive task invitation which belongs to previous lane.",
"TaskInvitation",
".",
"objects",
".",
"filter",
"(",
"instance",
"=",
"wfi",
",",
"role",
"=",
"current",
".",
"role",
",",
"wf_name",
"=",
"wfi",
".",
"wf",
".",
"name",
")",
".",
"delete",
"(",
")",
"today",
"=",
"datetime",
".",
"today",
"(",
")",
"for",
"recipient",
"in",
"owners",
":",
"inv",
"=",
"TaskInvitation",
"(",
"instance",
"=",
"wfi",
",",
"role",
"=",
"recipient",
",",
"wf_name",
"=",
"wfi",
".",
"wf",
".",
"name",
",",
"progress",
"=",
"30",
",",
"start_date",
"=",
"today",
",",
"finish_date",
"=",
"today",
"+",
"timedelta",
"(",
"15",
")",
")",
"inv",
".",
"title",
"=",
"current",
".",
"task_data",
".",
"get",
"(",
"'INVITATION_TITLE'",
")",
"or",
"wfi",
".",
"wf",
".",
"title",
"inv",
".",
"save",
"(",
")",
"# try to send notification, if it fails go on",
"try",
":",
"recipient",
".",
"send_notification",
"(",
"title",
"=",
"msg_context",
"[",
"'title'",
"]",
",",
"message",
"=",
"\"%s %s\"",
"%",
"(",
"wfi",
".",
"wf",
".",
"title",
",",
"msg_context",
"[",
"'body'",
"]",
")",
",",
"typ",
"=",
"1",
",",
"# info",
"url",
"=",
"''",
",",
"sender",
"=",
"sender",
")",
"except",
":",
"# todo: specify which exception",
"pass"
] |
Sends a message to possible owners of the current workflows
next lane.
Args:
**kwargs: ``current`` and ``possible_owners`` are required.
sender (User): User object
|
[
"Sends",
"a",
"message",
"to",
"possible",
"owners",
"of",
"the",
"current",
"workflows",
"next",
"lane",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/receivers.py#L29-L73
|
train
|
zetaops/zengine
|
zengine/receivers.py
|
set_password
|
def set_password(sender, **kwargs):
"""
Encrypts password of the user.
"""
if sender.model_class.__name__ == 'User':
usr = kwargs['object']
if not usr.password.startswith('$pbkdf2'):
usr.set_password(usr.password)
usr.save()
|
python
|
def set_password(sender, **kwargs):
"""
Encrypts password of the user.
"""
if sender.model_class.__name__ == 'User':
usr = kwargs['object']
if not usr.password.startswith('$pbkdf2'):
usr.set_password(usr.password)
usr.save()
|
[
"def",
"set_password",
"(",
"sender",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"sender",
".",
"model_class",
".",
"__name__",
"==",
"'User'",
":",
"usr",
"=",
"kwargs",
"[",
"'object'",
"]",
"if",
"not",
"usr",
".",
"password",
".",
"startswith",
"(",
"'$pbkdf2'",
")",
":",
"usr",
".",
"set_password",
"(",
"usr",
".",
"password",
")",
"usr",
".",
"save",
"(",
")"
] |
Encrypts password of the user.
|
[
"Encrypts",
"password",
"of",
"the",
"user",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/receivers.py#L78-L86
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.channel_list
|
def channel_list(self):
"""
Main screen for channel management.
Channels listed and operations can be chosen on the screen.
If there is an error message like non-choice,
it is shown here.
"""
if self.current.task_data.get('msg', False):
if self.current.task_data.get('target_channel_key', False):
self.current.output['msgbox'] = {'type': 'info',
"title": _(u"Successful Operation"),
"msg": self.current.task_data['msg']}
del self.current.task_data['msg']
else:
self.show_warning_messages()
self.current.task_data['new_channel'] = False
_form = ChannelListForm(title=_(u'Public Channel List'), help_text=CHANNEL_CHOICE_TEXT)
for channel in Channel.objects.filter(typ=15):
owner_name = channel.owner.username
_form.ChannelList(choice=False, name=channel.name, owner=owner_name, key=channel.key)
_form.new_channel = fields.Button(_(u"Merge At New Channel"), cmd="create_new_channel")
_form.existing_channel = fields.Button(_(u"Merge With An Existing Channel"),
cmd="choose_existing_channel")
_form.find_chosen_channel = fields.Button(_(u"Split Channel"), cmd="find_chosen_channel")
self.form_out(_form)
|
python
|
def channel_list(self):
"""
Main screen for channel management.
Channels listed and operations can be chosen on the screen.
If there is an error message like non-choice,
it is shown here.
"""
if self.current.task_data.get('msg', False):
if self.current.task_data.get('target_channel_key', False):
self.current.output['msgbox'] = {'type': 'info',
"title": _(u"Successful Operation"),
"msg": self.current.task_data['msg']}
del self.current.task_data['msg']
else:
self.show_warning_messages()
self.current.task_data['new_channel'] = False
_form = ChannelListForm(title=_(u'Public Channel List'), help_text=CHANNEL_CHOICE_TEXT)
for channel in Channel.objects.filter(typ=15):
owner_name = channel.owner.username
_form.ChannelList(choice=False, name=channel.name, owner=owner_name, key=channel.key)
_form.new_channel = fields.Button(_(u"Merge At New Channel"), cmd="create_new_channel")
_form.existing_channel = fields.Button(_(u"Merge With An Existing Channel"),
cmd="choose_existing_channel")
_form.find_chosen_channel = fields.Button(_(u"Split Channel"), cmd="find_chosen_channel")
self.form_out(_form)
|
[
"def",
"channel_list",
"(",
"self",
")",
":",
"if",
"self",
".",
"current",
".",
"task_data",
".",
"get",
"(",
"'msg'",
",",
"False",
")",
":",
"if",
"self",
".",
"current",
".",
"task_data",
".",
"get",
"(",
"'target_channel_key'",
",",
"False",
")",
":",
"self",
".",
"current",
".",
"output",
"[",
"'msgbox'",
"]",
"=",
"{",
"'type'",
":",
"'info'",
",",
"\"title\"",
":",
"_",
"(",
"u\"Successful Operation\"",
")",
",",
"\"msg\"",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"}",
"del",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"else",
":",
"self",
".",
"show_warning_messages",
"(",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'new_channel'",
"]",
"=",
"False",
"_form",
"=",
"ChannelListForm",
"(",
"title",
"=",
"_",
"(",
"u'Public Channel List'",
")",
",",
"help_text",
"=",
"CHANNEL_CHOICE_TEXT",
")",
"for",
"channel",
"in",
"Channel",
".",
"objects",
".",
"filter",
"(",
"typ",
"=",
"15",
")",
":",
"owner_name",
"=",
"channel",
".",
"owner",
".",
"username",
"_form",
".",
"ChannelList",
"(",
"choice",
"=",
"False",
",",
"name",
"=",
"channel",
".",
"name",
",",
"owner",
"=",
"owner_name",
",",
"key",
"=",
"channel",
".",
"key",
")",
"_form",
".",
"new_channel",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Merge At New Channel\"",
")",
",",
"cmd",
"=",
"\"create_new_channel\"",
")",
"_form",
".",
"existing_channel",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Merge With An Existing Channel\"",
")",
",",
"cmd",
"=",
"\"choose_existing_channel\"",
")",
"_form",
".",
"find_chosen_channel",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Split Channel\"",
")",
",",
"cmd",
"=",
"\"find_chosen_channel\"",
")",
"self",
".",
"form_out",
"(",
"_form",
")"
] |
Main screen for channel management.
Channels listed and operations can be chosen on the screen.
If there is an error message like non-choice,
it is shown here.
|
[
"Main",
"screen",
"for",
"channel",
"management",
".",
"Channels",
"listed",
"and",
"operations",
"can",
"be",
"chosen",
"on",
"the",
"screen",
".",
"If",
"there",
"is",
"an",
"error",
"message",
"like",
"non",
"-",
"choice",
"it",
"is",
"shown",
"here",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L57-L86
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.channel_choice_control
|
def channel_choice_control(self):
"""
It controls errors. If there is an error,
returns channel list screen with error message.
"""
self.current.task_data['control'], self.current.task_data['msg'] \
= self.selection_error_control(self.input['form'])
if self.current.task_data['control']:
self.current.task_data['option'] = self.input['cmd']
self.current.task_data['split_operation'] = False
keys, names = self.return_selected_form_items(self.input['form']['ChannelList'])
self.current.task_data['chosen_channels'] = keys
self.current.task_data['chosen_channels_names'] = names
|
python
|
def channel_choice_control(self):
"""
It controls errors. If there is an error,
returns channel list screen with error message.
"""
self.current.task_data['control'], self.current.task_data['msg'] \
= self.selection_error_control(self.input['form'])
if self.current.task_data['control']:
self.current.task_data['option'] = self.input['cmd']
self.current.task_data['split_operation'] = False
keys, names = self.return_selected_form_items(self.input['form']['ChannelList'])
self.current.task_data['chosen_channels'] = keys
self.current.task_data['chosen_channels_names'] = names
|
[
"def",
"channel_choice_control",
"(",
"self",
")",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'control'",
"]",
",",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"=",
"self",
".",
"selection_error_control",
"(",
"self",
".",
"input",
"[",
"'form'",
"]",
")",
"if",
"self",
".",
"current",
".",
"task_data",
"[",
"'control'",
"]",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'option'",
"]",
"=",
"self",
".",
"input",
"[",
"'cmd'",
"]",
"self",
".",
"current",
".",
"task_data",
"[",
"'split_operation'",
"]",
"=",
"False",
"keys",
",",
"names",
"=",
"self",
".",
"return_selected_form_items",
"(",
"self",
".",
"input",
"[",
"'form'",
"]",
"[",
"'ChannelList'",
"]",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels'",
"]",
"=",
"keys",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels_names'",
"]",
"=",
"names"
] |
It controls errors. If there is an error,
returns channel list screen with error message.
|
[
"It",
"controls",
"errors",
".",
"If",
"there",
"is",
"an",
"error",
"returns",
"channel",
"list",
"screen",
"with",
"error",
"message",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L88-L100
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.create_new_channel
|
def create_new_channel(self):
"""
Features of new channel are specified like channel's name, owner etc.
"""
self.current.task_data['new_channel'] = True
_form = NewChannelForm(Channel(), current=self.current)
_form.title = _(u"Specify Features of New Channel to Create")
_form.forward = fields.Button(_(u"Create"), flow="find_target_channel")
self.form_out(_form)
|
python
|
def create_new_channel(self):
"""
Features of new channel are specified like channel's name, owner etc.
"""
self.current.task_data['new_channel'] = True
_form = NewChannelForm(Channel(), current=self.current)
_form.title = _(u"Specify Features of New Channel to Create")
_form.forward = fields.Button(_(u"Create"), flow="find_target_channel")
self.form_out(_form)
|
[
"def",
"create_new_channel",
"(",
"self",
")",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'new_channel'",
"]",
"=",
"True",
"_form",
"=",
"NewChannelForm",
"(",
"Channel",
"(",
")",
",",
"current",
"=",
"self",
".",
"current",
")",
"_form",
".",
"title",
"=",
"_",
"(",
"u\"Specify Features of New Channel to Create\"",
")",
"_form",
".",
"forward",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Create\"",
")",
",",
"flow",
"=",
"\"find_target_channel\"",
")",
"self",
".",
"form_out",
"(",
"_form",
")"
] |
Features of new channel are specified like channel's name, owner etc.
|
[
"Features",
"of",
"new",
"channel",
"are",
"specified",
"like",
"channel",
"s",
"name",
"owner",
"etc",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L102-L111
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.save_new_channel
|
def save_new_channel(self):
"""
It saves new channel according to specified channel features.
"""
form_info = self.input['form']
channel = Channel(typ=15, name=form_info['name'],
description=form_info['description'],
owner_id=form_info['owner_id'])
channel.blocking_save()
self.current.task_data['target_channel_key'] = channel.key
|
python
|
def save_new_channel(self):
"""
It saves new channel according to specified channel features.
"""
form_info = self.input['form']
channel = Channel(typ=15, name=form_info['name'],
description=form_info['description'],
owner_id=form_info['owner_id'])
channel.blocking_save()
self.current.task_data['target_channel_key'] = channel.key
|
[
"def",
"save_new_channel",
"(",
"self",
")",
":",
"form_info",
"=",
"self",
".",
"input",
"[",
"'form'",
"]",
"channel",
"=",
"Channel",
"(",
"typ",
"=",
"15",
",",
"name",
"=",
"form_info",
"[",
"'name'",
"]",
",",
"description",
"=",
"form_info",
"[",
"'description'",
"]",
",",
"owner_id",
"=",
"form_info",
"[",
"'owner_id'",
"]",
")",
"channel",
".",
"blocking_save",
"(",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'target_channel_key'",
"]",
"=",
"channel",
".",
"key"
] |
It saves new channel according to specified channel features.
|
[
"It",
"saves",
"new",
"channel",
"according",
"to",
"specified",
"channel",
"features",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L113-L123
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.choose_existing_channel
|
def choose_existing_channel(self):
"""
It is a channel choice list and chosen channels
at previous step shouldn't be on the screen.
"""
if self.current.task_data.get('msg', False):
self.show_warning_messages()
_form = ChannelListForm()
_form.title = _(u"Choose a Channel Which Will Be Merged With Chosen Channels")
for channel in Channel.objects.filter(typ=15).exclude(
key__in=self.current.task_data['chosen_channels']):
owner_name = channel.owner.username
_form.ChannelList(choice=False, name=channel.name, owner=owner_name,
key=channel.key)
_form.choose = fields.Button(_(u"Choose"))
self.form_out(_form)
|
python
|
def choose_existing_channel(self):
"""
It is a channel choice list and chosen channels
at previous step shouldn't be on the screen.
"""
if self.current.task_data.get('msg', False):
self.show_warning_messages()
_form = ChannelListForm()
_form.title = _(u"Choose a Channel Which Will Be Merged With Chosen Channels")
for channel in Channel.objects.filter(typ=15).exclude(
key__in=self.current.task_data['chosen_channels']):
owner_name = channel.owner.username
_form.ChannelList(choice=False, name=channel.name, owner=owner_name,
key=channel.key)
_form.choose = fields.Button(_(u"Choose"))
self.form_out(_form)
|
[
"def",
"choose_existing_channel",
"(",
"self",
")",
":",
"if",
"self",
".",
"current",
".",
"task_data",
".",
"get",
"(",
"'msg'",
",",
"False",
")",
":",
"self",
".",
"show_warning_messages",
"(",
")",
"_form",
"=",
"ChannelListForm",
"(",
")",
"_form",
".",
"title",
"=",
"_",
"(",
"u\"Choose a Channel Which Will Be Merged With Chosen Channels\"",
")",
"for",
"channel",
"in",
"Channel",
".",
"objects",
".",
"filter",
"(",
"typ",
"=",
"15",
")",
".",
"exclude",
"(",
"key__in",
"=",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels'",
"]",
")",
":",
"owner_name",
"=",
"channel",
".",
"owner",
".",
"username",
"_form",
".",
"ChannelList",
"(",
"choice",
"=",
"False",
",",
"name",
"=",
"channel",
".",
"name",
",",
"owner",
"=",
"owner_name",
",",
"key",
"=",
"channel",
".",
"key",
")",
"_form",
".",
"choose",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Choose\"",
")",
")",
"self",
".",
"form_out",
"(",
"_form",
")"
] |
It is a channel choice list and chosen channels
at previous step shouldn't be on the screen.
|
[
"It",
"is",
"a",
"channel",
"choice",
"list",
"and",
"chosen",
"channels",
"at",
"previous",
"step",
"shouldn",
"t",
"be",
"on",
"the",
"screen",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L125-L144
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.existing_choice_control
|
def existing_choice_control(self):
"""
It controls errors. It generates an error message
if zero or more than one channels are selected.
"""
self.current.task_data['existing'] = False
self.current.task_data['msg'] = _(u"You should choose just one channel to do operation.")
keys, names = self.return_selected_form_items(self.input['form']['ChannelList'])
if len(keys) == 1:
self.current.task_data['existing'] = True
self.current.task_data['target_channel_key'] = keys[0]
|
python
|
def existing_choice_control(self):
"""
It controls errors. It generates an error message
if zero or more than one channels are selected.
"""
self.current.task_data['existing'] = False
self.current.task_data['msg'] = _(u"You should choose just one channel to do operation.")
keys, names = self.return_selected_form_items(self.input['form']['ChannelList'])
if len(keys) == 1:
self.current.task_data['existing'] = True
self.current.task_data['target_channel_key'] = keys[0]
|
[
"def",
"existing_choice_control",
"(",
"self",
")",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'existing'",
"]",
"=",
"False",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"=",
"_",
"(",
"u\"You should choose just one channel to do operation.\"",
")",
"keys",
",",
"names",
"=",
"self",
".",
"return_selected_form_items",
"(",
"self",
".",
"input",
"[",
"'form'",
"]",
"[",
"'ChannelList'",
"]",
")",
"if",
"len",
"(",
"keys",
")",
"==",
"1",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'existing'",
"]",
"=",
"True",
"self",
".",
"current",
".",
"task_data",
"[",
"'target_channel_key'",
"]",
"=",
"keys",
"[",
"0",
"]"
] |
It controls errors. It generates an error message
if zero or more than one channels are selected.
|
[
"It",
"controls",
"errors",
".",
"It",
"generates",
"an",
"error",
"message",
"if",
"zero",
"or",
"more",
"than",
"one",
"channels",
"are",
"selected",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L146-L156
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.split_channel
|
def split_channel(self):
"""
A channel can be splitted to new channel or other existing channel.
It creates subscribers list as selectable to moved.
"""
if self.current.task_data.get('msg', False):
self.show_warning_messages()
self.current.task_data['split_operation'] = True
channel = Channel.objects.get(self.current.task_data['chosen_channels'][0])
_form = SubscriberListForm(title=_(u'Choose Subscribers to Migrate'))
for subscriber in Subscriber.objects.filter(channel=channel):
subscriber_name = subscriber.user.username
_form.SubscriberList(choice=False, name=subscriber_name, key=subscriber.key)
_form.new_channel = fields.Button(_(u"Move to a New Channel"), cmd="create_new_channel")
_form.existing_channel = fields.Button(_(u"Move to an Existing Channel"),
cmd="choose_existing_channel")
self.form_out(_form)
|
python
|
def split_channel(self):
"""
A channel can be splitted to new channel or other existing channel.
It creates subscribers list as selectable to moved.
"""
if self.current.task_data.get('msg', False):
self.show_warning_messages()
self.current.task_data['split_operation'] = True
channel = Channel.objects.get(self.current.task_data['chosen_channels'][0])
_form = SubscriberListForm(title=_(u'Choose Subscribers to Migrate'))
for subscriber in Subscriber.objects.filter(channel=channel):
subscriber_name = subscriber.user.username
_form.SubscriberList(choice=False, name=subscriber_name, key=subscriber.key)
_form.new_channel = fields.Button(_(u"Move to a New Channel"), cmd="create_new_channel")
_form.existing_channel = fields.Button(_(u"Move to an Existing Channel"),
cmd="choose_existing_channel")
self.form_out(_form)
|
[
"def",
"split_channel",
"(",
"self",
")",
":",
"if",
"self",
".",
"current",
".",
"task_data",
".",
"get",
"(",
"'msg'",
",",
"False",
")",
":",
"self",
".",
"show_warning_messages",
"(",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'split_operation'",
"]",
"=",
"True",
"channel",
"=",
"Channel",
".",
"objects",
".",
"get",
"(",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels'",
"]",
"[",
"0",
"]",
")",
"_form",
"=",
"SubscriberListForm",
"(",
"title",
"=",
"_",
"(",
"u'Choose Subscribers to Migrate'",
")",
")",
"for",
"subscriber",
"in",
"Subscriber",
".",
"objects",
".",
"filter",
"(",
"channel",
"=",
"channel",
")",
":",
"subscriber_name",
"=",
"subscriber",
".",
"user",
".",
"username",
"_form",
".",
"SubscriberList",
"(",
"choice",
"=",
"False",
",",
"name",
"=",
"subscriber_name",
",",
"key",
"=",
"subscriber",
".",
"key",
")",
"_form",
".",
"new_channel",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Move to a New Channel\"",
")",
",",
"cmd",
"=",
"\"create_new_channel\"",
")",
"_form",
".",
"existing_channel",
"=",
"fields",
".",
"Button",
"(",
"_",
"(",
"u\"Move to an Existing Channel\"",
")",
",",
"cmd",
"=",
"\"choose_existing_channel\"",
")",
"self",
".",
"form_out",
"(",
"_form",
")"
] |
A channel can be splitted to new channel or other existing channel.
It creates subscribers list as selectable to moved.
|
[
"A",
"channel",
"can",
"be",
"splitted",
"to",
"new",
"channel",
"or",
"other",
"existing",
"channel",
".",
"It",
"creates",
"subscribers",
"list",
"as",
"selectable",
"to",
"moved",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L158-L179
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.subscriber_choice_control
|
def subscriber_choice_control(self):
"""
It controls subscribers choice and generates
error message if there is a non-choice.
"""
self.current.task_data['option'] = None
self.current.task_data['chosen_subscribers'], names = self.return_selected_form_items(
self.input['form']['SubscriberList'])
self.current.task_data[
'msg'] = "You should choose at least one subscriber for migration operation."
if self.current.task_data['chosen_subscribers']:
self.current.task_data['option'] = self.input['cmd']
del self.current.task_data['msg']
|
python
|
def subscriber_choice_control(self):
"""
It controls subscribers choice and generates
error message if there is a non-choice.
"""
self.current.task_data['option'] = None
self.current.task_data['chosen_subscribers'], names = self.return_selected_form_items(
self.input['form']['SubscriberList'])
self.current.task_data[
'msg'] = "You should choose at least one subscriber for migration operation."
if self.current.task_data['chosen_subscribers']:
self.current.task_data['option'] = self.input['cmd']
del self.current.task_data['msg']
|
[
"def",
"subscriber_choice_control",
"(",
"self",
")",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'option'",
"]",
"=",
"None",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_subscribers'",
"]",
",",
"names",
"=",
"self",
".",
"return_selected_form_items",
"(",
"self",
".",
"input",
"[",
"'form'",
"]",
"[",
"'SubscriberList'",
"]",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"=",
"\"You should choose at least one subscriber for migration operation.\"",
"if",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_subscribers'",
"]",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"'option'",
"]",
"=",
"self",
".",
"input",
"[",
"'cmd'",
"]",
"del",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]"
] |
It controls subscribers choice and generates
error message if there is a non-choice.
|
[
"It",
"controls",
"subscribers",
"choice",
"and",
"generates",
"error",
"message",
"if",
"there",
"is",
"a",
"non",
"-",
"choice",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L181-L193
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.move_complete_channel
|
def move_complete_channel(self):
"""
Channels and theirs subscribers are moved
completely to new channel or existing channel.
"""
to_channel = Channel.objects.get(self.current.task_data['target_channel_key'])
chosen_channels = self.current.task_data['chosen_channels']
chosen_channels_names = self.current.task_data['chosen_channels_names']
with BlockSave(Subscriber, query_dict={'channel_id': to_channel.key}):
for s in Subscriber.objects.filter(channel_id__in=chosen_channels, typ=15):
s.channel = to_channel
s.save()
with BlockDelete(Message):
Message.objects.filter(channel_id__in=chosen_channels, typ=15).delete()
with BlockDelete(Channel):
Channel.objects.filter(key__in=chosen_channels).delete()
self.current.task_data[
'msg'] = _(u"Chosen channels(%s) have been merged to '%s' channel successfully.") % \
(', '.join(chosen_channels_names), to_channel.name)
|
python
|
def move_complete_channel(self):
"""
Channels and theirs subscribers are moved
completely to new channel or existing channel.
"""
to_channel = Channel.objects.get(self.current.task_data['target_channel_key'])
chosen_channels = self.current.task_data['chosen_channels']
chosen_channels_names = self.current.task_data['chosen_channels_names']
with BlockSave(Subscriber, query_dict={'channel_id': to_channel.key}):
for s in Subscriber.objects.filter(channel_id__in=chosen_channels, typ=15):
s.channel = to_channel
s.save()
with BlockDelete(Message):
Message.objects.filter(channel_id__in=chosen_channels, typ=15).delete()
with BlockDelete(Channel):
Channel.objects.filter(key__in=chosen_channels).delete()
self.current.task_data[
'msg'] = _(u"Chosen channels(%s) have been merged to '%s' channel successfully.") % \
(', '.join(chosen_channels_names), to_channel.name)
|
[
"def",
"move_complete_channel",
"(",
"self",
")",
":",
"to_channel",
"=",
"Channel",
".",
"objects",
".",
"get",
"(",
"self",
".",
"current",
".",
"task_data",
"[",
"'target_channel_key'",
"]",
")",
"chosen_channels",
"=",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels'",
"]",
"chosen_channels_names",
"=",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels_names'",
"]",
"with",
"BlockSave",
"(",
"Subscriber",
",",
"query_dict",
"=",
"{",
"'channel_id'",
":",
"to_channel",
".",
"key",
"}",
")",
":",
"for",
"s",
"in",
"Subscriber",
".",
"objects",
".",
"filter",
"(",
"channel_id__in",
"=",
"chosen_channels",
",",
"typ",
"=",
"15",
")",
":",
"s",
".",
"channel",
"=",
"to_channel",
"s",
".",
"save",
"(",
")",
"with",
"BlockDelete",
"(",
"Message",
")",
":",
"Message",
".",
"objects",
".",
"filter",
"(",
"channel_id__in",
"=",
"chosen_channels",
",",
"typ",
"=",
"15",
")",
".",
"delete",
"(",
")",
"with",
"BlockDelete",
"(",
"Channel",
")",
":",
"Channel",
".",
"objects",
".",
"filter",
"(",
"key__in",
"=",
"chosen_channels",
")",
".",
"delete",
"(",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"=",
"_",
"(",
"u\"Chosen channels(%s) have been merged to '%s' channel successfully.\"",
")",
"%",
"(",
"', '",
".",
"join",
"(",
"chosen_channels_names",
")",
",",
"to_channel",
".",
"name",
")"
] |
Channels and theirs subscribers are moved
completely to new channel or existing channel.
|
[
"Channels",
"and",
"theirs",
"subscribers",
"are",
"moved",
"completely",
"to",
"new",
"channel",
"or",
"existing",
"channel",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L195-L218
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.move_chosen_subscribers
|
def move_chosen_subscribers(self):
"""
After splitting operation, only chosen subscribers
are moved to new channel or existing channel.
"""
from_channel = Channel.objects.get(self.current.task_data['chosen_channels'][0])
to_channel = Channel.objects.get(self.current.task_data['target_channel_key'])
with BlockSave(Subscriber, query_dict={'channel_id': to_channel.key}):
for subscriber in Subscriber.objects.filter(
key__in=self.current.task_data['chosen_subscribers']):
subscriber.channel = to_channel
subscriber.save()
if self.current.task_data['new_channel']:
self.copy_and_move_messages(from_channel, to_channel)
self.current.task_data[
'msg'] = _(u"Chosen subscribers and messages of them migrated from '%s' channel to "
u"'%s' channel successfully.") % (from_channel.name, to_channel.name)
|
python
|
def move_chosen_subscribers(self):
"""
After splitting operation, only chosen subscribers
are moved to new channel or existing channel.
"""
from_channel = Channel.objects.get(self.current.task_data['chosen_channels'][0])
to_channel = Channel.objects.get(self.current.task_data['target_channel_key'])
with BlockSave(Subscriber, query_dict={'channel_id': to_channel.key}):
for subscriber in Subscriber.objects.filter(
key__in=self.current.task_data['chosen_subscribers']):
subscriber.channel = to_channel
subscriber.save()
if self.current.task_data['new_channel']:
self.copy_and_move_messages(from_channel, to_channel)
self.current.task_data[
'msg'] = _(u"Chosen subscribers and messages of them migrated from '%s' channel to "
u"'%s' channel successfully.") % (from_channel.name, to_channel.name)
|
[
"def",
"move_chosen_subscribers",
"(",
"self",
")",
":",
"from_channel",
"=",
"Channel",
".",
"objects",
".",
"get",
"(",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_channels'",
"]",
"[",
"0",
"]",
")",
"to_channel",
"=",
"Channel",
".",
"objects",
".",
"get",
"(",
"self",
".",
"current",
".",
"task_data",
"[",
"'target_channel_key'",
"]",
")",
"with",
"BlockSave",
"(",
"Subscriber",
",",
"query_dict",
"=",
"{",
"'channel_id'",
":",
"to_channel",
".",
"key",
"}",
")",
":",
"for",
"subscriber",
"in",
"Subscriber",
".",
"objects",
".",
"filter",
"(",
"key__in",
"=",
"self",
".",
"current",
".",
"task_data",
"[",
"'chosen_subscribers'",
"]",
")",
":",
"subscriber",
".",
"channel",
"=",
"to_channel",
"subscriber",
".",
"save",
"(",
")",
"if",
"self",
".",
"current",
".",
"task_data",
"[",
"'new_channel'",
"]",
":",
"self",
".",
"copy_and_move_messages",
"(",
"from_channel",
",",
"to_channel",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"=",
"_",
"(",
"u\"Chosen subscribers and messages of them migrated from '%s' channel to \"",
"u\"'%s' channel successfully.\"",
")",
"%",
"(",
"from_channel",
".",
"name",
",",
"to_channel",
".",
"name",
")"
] |
After splitting operation, only chosen subscribers
are moved to new channel or existing channel.
|
[
"After",
"splitting",
"operation",
"only",
"chosen",
"subscribers",
"are",
"moved",
"to",
"new",
"channel",
"or",
"existing",
"channel",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L220-L239
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.copy_and_move_messages
|
def copy_and_move_messages(from_channel, to_channel):
"""
While splitting channel and moving chosen subscribers to new channel,
old channel's messages are copied and moved to new channel.
Args:
from_channel (Channel object): move messages from channel
to_channel (Channel object): move messages to channel
"""
with BlockSave(Message, query_dict={'channel_id': to_channel.key}):
for message in Message.objects.filter(channel=from_channel, typ=15):
message.key = ''
message.channel = to_channel
message.save()
|
python
|
def copy_and_move_messages(from_channel, to_channel):
"""
While splitting channel and moving chosen subscribers to new channel,
old channel's messages are copied and moved to new channel.
Args:
from_channel (Channel object): move messages from channel
to_channel (Channel object): move messages to channel
"""
with BlockSave(Message, query_dict={'channel_id': to_channel.key}):
for message in Message.objects.filter(channel=from_channel, typ=15):
message.key = ''
message.channel = to_channel
message.save()
|
[
"def",
"copy_and_move_messages",
"(",
"from_channel",
",",
"to_channel",
")",
":",
"with",
"BlockSave",
"(",
"Message",
",",
"query_dict",
"=",
"{",
"'channel_id'",
":",
"to_channel",
".",
"key",
"}",
")",
":",
"for",
"message",
"in",
"Message",
".",
"objects",
".",
"filter",
"(",
"channel",
"=",
"from_channel",
",",
"typ",
"=",
"15",
")",
":",
"message",
".",
"key",
"=",
"''",
"message",
".",
"channel",
"=",
"to_channel",
"message",
".",
"save",
"(",
")"
] |
While splitting channel and moving chosen subscribers to new channel,
old channel's messages are copied and moved to new channel.
Args:
from_channel (Channel object): move messages from channel
to_channel (Channel object): move messages to channel
|
[
"While",
"splitting",
"channel",
"and",
"moving",
"chosen",
"subscribers",
"to",
"new",
"channel",
"old",
"channel",
"s",
"messages",
"are",
"copied",
"and",
"moved",
"to",
"new",
"channel",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L242-L255
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.show_warning_messages
|
def show_warning_messages(self, title=_(u"Incorrect Operation"), box_type='warning'):
"""
It shows incorrect operations or successful operation messages.
Args:
title (string): title of message box
box_type (string): type of message box (warning, info)
"""
msg = self.current.task_data['msg']
self.current.output['msgbox'] = {'type': box_type, "title": title, "msg": msg}
del self.current.task_data['msg']
|
python
|
def show_warning_messages(self, title=_(u"Incorrect Operation"), box_type='warning'):
"""
It shows incorrect operations or successful operation messages.
Args:
title (string): title of message box
box_type (string): type of message box (warning, info)
"""
msg = self.current.task_data['msg']
self.current.output['msgbox'] = {'type': box_type, "title": title, "msg": msg}
del self.current.task_data['msg']
|
[
"def",
"show_warning_messages",
"(",
"self",
",",
"title",
"=",
"_",
"(",
"u\"Incorrect Operation\"",
")",
",",
"box_type",
"=",
"'warning'",
")",
":",
"msg",
"=",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]",
"self",
".",
"current",
".",
"output",
"[",
"'msgbox'",
"]",
"=",
"{",
"'type'",
":",
"box_type",
",",
"\"title\"",
":",
"title",
",",
"\"msg\"",
":",
"msg",
"}",
"del",
"self",
".",
"current",
".",
"task_data",
"[",
"'msg'",
"]"
] |
It shows incorrect operations or successful operation messages.
Args:
title (string): title of message box
box_type (string): type of message box (warning, info)
|
[
"It",
"shows",
"incorrect",
"operations",
"or",
"successful",
"operation",
"messages",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L257-L267
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.return_selected_form_items
|
def return_selected_form_items(form_info):
"""
It returns chosen keys list from a given form.
Args:
form_info: serialized list of dict form data
Returns:
selected_keys(list): Chosen keys list
selected_names(list): Chosen channels' or subscribers' names.
"""
selected_keys = []
selected_names = []
for chosen in form_info:
if chosen['choice']:
selected_keys.append(chosen['key'])
selected_names.append(chosen['name'])
return selected_keys, selected_names
|
python
|
def return_selected_form_items(form_info):
"""
It returns chosen keys list from a given form.
Args:
form_info: serialized list of dict form data
Returns:
selected_keys(list): Chosen keys list
selected_names(list): Chosen channels' or subscribers' names.
"""
selected_keys = []
selected_names = []
for chosen in form_info:
if chosen['choice']:
selected_keys.append(chosen['key'])
selected_names.append(chosen['name'])
return selected_keys, selected_names
|
[
"def",
"return_selected_form_items",
"(",
"form_info",
")",
":",
"selected_keys",
"=",
"[",
"]",
"selected_names",
"=",
"[",
"]",
"for",
"chosen",
"in",
"form_info",
":",
"if",
"chosen",
"[",
"'choice'",
"]",
":",
"selected_keys",
".",
"append",
"(",
"chosen",
"[",
"'key'",
"]",
")",
"selected_names",
".",
"append",
"(",
"chosen",
"[",
"'name'",
"]",
")",
"return",
"selected_keys",
",",
"selected_names"
] |
It returns chosen keys list from a given form.
Args:
form_info: serialized list of dict form data
Returns:
selected_keys(list): Chosen keys list
selected_names(list): Chosen channels' or subscribers' names.
|
[
"It",
"returns",
"chosen",
"keys",
"list",
"from",
"a",
"given",
"form",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L270-L287
|
train
|
zetaops/zengine
|
zengine/views/channel_management.py
|
ChannelManagement.selection_error_control
|
def selection_error_control(self, form_info):
"""
It controls the selection from the form according
to the operations, and returns an error message
if it does not comply with the rules.
Args:
form_info: Channel or subscriber form from the user
Returns: True or False
error message
"""
keys, names = self.return_selected_form_items(form_info['ChannelList'])
chosen_channels_number = len(keys)
if form_info['new_channel'] and chosen_channels_number < 2:
return False, _(
u"You should choose at least two channel to merge operation at a new channel.")
elif form_info['existing_channel'] and chosen_channels_number == 0:
return False, _(
u"You should choose at least one channel to merge operation with existing channel.")
elif form_info['find_chosen_channel'] and chosen_channels_number != 1:
return False, _(u"You should choose one channel for split operation.")
return True, None
|
python
|
def selection_error_control(self, form_info):
"""
It controls the selection from the form according
to the operations, and returns an error message
if it does not comply with the rules.
Args:
form_info: Channel or subscriber form from the user
Returns: True or False
error message
"""
keys, names = self.return_selected_form_items(form_info['ChannelList'])
chosen_channels_number = len(keys)
if form_info['new_channel'] and chosen_channels_number < 2:
return False, _(
u"You should choose at least two channel to merge operation at a new channel.")
elif form_info['existing_channel'] and chosen_channels_number == 0:
return False, _(
u"You should choose at least one channel to merge operation with existing channel.")
elif form_info['find_chosen_channel'] and chosen_channels_number != 1:
return False, _(u"You should choose one channel for split operation.")
return True, None
|
[
"def",
"selection_error_control",
"(",
"self",
",",
"form_info",
")",
":",
"keys",
",",
"names",
"=",
"self",
".",
"return_selected_form_items",
"(",
"form_info",
"[",
"'ChannelList'",
"]",
")",
"chosen_channels_number",
"=",
"len",
"(",
"keys",
")",
"if",
"form_info",
"[",
"'new_channel'",
"]",
"and",
"chosen_channels_number",
"<",
"2",
":",
"return",
"False",
",",
"_",
"(",
"u\"You should choose at least two channel to merge operation at a new channel.\"",
")",
"elif",
"form_info",
"[",
"'existing_channel'",
"]",
"and",
"chosen_channels_number",
"==",
"0",
":",
"return",
"False",
",",
"_",
"(",
"u\"You should choose at least one channel to merge operation with existing channel.\"",
")",
"elif",
"form_info",
"[",
"'find_chosen_channel'",
"]",
"and",
"chosen_channels_number",
"!=",
"1",
":",
"return",
"False",
",",
"_",
"(",
"u\"You should choose one channel for split operation.\"",
")",
"return",
"True",
",",
"None"
] |
It controls the selection from the form according
to the operations, and returns an error message
if it does not comply with the rules.
Args:
form_info: Channel or subscriber form from the user
Returns: True or False
error message
|
[
"It",
"controls",
"the",
"selection",
"from",
"the",
"form",
"according",
"to",
"the",
"operations",
"and",
"returns",
"an",
"error",
"message",
"if",
"it",
"does",
"not",
"comply",
"with",
"the",
"rules",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/channel_management.py#L289-L314
|
train
|
cimm-kzn/CGRtools
|
CGRtools/algorithms/morgan.py
|
_eratosthenes
|
def _eratosthenes():
"""Yields the sequence of prime numbers via the Sieve of Eratosthenes."""
d = {} # map each composite integer to its first-found prime factor
for q in count(2): # q gets 2, 3, 4, 5, ... ad infinitum
p = d.pop(q, None)
if p is None:
# q not a key in D, so q is prime, therefore, yield it
yield q
# mark q squared as not-prime (with q as first-found prime factor)
d[q * q] = q
else:
# let x <- smallest (N*p)+q which wasn't yet known to be composite
# we just learned x is composite, with p first-found prime factor,
# since p is the first-found prime factor of q -- find and mark it
x = p + q
while x in d:
x += p
d[x] = p
|
python
|
def _eratosthenes():
"""Yields the sequence of prime numbers via the Sieve of Eratosthenes."""
d = {} # map each composite integer to its first-found prime factor
for q in count(2): # q gets 2, 3, 4, 5, ... ad infinitum
p = d.pop(q, None)
if p is None:
# q not a key in D, so q is prime, therefore, yield it
yield q
# mark q squared as not-prime (with q as first-found prime factor)
d[q * q] = q
else:
# let x <- smallest (N*p)+q which wasn't yet known to be composite
# we just learned x is composite, with p first-found prime factor,
# since p is the first-found prime factor of q -- find and mark it
x = p + q
while x in d:
x += p
d[x] = p
|
[
"def",
"_eratosthenes",
"(",
")",
":",
"d",
"=",
"{",
"}",
"# map each composite integer to its first-found prime factor",
"for",
"q",
"in",
"count",
"(",
"2",
")",
":",
"# q gets 2, 3, 4, 5, ... ad infinitum",
"p",
"=",
"d",
".",
"pop",
"(",
"q",
",",
"None",
")",
"if",
"p",
"is",
"None",
":",
"# q not a key in D, so q is prime, therefore, yield it",
"yield",
"q",
"# mark q squared as not-prime (with q as first-found prime factor)",
"d",
"[",
"q",
"*",
"q",
"]",
"=",
"q",
"else",
":",
"# let x <- smallest (N*p)+q which wasn't yet known to be composite",
"# we just learned x is composite, with p first-found prime factor,",
"# since p is the first-found prime factor of q -- find and mark it",
"x",
"=",
"p",
"+",
"q",
"while",
"x",
"in",
"d",
":",
"x",
"+=",
"p",
"d",
"[",
"x",
"]",
"=",
"p"
] |
Yields the sequence of prime numbers via the Sieve of Eratosthenes.
|
[
"Yields",
"the",
"sequence",
"of",
"prime",
"numbers",
"via",
"the",
"Sieve",
"of",
"Eratosthenes",
"."
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/algorithms/morgan.py#L88-L105
|
train
|
cimm-kzn/CGRtools
|
CGRtools/algorithms/morgan.py
|
Morgan.atoms_order
|
def atoms_order(self):
"""
Morgan like algorithm for graph nodes ordering
:return: dict of atom-weight pairs
"""
if not len(self): # for empty containers
return {}
elif len(self) == 1: # optimize single atom containers
return dict.fromkeys(self, 2)
params = {n: (int(node), tuple(sorted(int(edge) for edge in self._adj[n].values())))
for n, node in self.atoms()}
newlevels = {}
countprime = iter(primes)
weights = {x: newlevels.get(y) or newlevels.setdefault(y, next(countprime))
for x, y in sorted(params.items(), key=itemgetter(1))}
tries = len(self) * 4
numb = len(set(weights.values()))
stab = 0
while tries:
oldnumb = numb
neweights = {}
countprime = iter(primes)
# weights[n] ** 2 NEED for differentiation of molecules like A-B or any other complete graphs.
tmp = {n: reduce(mul, (weights[x] for x in m), weights[n] ** 2) for n, m in self._adj.items()}
weights = {x: (neweights.get(y) or neweights.setdefault(y, next(countprime)))
for x, y in sorted(tmp.items(), key=itemgetter(1))}
numb = len(set(weights.values()))
if numb == len(self): # each atom now unique
break
elif numb == oldnumb:
x = Counter(weights.values())
if x[min(x)] > 1:
if stab == 3:
break
elif stab >= 2:
break
stab += 1
elif stab:
stab = 0
tries -= 1
if not tries and numb < oldnumb:
warning('morgan. number of attempts exceeded. uniqueness has decreased. next attempt will be made')
tries = 1
else:
warning('morgan. number of attempts exceeded')
return weights
|
python
|
def atoms_order(self):
"""
Morgan like algorithm for graph nodes ordering
:return: dict of atom-weight pairs
"""
if not len(self): # for empty containers
return {}
elif len(self) == 1: # optimize single atom containers
return dict.fromkeys(self, 2)
params = {n: (int(node), tuple(sorted(int(edge) for edge in self._adj[n].values())))
for n, node in self.atoms()}
newlevels = {}
countprime = iter(primes)
weights = {x: newlevels.get(y) or newlevels.setdefault(y, next(countprime))
for x, y in sorted(params.items(), key=itemgetter(1))}
tries = len(self) * 4
numb = len(set(weights.values()))
stab = 0
while tries:
oldnumb = numb
neweights = {}
countprime = iter(primes)
# weights[n] ** 2 NEED for differentiation of molecules like A-B or any other complete graphs.
tmp = {n: reduce(mul, (weights[x] for x in m), weights[n] ** 2) for n, m in self._adj.items()}
weights = {x: (neweights.get(y) or neweights.setdefault(y, next(countprime)))
for x, y in sorted(tmp.items(), key=itemgetter(1))}
numb = len(set(weights.values()))
if numb == len(self): # each atom now unique
break
elif numb == oldnumb:
x = Counter(weights.values())
if x[min(x)] > 1:
if stab == 3:
break
elif stab >= 2:
break
stab += 1
elif stab:
stab = 0
tries -= 1
if not tries and numb < oldnumb:
warning('morgan. number of attempts exceeded. uniqueness has decreased. next attempt will be made')
tries = 1
else:
warning('morgan. number of attempts exceeded')
return weights
|
[
"def",
"atoms_order",
"(",
"self",
")",
":",
"if",
"not",
"len",
"(",
"self",
")",
":",
"# for empty containers",
"return",
"{",
"}",
"elif",
"len",
"(",
"self",
")",
"==",
"1",
":",
"# optimize single atom containers",
"return",
"dict",
".",
"fromkeys",
"(",
"self",
",",
"2",
")",
"params",
"=",
"{",
"n",
":",
"(",
"int",
"(",
"node",
")",
",",
"tuple",
"(",
"sorted",
"(",
"int",
"(",
"edge",
")",
"for",
"edge",
"in",
"self",
".",
"_adj",
"[",
"n",
"]",
".",
"values",
"(",
")",
")",
")",
")",
"for",
"n",
",",
"node",
"in",
"self",
".",
"atoms",
"(",
")",
"}",
"newlevels",
"=",
"{",
"}",
"countprime",
"=",
"iter",
"(",
"primes",
")",
"weights",
"=",
"{",
"x",
":",
"newlevels",
".",
"get",
"(",
"y",
")",
"or",
"newlevels",
".",
"setdefault",
"(",
"y",
",",
"next",
"(",
"countprime",
")",
")",
"for",
"x",
",",
"y",
"in",
"sorted",
"(",
"params",
".",
"items",
"(",
")",
",",
"key",
"=",
"itemgetter",
"(",
"1",
")",
")",
"}",
"tries",
"=",
"len",
"(",
"self",
")",
"*",
"4",
"numb",
"=",
"len",
"(",
"set",
"(",
"weights",
".",
"values",
"(",
")",
")",
")",
"stab",
"=",
"0",
"while",
"tries",
":",
"oldnumb",
"=",
"numb",
"neweights",
"=",
"{",
"}",
"countprime",
"=",
"iter",
"(",
"primes",
")",
"# weights[n] ** 2 NEED for differentiation of molecules like A-B or any other complete graphs.",
"tmp",
"=",
"{",
"n",
":",
"reduce",
"(",
"mul",
",",
"(",
"weights",
"[",
"x",
"]",
"for",
"x",
"in",
"m",
")",
",",
"weights",
"[",
"n",
"]",
"**",
"2",
")",
"for",
"n",
",",
"m",
"in",
"self",
".",
"_adj",
".",
"items",
"(",
")",
"}",
"weights",
"=",
"{",
"x",
":",
"(",
"neweights",
".",
"get",
"(",
"y",
")",
"or",
"neweights",
".",
"setdefault",
"(",
"y",
",",
"next",
"(",
"countprime",
")",
")",
")",
"for",
"x",
",",
"y",
"in",
"sorted",
"(",
"tmp",
".",
"items",
"(",
")",
",",
"key",
"=",
"itemgetter",
"(",
"1",
")",
")",
"}",
"numb",
"=",
"len",
"(",
"set",
"(",
"weights",
".",
"values",
"(",
")",
")",
")",
"if",
"numb",
"==",
"len",
"(",
"self",
")",
":",
"# each atom now unique",
"break",
"elif",
"numb",
"==",
"oldnumb",
":",
"x",
"=",
"Counter",
"(",
"weights",
".",
"values",
"(",
")",
")",
"if",
"x",
"[",
"min",
"(",
"x",
")",
"]",
">",
"1",
":",
"if",
"stab",
"==",
"3",
":",
"break",
"elif",
"stab",
">=",
"2",
":",
"break",
"stab",
"+=",
"1",
"elif",
"stab",
":",
"stab",
"=",
"0",
"tries",
"-=",
"1",
"if",
"not",
"tries",
"and",
"numb",
"<",
"oldnumb",
":",
"warning",
"(",
"'morgan. number of attempts exceeded. uniqueness has decreased. next attempt will be made'",
")",
"tries",
"=",
"1",
"else",
":",
"warning",
"(",
"'morgan. number of attempts exceeded'",
")",
"return",
"weights"
] |
Morgan like algorithm for graph nodes ordering
:return: dict of atom-weight pairs
|
[
"Morgan",
"like",
"algorithm",
"for",
"graph",
"nodes",
"ordering"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/algorithms/morgan.py#L29-L85
|
train
|
LordDarkula/chess_py
|
chess_py/pieces/piece_const.py
|
PieceValues.init_manual
|
def init_manual(cls, pawn_value, knight_value, bishop_value, rook_value, queen_value, king_value):
"""
Manual init method for external piece values
:type: PAWN_VALUE: int
:type: KNIGHT_VALUE: int
:type: BISHOP_VALUE: int
:type: ROOK_VALUE: int
:type: QUEEN_VALUE: int
"""
piece_values = cls()
piece_values.PAWN_VALUE = pawn_value
piece_values.KNIGHT_VALUE = knight_value
piece_values.BISHOP_VALUE = bishop_value
piece_values.ROOK_VALUE = rook_value
piece_values.QUEEN_VALUE = queen_value
piece_values.KING_VALUE = king_value
return piece_values
|
python
|
def init_manual(cls, pawn_value, knight_value, bishop_value, rook_value, queen_value, king_value):
"""
Manual init method for external piece values
:type: PAWN_VALUE: int
:type: KNIGHT_VALUE: int
:type: BISHOP_VALUE: int
:type: ROOK_VALUE: int
:type: QUEEN_VALUE: int
"""
piece_values = cls()
piece_values.PAWN_VALUE = pawn_value
piece_values.KNIGHT_VALUE = knight_value
piece_values.BISHOP_VALUE = bishop_value
piece_values.ROOK_VALUE = rook_value
piece_values.QUEEN_VALUE = queen_value
piece_values.KING_VALUE = king_value
return piece_values
|
[
"def",
"init_manual",
"(",
"cls",
",",
"pawn_value",
",",
"knight_value",
",",
"bishop_value",
",",
"rook_value",
",",
"queen_value",
",",
"king_value",
")",
":",
"piece_values",
"=",
"cls",
"(",
")",
"piece_values",
".",
"PAWN_VALUE",
"=",
"pawn_value",
"piece_values",
".",
"KNIGHT_VALUE",
"=",
"knight_value",
"piece_values",
".",
"BISHOP_VALUE",
"=",
"bishop_value",
"piece_values",
".",
"ROOK_VALUE",
"=",
"rook_value",
"piece_values",
".",
"QUEEN_VALUE",
"=",
"queen_value",
"piece_values",
".",
"KING_VALUE",
"=",
"king_value",
"return",
"piece_values"
] |
Manual init method for external piece values
:type: PAWN_VALUE: int
:type: KNIGHT_VALUE: int
:type: BISHOP_VALUE: int
:type: ROOK_VALUE: int
:type: QUEEN_VALUE: int
|
[
"Manual",
"init",
"method",
"for",
"external",
"piece",
"values"
] |
14bebc2f8c49ae25c59375cc83d0b38d8ff7281d
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/pieces/piece_const.py#L27-L44
|
train
|
LordDarkula/chess_py
|
chess_py/pieces/piece_const.py
|
PieceValues.val
|
def val(self, piece, ref_color):
"""
Finds value of ``Piece``
:type: piece: Piece
:type: ref_color: Color
:rtype: int
"""
if piece is None:
return 0
if ref_color == piece.color:
const = 1
else:
const = -1
if isinstance(piece, Pawn):
return self.PAWN_VALUE * const
elif isinstance(piece, Queen):
return self.QUEEN_VALUE * const
elif isinstance(piece, Bishop):
return self.BISHOP_VALUE * const
elif isinstance(piece, Rook):
return self.ROOK_VALUE * const
elif isinstance(piece, Knight):
return self.KNIGHT_VALUE * const
elif isinstance(piece, King):
return self.KING_VALUE * const
return 0
|
python
|
def val(self, piece, ref_color):
"""
Finds value of ``Piece``
:type: piece: Piece
:type: ref_color: Color
:rtype: int
"""
if piece is None:
return 0
if ref_color == piece.color:
const = 1
else:
const = -1
if isinstance(piece, Pawn):
return self.PAWN_VALUE * const
elif isinstance(piece, Queen):
return self.QUEEN_VALUE * const
elif isinstance(piece, Bishop):
return self.BISHOP_VALUE * const
elif isinstance(piece, Rook):
return self.ROOK_VALUE * const
elif isinstance(piece, Knight):
return self.KNIGHT_VALUE * const
elif isinstance(piece, King):
return self.KING_VALUE * const
return 0
|
[
"def",
"val",
"(",
"self",
",",
"piece",
",",
"ref_color",
")",
":",
"if",
"piece",
"is",
"None",
":",
"return",
"0",
"if",
"ref_color",
"==",
"piece",
".",
"color",
":",
"const",
"=",
"1",
"else",
":",
"const",
"=",
"-",
"1",
"if",
"isinstance",
"(",
"piece",
",",
"Pawn",
")",
":",
"return",
"self",
".",
"PAWN_VALUE",
"*",
"const",
"elif",
"isinstance",
"(",
"piece",
",",
"Queen",
")",
":",
"return",
"self",
".",
"QUEEN_VALUE",
"*",
"const",
"elif",
"isinstance",
"(",
"piece",
",",
"Bishop",
")",
":",
"return",
"self",
".",
"BISHOP_VALUE",
"*",
"const",
"elif",
"isinstance",
"(",
"piece",
",",
"Rook",
")",
":",
"return",
"self",
".",
"ROOK_VALUE",
"*",
"const",
"elif",
"isinstance",
"(",
"piece",
",",
"Knight",
")",
":",
"return",
"self",
".",
"KNIGHT_VALUE",
"*",
"const",
"elif",
"isinstance",
"(",
"piece",
",",
"King",
")",
":",
"return",
"self",
".",
"KING_VALUE",
"*",
"const",
"return",
"0"
] |
Finds value of ``Piece``
:type: piece: Piece
:type: ref_color: Color
:rtype: int
|
[
"Finds",
"value",
"of",
"Piece"
] |
14bebc2f8c49ae25c59375cc83d0b38d8ff7281d
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/pieces/piece_const.py#L46-L74
|
train
|
LordDarkula/chess_py
|
chess_py/game/game.py
|
Game.play
|
def play(self):
"""
Starts game and returns one of 3 results .
Iterates between methods ``white_move()`` and
``black_move()`` until game ends. Each
method calls the respective player's ``generate_move()``
method.
:rtype: int
"""
colors = [lambda: self.white_move(), lambda: self.black_move()]
colors = itertools.cycle(colors)
while True:
color_fn = next(colors)
if game_state.no_moves(self.position):
if self.position.get_king(color.white).in_check(self.position):
return 1
elif self.position.get_king(color.black).in_check(self.position):
return 0
else:
return 0.5
color_fn()
|
python
|
def play(self):
"""
Starts game and returns one of 3 results .
Iterates between methods ``white_move()`` and
``black_move()`` until game ends. Each
method calls the respective player's ``generate_move()``
method.
:rtype: int
"""
colors = [lambda: self.white_move(), lambda: self.black_move()]
colors = itertools.cycle(colors)
while True:
color_fn = next(colors)
if game_state.no_moves(self.position):
if self.position.get_king(color.white).in_check(self.position):
return 1
elif self.position.get_king(color.black).in_check(self.position):
return 0
else:
return 0.5
color_fn()
|
[
"def",
"play",
"(",
"self",
")",
":",
"colors",
"=",
"[",
"lambda",
":",
"self",
".",
"white_move",
"(",
")",
",",
"lambda",
":",
"self",
".",
"black_move",
"(",
")",
"]",
"colors",
"=",
"itertools",
".",
"cycle",
"(",
"colors",
")",
"while",
"True",
":",
"color_fn",
"=",
"next",
"(",
"colors",
")",
"if",
"game_state",
".",
"no_moves",
"(",
"self",
".",
"position",
")",
":",
"if",
"self",
".",
"position",
".",
"get_king",
"(",
"color",
".",
"white",
")",
".",
"in_check",
"(",
"self",
".",
"position",
")",
":",
"return",
"1",
"elif",
"self",
".",
"position",
".",
"get_king",
"(",
"color",
".",
"black",
")",
".",
"in_check",
"(",
"self",
".",
"position",
")",
":",
"return",
"0",
"else",
":",
"return",
"0.5",
"color_fn",
"(",
")"
] |
Starts game and returns one of 3 results .
Iterates between methods ``white_move()`` and
``black_move()`` until game ends. Each
method calls the respective player's ``generate_move()``
method.
:rtype: int
|
[
"Starts",
"game",
"and",
"returns",
"one",
"of",
"3",
"results",
".",
"Iterates",
"between",
"methods",
"white_move",
"()",
"and",
"black_move",
"()",
"until",
"game",
"ends",
".",
"Each",
"method",
"calls",
"the",
"respective",
"player",
"s",
"generate_move",
"()",
"method",
"."
] |
14bebc2f8c49ae25c59375cc83d0b38d8ff7281d
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game.py#L39-L64
|
train
|
LordDarkula/chess_py
|
chess_py/game/game.py
|
Game.white_move
|
def white_move(self):
"""
Calls the white player's ``generate_move()``
method and updates the board with the move returned.
"""
move = self.player_white.generate_move(self.position)
move = make_legal(move, self.position)
self.position.update(move)
|
python
|
def white_move(self):
"""
Calls the white player's ``generate_move()``
method and updates the board with the move returned.
"""
move = self.player_white.generate_move(self.position)
move = make_legal(move, self.position)
self.position.update(move)
|
[
"def",
"white_move",
"(",
"self",
")",
":",
"move",
"=",
"self",
".",
"player_white",
".",
"generate_move",
"(",
"self",
".",
"position",
")",
"move",
"=",
"make_legal",
"(",
"move",
",",
"self",
".",
"position",
")",
"self",
".",
"position",
".",
"update",
"(",
"move",
")"
] |
Calls the white player's ``generate_move()``
method and updates the board with the move returned.
|
[
"Calls",
"the",
"white",
"player",
"s",
"generate_move",
"()",
"method",
"and",
"updates",
"the",
"board",
"with",
"the",
"move",
"returned",
"."
] |
14bebc2f8c49ae25c59375cc83d0b38d8ff7281d
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game.py#L66-L73
|
train
|
LordDarkula/chess_py
|
chess_py/game/game.py
|
Game.black_move
|
def black_move(self):
"""
Calls the black player's ``generate_move()``
method and updates the board with the move returned.
"""
move = self.player_black.generate_move(self.position)
move = make_legal(move, self.position)
self.position.update(move)
|
python
|
def black_move(self):
"""
Calls the black player's ``generate_move()``
method and updates the board with the move returned.
"""
move = self.player_black.generate_move(self.position)
move = make_legal(move, self.position)
self.position.update(move)
|
[
"def",
"black_move",
"(",
"self",
")",
":",
"move",
"=",
"self",
".",
"player_black",
".",
"generate_move",
"(",
"self",
".",
"position",
")",
"move",
"=",
"make_legal",
"(",
"move",
",",
"self",
".",
"position",
")",
"self",
".",
"position",
".",
"update",
"(",
"move",
")"
] |
Calls the black player's ``generate_move()``
method and updates the board with the move returned.
|
[
"Calls",
"the",
"black",
"player",
"s",
"generate_move",
"()",
"method",
"and",
"updates",
"the",
"board",
"with",
"the",
"move",
"returned",
"."
] |
14bebc2f8c49ae25c59375cc83d0b38d8ff7281d
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game.py#L75-L82
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.get_field_cache
|
def get_field_cache(self, cache_type='es'):
"""Return a list of fields' mappings"""
if cache_type == 'kibana':
try:
search_results = urlopen(self.get_url).read().decode('utf-8')
except HTTPError: # as e:
# self.pr_err("get_field_cache(kibana), HTTPError: %s" % e)
return []
index_pattern = json.loads(search_results)
# Results look like: {"_index":".kibana","_type":"index-pattern","_id":"aaa*","_version":6,"found":true,"_source":{"title":"aaa*","fields":"<what we want>"}} # noqa
fields_str = index_pattern['_source']['fields']
return json.loads(fields_str)
elif cache_type == 'es' or cache_type.startswith('elastic'):
search_results = urlopen(self.es_get_url).read().decode('utf-8')
es_mappings = json.loads(search_results)
# Results look like: {"<index_name>":{"mappings":{"<doc_type>":{"<field_name>":{"full_name":"<field_name>","mapping":{"<sub-field_name>":{"type":"date","index_name":"<sub-field_name>","boost":1.0,"index":"not_analyzed","store":false,"doc_values":false,"term_vector":"no","norms":{"enabled":false},"index_options":"docs","index_analyzer":"_date/16","search_analyzer":"_date/max","postings_format":"default","doc_values_format":"default","similarity":"default","fielddata":{},"ignore_malformed":false,"coerce":true,"precision_step":16,"format":"dateOptionalTime","null_value":null,"include_in_all":false,"numeric_resolution":"milliseconds","locale":""}}}, # noqa
# now convert the mappings into the .kibana format
field_cache = []
for (index_name, val) in iteritems(es_mappings):
if index_name != self.index: # only get non-'.kibana' indices
# self.pr_dbg("index: %s" % index_name)
m_dict = es_mappings[index_name]['mappings']
# self.pr_dbg('m_dict %s' % m_dict)
mappings = self.get_index_mappings(m_dict)
# self.pr_dbg('mappings %s' % mappings)
field_cache.extend(mappings)
field_cache = self.dedup_field_cache(field_cache)
return field_cache
self.pr_err("Unknown cache type: %s" % cache_type)
return None
|
python
|
def get_field_cache(self, cache_type='es'):
"""Return a list of fields' mappings"""
if cache_type == 'kibana':
try:
search_results = urlopen(self.get_url).read().decode('utf-8')
except HTTPError: # as e:
# self.pr_err("get_field_cache(kibana), HTTPError: %s" % e)
return []
index_pattern = json.loads(search_results)
# Results look like: {"_index":".kibana","_type":"index-pattern","_id":"aaa*","_version":6,"found":true,"_source":{"title":"aaa*","fields":"<what we want>"}} # noqa
fields_str = index_pattern['_source']['fields']
return json.loads(fields_str)
elif cache_type == 'es' or cache_type.startswith('elastic'):
search_results = urlopen(self.es_get_url).read().decode('utf-8')
es_mappings = json.loads(search_results)
# Results look like: {"<index_name>":{"mappings":{"<doc_type>":{"<field_name>":{"full_name":"<field_name>","mapping":{"<sub-field_name>":{"type":"date","index_name":"<sub-field_name>","boost":1.0,"index":"not_analyzed","store":false,"doc_values":false,"term_vector":"no","norms":{"enabled":false},"index_options":"docs","index_analyzer":"_date/16","search_analyzer":"_date/max","postings_format":"default","doc_values_format":"default","similarity":"default","fielddata":{},"ignore_malformed":false,"coerce":true,"precision_step":16,"format":"dateOptionalTime","null_value":null,"include_in_all":false,"numeric_resolution":"milliseconds","locale":""}}}, # noqa
# now convert the mappings into the .kibana format
field_cache = []
for (index_name, val) in iteritems(es_mappings):
if index_name != self.index: # only get non-'.kibana' indices
# self.pr_dbg("index: %s" % index_name)
m_dict = es_mappings[index_name]['mappings']
# self.pr_dbg('m_dict %s' % m_dict)
mappings = self.get_index_mappings(m_dict)
# self.pr_dbg('mappings %s' % mappings)
field_cache.extend(mappings)
field_cache = self.dedup_field_cache(field_cache)
return field_cache
self.pr_err("Unknown cache type: %s" % cache_type)
return None
|
[
"def",
"get_field_cache",
"(",
"self",
",",
"cache_type",
"=",
"'es'",
")",
":",
"if",
"cache_type",
"==",
"'kibana'",
":",
"try",
":",
"search_results",
"=",
"urlopen",
"(",
"self",
".",
"get_url",
")",
".",
"read",
"(",
")",
".",
"decode",
"(",
"'utf-8'",
")",
"except",
"HTTPError",
":",
"# as e:",
"# self.pr_err(\"get_field_cache(kibana), HTTPError: %s\" % e)",
"return",
"[",
"]",
"index_pattern",
"=",
"json",
".",
"loads",
"(",
"search_results",
")",
"# Results look like: {\"_index\":\".kibana\",\"_type\":\"index-pattern\",\"_id\":\"aaa*\",\"_version\":6,\"found\":true,\"_source\":{\"title\":\"aaa*\",\"fields\":\"<what we want>\"}} # noqa",
"fields_str",
"=",
"index_pattern",
"[",
"'_source'",
"]",
"[",
"'fields'",
"]",
"return",
"json",
".",
"loads",
"(",
"fields_str",
")",
"elif",
"cache_type",
"==",
"'es'",
"or",
"cache_type",
".",
"startswith",
"(",
"'elastic'",
")",
":",
"search_results",
"=",
"urlopen",
"(",
"self",
".",
"es_get_url",
")",
".",
"read",
"(",
")",
".",
"decode",
"(",
"'utf-8'",
")",
"es_mappings",
"=",
"json",
".",
"loads",
"(",
"search_results",
")",
"# Results look like: {\"<index_name>\":{\"mappings\":{\"<doc_type>\":{\"<field_name>\":{\"full_name\":\"<field_name>\",\"mapping\":{\"<sub-field_name>\":{\"type\":\"date\",\"index_name\":\"<sub-field_name>\",\"boost\":1.0,\"index\":\"not_analyzed\",\"store\":false,\"doc_values\":false,\"term_vector\":\"no\",\"norms\":{\"enabled\":false},\"index_options\":\"docs\",\"index_analyzer\":\"_date/16\",\"search_analyzer\":\"_date/max\",\"postings_format\":\"default\",\"doc_values_format\":\"default\",\"similarity\":\"default\",\"fielddata\":{},\"ignore_malformed\":false,\"coerce\":true,\"precision_step\":16,\"format\":\"dateOptionalTime\",\"null_value\":null,\"include_in_all\":false,\"numeric_resolution\":\"milliseconds\",\"locale\":\"\"}}}, # noqa",
"# now convert the mappings into the .kibana format",
"field_cache",
"=",
"[",
"]",
"for",
"(",
"index_name",
",",
"val",
")",
"in",
"iteritems",
"(",
"es_mappings",
")",
":",
"if",
"index_name",
"!=",
"self",
".",
"index",
":",
"# only get non-'.kibana' indices",
"# self.pr_dbg(\"index: %s\" % index_name)",
"m_dict",
"=",
"es_mappings",
"[",
"index_name",
"]",
"[",
"'mappings'",
"]",
"# self.pr_dbg('m_dict %s' % m_dict)",
"mappings",
"=",
"self",
".",
"get_index_mappings",
"(",
"m_dict",
")",
"# self.pr_dbg('mappings %s' % mappings)",
"field_cache",
".",
"extend",
"(",
"mappings",
")",
"field_cache",
"=",
"self",
".",
"dedup_field_cache",
"(",
"field_cache",
")",
"return",
"field_cache",
"self",
".",
"pr_err",
"(",
"\"Unknown cache type: %s\"",
"%",
"cache_type",
")",
"return",
"None"
] |
Return a list of fields' mappings
|
[
"Return",
"a",
"list",
"of",
"fields",
"mappings"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L96-L125
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.post_field_cache
|
def post_field_cache(self, field_cache):
"""Where field_cache is a list of fields' mappings"""
index_pattern = self.field_cache_to_index_pattern(field_cache)
# self.pr_dbg("request/post: %s" % index_pattern)
resp = requests.post(self.post_url, data=index_pattern).text
# resp = {"_index":".kibana","_type":"index-pattern","_id":"aaa*","_version":1,"created":true} # noqa
resp = json.loads(resp)
return 0
|
python
|
def post_field_cache(self, field_cache):
"""Where field_cache is a list of fields' mappings"""
index_pattern = self.field_cache_to_index_pattern(field_cache)
# self.pr_dbg("request/post: %s" % index_pattern)
resp = requests.post(self.post_url, data=index_pattern).text
# resp = {"_index":".kibana","_type":"index-pattern","_id":"aaa*","_version":1,"created":true} # noqa
resp = json.loads(resp)
return 0
|
[
"def",
"post_field_cache",
"(",
"self",
",",
"field_cache",
")",
":",
"index_pattern",
"=",
"self",
".",
"field_cache_to_index_pattern",
"(",
"field_cache",
")",
"# self.pr_dbg(\"request/post: %s\" % index_pattern)",
"resp",
"=",
"requests",
".",
"post",
"(",
"self",
".",
"post_url",
",",
"data",
"=",
"index_pattern",
")",
".",
"text",
"# resp = {\"_index\":\".kibana\",\"_type\":\"index-pattern\",\"_id\":\"aaa*\",\"_version\":1,\"created\":true} # noqa",
"resp",
"=",
"json",
".",
"loads",
"(",
"resp",
")",
"return",
"0"
] |
Where field_cache is a list of fields' mappings
|
[
"Where",
"field_cache",
"is",
"a",
"list",
"of",
"fields",
"mappings"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L142-L149
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.field_cache_to_index_pattern
|
def field_cache_to_index_pattern(self, field_cache):
"""Return a .kibana index-pattern doc_type"""
mapping_dict = {}
mapping_dict['customFormats'] = "{}"
mapping_dict['title'] = self.index_pattern
# now post the data into .kibana
mapping_dict['fields'] = json.dumps(field_cache, separators=(',', ':'))
# in order to post, we need to create the post string
mapping_str = json.dumps(mapping_dict, separators=(',', ':'))
return mapping_str
|
python
|
def field_cache_to_index_pattern(self, field_cache):
"""Return a .kibana index-pattern doc_type"""
mapping_dict = {}
mapping_dict['customFormats'] = "{}"
mapping_dict['title'] = self.index_pattern
# now post the data into .kibana
mapping_dict['fields'] = json.dumps(field_cache, separators=(',', ':'))
# in order to post, we need to create the post string
mapping_str = json.dumps(mapping_dict, separators=(',', ':'))
return mapping_str
|
[
"def",
"field_cache_to_index_pattern",
"(",
"self",
",",
"field_cache",
")",
":",
"mapping_dict",
"=",
"{",
"}",
"mapping_dict",
"[",
"'customFormats'",
"]",
"=",
"\"{}\"",
"mapping_dict",
"[",
"'title'",
"]",
"=",
"self",
".",
"index_pattern",
"# now post the data into .kibana",
"mapping_dict",
"[",
"'fields'",
"]",
"=",
"json",
".",
"dumps",
"(",
"field_cache",
",",
"separators",
"=",
"(",
"','",
",",
"':'",
")",
")",
"# in order to post, we need to create the post string",
"mapping_str",
"=",
"json",
".",
"dumps",
"(",
"mapping_dict",
",",
"separators",
"=",
"(",
"','",
",",
"':'",
")",
")",
"return",
"mapping_str"
] |
Return a .kibana index-pattern doc_type
|
[
"Return",
"a",
".",
"kibana",
"index",
"-",
"pattern",
"doc_type"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L152-L161
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.check_mapping
|
def check_mapping(self, m):
"""Assert minimum set of fields in cache, does not validate contents"""
if 'name' not in m:
self.pr_dbg("Missing %s" % "name")
return False
# self.pr_dbg("Checking %s" % m['name'])
for x in ['analyzed', 'indexed', 'type', 'scripted', 'count']:
if x not in m or m[x] == "":
self.pr_dbg("Missing %s" % x)
self.pr_dbg("Full %s" % m)
return False
if 'doc_values' not in m or m['doc_values'] == "":
if not m['name'].startswith('_'):
self.pr_dbg("Missing %s" % "doc_values")
return False
m['doc_values'] = False
return True
|
python
|
def check_mapping(self, m):
"""Assert minimum set of fields in cache, does not validate contents"""
if 'name' not in m:
self.pr_dbg("Missing %s" % "name")
return False
# self.pr_dbg("Checking %s" % m['name'])
for x in ['analyzed', 'indexed', 'type', 'scripted', 'count']:
if x not in m or m[x] == "":
self.pr_dbg("Missing %s" % x)
self.pr_dbg("Full %s" % m)
return False
if 'doc_values' not in m or m['doc_values'] == "":
if not m['name'].startswith('_'):
self.pr_dbg("Missing %s" % "doc_values")
return False
m['doc_values'] = False
return True
|
[
"def",
"check_mapping",
"(",
"self",
",",
"m",
")",
":",
"if",
"'name'",
"not",
"in",
"m",
":",
"self",
".",
"pr_dbg",
"(",
"\"Missing %s\"",
"%",
"\"name\"",
")",
"return",
"False",
"# self.pr_dbg(\"Checking %s\" % m['name'])",
"for",
"x",
"in",
"[",
"'analyzed'",
",",
"'indexed'",
",",
"'type'",
",",
"'scripted'",
",",
"'count'",
"]",
":",
"if",
"x",
"not",
"in",
"m",
"or",
"m",
"[",
"x",
"]",
"==",
"\"\"",
":",
"self",
".",
"pr_dbg",
"(",
"\"Missing %s\"",
"%",
"x",
")",
"self",
".",
"pr_dbg",
"(",
"\"Full %s\"",
"%",
"m",
")",
"return",
"False",
"if",
"'doc_values'",
"not",
"in",
"m",
"or",
"m",
"[",
"'doc_values'",
"]",
"==",
"\"\"",
":",
"if",
"not",
"m",
"[",
"'name'",
"]",
".",
"startswith",
"(",
"'_'",
")",
":",
"self",
".",
"pr_dbg",
"(",
"\"Missing %s\"",
"%",
"\"doc_values\"",
")",
"return",
"False",
"m",
"[",
"'doc_values'",
"]",
"=",
"False",
"return",
"True"
] |
Assert minimum set of fields in cache, does not validate contents
|
[
"Assert",
"minimum",
"set",
"of",
"fields",
"in",
"cache",
"does",
"not",
"validate",
"contents"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L163-L179
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.get_index_mappings
|
def get_index_mappings(self, index):
"""Converts all index's doc_types to .kibana"""
fields_arr = []
for (key, val) in iteritems(index):
# self.pr_dbg("\tdoc_type: %s" % key)
doc_mapping = self.get_doc_type_mappings(index[key])
# self.pr_dbg("\tdoc_mapping: %s" % doc_mapping)
if doc_mapping is None:
return None
# keep adding to the fields array
fields_arr.extend(doc_mapping)
return fields_arr
|
python
|
def get_index_mappings(self, index):
"""Converts all index's doc_types to .kibana"""
fields_arr = []
for (key, val) in iteritems(index):
# self.pr_dbg("\tdoc_type: %s" % key)
doc_mapping = self.get_doc_type_mappings(index[key])
# self.pr_dbg("\tdoc_mapping: %s" % doc_mapping)
if doc_mapping is None:
return None
# keep adding to the fields array
fields_arr.extend(doc_mapping)
return fields_arr
|
[
"def",
"get_index_mappings",
"(",
"self",
",",
"index",
")",
":",
"fields_arr",
"=",
"[",
"]",
"for",
"(",
"key",
",",
"val",
")",
"in",
"iteritems",
"(",
"index",
")",
":",
"# self.pr_dbg(\"\\tdoc_type: %s\" % key)",
"doc_mapping",
"=",
"self",
".",
"get_doc_type_mappings",
"(",
"index",
"[",
"key",
"]",
")",
"# self.pr_dbg(\"\\tdoc_mapping: %s\" % doc_mapping)",
"if",
"doc_mapping",
"is",
"None",
":",
"return",
"None",
"# keep adding to the fields array",
"fields_arr",
".",
"extend",
"(",
"doc_mapping",
")",
"return",
"fields_arr"
] |
Converts all index's doc_types to .kibana
|
[
"Converts",
"all",
"index",
"s",
"doc_types",
"to",
".",
"kibana"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L181-L192
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.get_doc_type_mappings
|
def get_doc_type_mappings(self, doc_type):
"""Converts all doc_types' fields to .kibana"""
doc_fields_arr = []
found_score = False
for (key, val) in iteritems(doc_type):
# self.pr_dbg("\t\tfield: %s" % key)
# self.pr_dbg("\tval: %s" % val)
add_it = False
retdict = {}
# _ are system
if not key.startswith('_'):
if 'mapping' not in doc_type[key]:
self.pr_err("No mapping in doc_type[%s]" % key)
return None
if key in doc_type[key]['mapping']:
subkey_name = key
else:
subkey_name = re.sub('.*\.', '', key)
if subkey_name not in doc_type[key]['mapping']:
self.pr_err(
"Couldn't find subkey " +
"doc_type[%s]['mapping'][%s]" % (key, subkey_name))
return None
# self.pr_dbg("\t\tsubkey_name: %s" % subkey_name)
retdict = self.get_field_mappings(
doc_type[key]['mapping'][subkey_name])
add_it = True
# system mappings don't list a type,
# but kibana makes them all strings
if key in self.sys_mappings:
retdict['analyzed'] = False
retdict['indexed'] = False
if key == '_source':
retdict = self.get_field_mappings(
doc_type[key]['mapping'][key])
retdict['type'] = "_source"
elif key == '_score':
retdict['type'] = "number"
elif 'type' not in retdict:
retdict['type'] = "string"
add_it = True
if add_it:
retdict['name'] = key
retdict['count'] = 0 # always init to 0
retdict['scripted'] = False # I haven't observed a True yet
if not self.check_mapping(retdict):
self.pr_err("Error, invalid mapping")
return None
# the fields element is an escaped array of json
# make the array here, after all collected, then escape it
doc_fields_arr.append(retdict)
if not found_score:
doc_fields_arr.append(
{"name": "_score",
"type": "number",
"count": 0,
"scripted": False,
"indexed": False,
"analyzed": False,
"doc_values": False})
return doc_fields_arr
|
python
|
def get_doc_type_mappings(self, doc_type):
"""Converts all doc_types' fields to .kibana"""
doc_fields_arr = []
found_score = False
for (key, val) in iteritems(doc_type):
# self.pr_dbg("\t\tfield: %s" % key)
# self.pr_dbg("\tval: %s" % val)
add_it = False
retdict = {}
# _ are system
if not key.startswith('_'):
if 'mapping' not in doc_type[key]:
self.pr_err("No mapping in doc_type[%s]" % key)
return None
if key in doc_type[key]['mapping']:
subkey_name = key
else:
subkey_name = re.sub('.*\.', '', key)
if subkey_name not in doc_type[key]['mapping']:
self.pr_err(
"Couldn't find subkey " +
"doc_type[%s]['mapping'][%s]" % (key, subkey_name))
return None
# self.pr_dbg("\t\tsubkey_name: %s" % subkey_name)
retdict = self.get_field_mappings(
doc_type[key]['mapping'][subkey_name])
add_it = True
# system mappings don't list a type,
# but kibana makes them all strings
if key in self.sys_mappings:
retdict['analyzed'] = False
retdict['indexed'] = False
if key == '_source':
retdict = self.get_field_mappings(
doc_type[key]['mapping'][key])
retdict['type'] = "_source"
elif key == '_score':
retdict['type'] = "number"
elif 'type' not in retdict:
retdict['type'] = "string"
add_it = True
if add_it:
retdict['name'] = key
retdict['count'] = 0 # always init to 0
retdict['scripted'] = False # I haven't observed a True yet
if not self.check_mapping(retdict):
self.pr_err("Error, invalid mapping")
return None
# the fields element is an escaped array of json
# make the array here, after all collected, then escape it
doc_fields_arr.append(retdict)
if not found_score:
doc_fields_arr.append(
{"name": "_score",
"type": "number",
"count": 0,
"scripted": False,
"indexed": False,
"analyzed": False,
"doc_values": False})
return doc_fields_arr
|
[
"def",
"get_doc_type_mappings",
"(",
"self",
",",
"doc_type",
")",
":",
"doc_fields_arr",
"=",
"[",
"]",
"found_score",
"=",
"False",
"for",
"(",
"key",
",",
"val",
")",
"in",
"iteritems",
"(",
"doc_type",
")",
":",
"# self.pr_dbg(\"\\t\\tfield: %s\" % key)",
"# self.pr_dbg(\"\\tval: %s\" % val)",
"add_it",
"=",
"False",
"retdict",
"=",
"{",
"}",
"# _ are system",
"if",
"not",
"key",
".",
"startswith",
"(",
"'_'",
")",
":",
"if",
"'mapping'",
"not",
"in",
"doc_type",
"[",
"key",
"]",
":",
"self",
".",
"pr_err",
"(",
"\"No mapping in doc_type[%s]\"",
"%",
"key",
")",
"return",
"None",
"if",
"key",
"in",
"doc_type",
"[",
"key",
"]",
"[",
"'mapping'",
"]",
":",
"subkey_name",
"=",
"key",
"else",
":",
"subkey_name",
"=",
"re",
".",
"sub",
"(",
"'.*\\.'",
",",
"''",
",",
"key",
")",
"if",
"subkey_name",
"not",
"in",
"doc_type",
"[",
"key",
"]",
"[",
"'mapping'",
"]",
":",
"self",
".",
"pr_err",
"(",
"\"Couldn't find subkey \"",
"+",
"\"doc_type[%s]['mapping'][%s]\"",
"%",
"(",
"key",
",",
"subkey_name",
")",
")",
"return",
"None",
"# self.pr_dbg(\"\\t\\tsubkey_name: %s\" % subkey_name)",
"retdict",
"=",
"self",
".",
"get_field_mappings",
"(",
"doc_type",
"[",
"key",
"]",
"[",
"'mapping'",
"]",
"[",
"subkey_name",
"]",
")",
"add_it",
"=",
"True",
"# system mappings don't list a type,",
"# but kibana makes them all strings",
"if",
"key",
"in",
"self",
".",
"sys_mappings",
":",
"retdict",
"[",
"'analyzed'",
"]",
"=",
"False",
"retdict",
"[",
"'indexed'",
"]",
"=",
"False",
"if",
"key",
"==",
"'_source'",
":",
"retdict",
"=",
"self",
".",
"get_field_mappings",
"(",
"doc_type",
"[",
"key",
"]",
"[",
"'mapping'",
"]",
"[",
"key",
"]",
")",
"retdict",
"[",
"'type'",
"]",
"=",
"\"_source\"",
"elif",
"key",
"==",
"'_score'",
":",
"retdict",
"[",
"'type'",
"]",
"=",
"\"number\"",
"elif",
"'type'",
"not",
"in",
"retdict",
":",
"retdict",
"[",
"'type'",
"]",
"=",
"\"string\"",
"add_it",
"=",
"True",
"if",
"add_it",
":",
"retdict",
"[",
"'name'",
"]",
"=",
"key",
"retdict",
"[",
"'count'",
"]",
"=",
"0",
"# always init to 0",
"retdict",
"[",
"'scripted'",
"]",
"=",
"False",
"# I haven't observed a True yet",
"if",
"not",
"self",
".",
"check_mapping",
"(",
"retdict",
")",
":",
"self",
".",
"pr_err",
"(",
"\"Error, invalid mapping\"",
")",
"return",
"None",
"# the fields element is an escaped array of json",
"# make the array here, after all collected, then escape it",
"doc_fields_arr",
".",
"append",
"(",
"retdict",
")",
"if",
"not",
"found_score",
":",
"doc_fields_arr",
".",
"append",
"(",
"{",
"\"name\"",
":",
"\"_score\"",
",",
"\"type\"",
":",
"\"number\"",
",",
"\"count\"",
":",
"0",
",",
"\"scripted\"",
":",
"False",
",",
"\"indexed\"",
":",
"False",
",",
"\"analyzed\"",
":",
"False",
",",
"\"doc_values\"",
":",
"False",
"}",
")",
"return",
"doc_fields_arr"
] |
Converts all doc_types' fields to .kibana
|
[
"Converts",
"all",
"doc_types",
"fields",
"to",
".",
"kibana"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L194-L254
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.get_field_mappings
|
def get_field_mappings(self, field):
"""Converts ES field mappings to .kibana field mappings"""
retdict = {}
retdict['indexed'] = False
retdict['analyzed'] = False
for (key, val) in iteritems(field):
if key in self.mappings:
if (key == 'type' and
(val == "long" or
val == "integer" or
val == "double" or
val == "float")):
val = "number"
# self.pr_dbg("\t\t\tkey: %s" % key)
# self.pr_dbg("\t\t\t\tval: %s" % val)
retdict[key] = val
if key == 'index' and val != "no":
retdict['indexed'] = True
# self.pr_dbg("\t\t\tkey: %s" % key)
# self.pr_dbg("\t\t\t\tval: %s" % val)
if val == "analyzed":
retdict['analyzed'] = True
return retdict
|
python
|
def get_field_mappings(self, field):
"""Converts ES field mappings to .kibana field mappings"""
retdict = {}
retdict['indexed'] = False
retdict['analyzed'] = False
for (key, val) in iteritems(field):
if key in self.mappings:
if (key == 'type' and
(val == "long" or
val == "integer" or
val == "double" or
val == "float")):
val = "number"
# self.pr_dbg("\t\t\tkey: %s" % key)
# self.pr_dbg("\t\t\t\tval: %s" % val)
retdict[key] = val
if key == 'index' and val != "no":
retdict['indexed'] = True
# self.pr_dbg("\t\t\tkey: %s" % key)
# self.pr_dbg("\t\t\t\tval: %s" % val)
if val == "analyzed":
retdict['analyzed'] = True
return retdict
|
[
"def",
"get_field_mappings",
"(",
"self",
",",
"field",
")",
":",
"retdict",
"=",
"{",
"}",
"retdict",
"[",
"'indexed'",
"]",
"=",
"False",
"retdict",
"[",
"'analyzed'",
"]",
"=",
"False",
"for",
"(",
"key",
",",
"val",
")",
"in",
"iteritems",
"(",
"field",
")",
":",
"if",
"key",
"in",
"self",
".",
"mappings",
":",
"if",
"(",
"key",
"==",
"'type'",
"and",
"(",
"val",
"==",
"\"long\"",
"or",
"val",
"==",
"\"integer\"",
"or",
"val",
"==",
"\"double\"",
"or",
"val",
"==",
"\"float\"",
")",
")",
":",
"val",
"=",
"\"number\"",
"# self.pr_dbg(\"\\t\\t\\tkey: %s\" % key)",
"# self.pr_dbg(\"\\t\\t\\t\\tval: %s\" % val)",
"retdict",
"[",
"key",
"]",
"=",
"val",
"if",
"key",
"==",
"'index'",
"and",
"val",
"!=",
"\"no\"",
":",
"retdict",
"[",
"'indexed'",
"]",
"=",
"True",
"# self.pr_dbg(\"\\t\\t\\tkey: %s\" % key)",
"# self.pr_dbg(\"\\t\\t\\t\\tval: %s\" % val)",
"if",
"val",
"==",
"\"analyzed\"",
":",
"retdict",
"[",
"'analyzed'",
"]",
"=",
"True",
"return",
"retdict"
] |
Converts ES field mappings to .kibana field mappings
|
[
"Converts",
"ES",
"field",
"mappings",
"to",
".",
"kibana",
"field",
"mappings"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L256-L278
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.is_kibana_cache_incomplete
|
def is_kibana_cache_incomplete(self, es_cache, k_cache):
"""Test if k_cache is incomplete
Assume k_cache is always correct, but could be missing new
fields that es_cache has
"""
# convert list into dict, with each item's ['name'] as key
k_dict = {}
for field in k_cache:
# self.pr_dbg("field: %s" % field)
k_dict[field['name']] = field
for ign_f in self.mappings_ignore:
k_dict[field['name']][ign_f] = 0
es_dict = {}
for field in es_cache:
es_dict[field['name']] = field
for ign_f in self.mappings_ignore:
es_dict[field['name']][ign_f] = 0
es_set = set(es_dict.keys())
k_set = set(k_dict.keys())
# reasons why kibana cache could be incomplete:
# k_dict is missing keys that are within es_dict
# We don't care if k has keys that es doesn't
# es {1,2} k {1,2,3}; intersection {1,2}; len(es-{}) 0
# es {1,2} k {1,2}; intersection {1,2}; len(es-{}) 0
# es {1,2} k {}; intersection {}; len(es-{}) 2
# es {1,2} k {1}; intersection {1}; len(es-{}) 1
# es {2,3} k {1}; intersection {}; len(es-{}) 2
# es {2,3} k {1,2}; intersection {2}; len(es-{}) 1
return len(es_set - k_set.intersection(es_set)) > 0
|
python
|
def is_kibana_cache_incomplete(self, es_cache, k_cache):
"""Test if k_cache is incomplete
Assume k_cache is always correct, but could be missing new
fields that es_cache has
"""
# convert list into dict, with each item's ['name'] as key
k_dict = {}
for field in k_cache:
# self.pr_dbg("field: %s" % field)
k_dict[field['name']] = field
for ign_f in self.mappings_ignore:
k_dict[field['name']][ign_f] = 0
es_dict = {}
for field in es_cache:
es_dict[field['name']] = field
for ign_f in self.mappings_ignore:
es_dict[field['name']][ign_f] = 0
es_set = set(es_dict.keys())
k_set = set(k_dict.keys())
# reasons why kibana cache could be incomplete:
# k_dict is missing keys that are within es_dict
# We don't care if k has keys that es doesn't
# es {1,2} k {1,2,3}; intersection {1,2}; len(es-{}) 0
# es {1,2} k {1,2}; intersection {1,2}; len(es-{}) 0
# es {1,2} k {}; intersection {}; len(es-{}) 2
# es {1,2} k {1}; intersection {1}; len(es-{}) 1
# es {2,3} k {1}; intersection {}; len(es-{}) 2
# es {2,3} k {1,2}; intersection {2}; len(es-{}) 1
return len(es_set - k_set.intersection(es_set)) > 0
|
[
"def",
"is_kibana_cache_incomplete",
"(",
"self",
",",
"es_cache",
",",
"k_cache",
")",
":",
"# convert list into dict, with each item's ['name'] as key",
"k_dict",
"=",
"{",
"}",
"for",
"field",
"in",
"k_cache",
":",
"# self.pr_dbg(\"field: %s\" % field)",
"k_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"=",
"field",
"for",
"ign_f",
"in",
"self",
".",
"mappings_ignore",
":",
"k_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"[",
"ign_f",
"]",
"=",
"0",
"es_dict",
"=",
"{",
"}",
"for",
"field",
"in",
"es_cache",
":",
"es_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"=",
"field",
"for",
"ign_f",
"in",
"self",
".",
"mappings_ignore",
":",
"es_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"[",
"ign_f",
"]",
"=",
"0",
"es_set",
"=",
"set",
"(",
"es_dict",
".",
"keys",
"(",
")",
")",
"k_set",
"=",
"set",
"(",
"k_dict",
".",
"keys",
"(",
")",
")",
"# reasons why kibana cache could be incomplete:",
"# k_dict is missing keys that are within es_dict",
"# We don't care if k has keys that es doesn't",
"# es {1,2} k {1,2,3}; intersection {1,2}; len(es-{}) 0",
"# es {1,2} k {1,2}; intersection {1,2}; len(es-{}) 0",
"# es {1,2} k {}; intersection {}; len(es-{}) 2",
"# es {1,2} k {1}; intersection {1}; len(es-{}) 1",
"# es {2,3} k {1}; intersection {}; len(es-{}) 2",
"# es {2,3} k {1,2}; intersection {2}; len(es-{}) 1",
"return",
"len",
"(",
"es_set",
"-",
"k_set",
".",
"intersection",
"(",
"es_set",
")",
")",
">",
"0"
] |
Test if k_cache is incomplete
Assume k_cache is always correct, but could be missing new
fields that es_cache has
|
[
"Test",
"if",
"k_cache",
"is",
"incomplete"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L310-L339
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.list_to_compare_dict
|
def list_to_compare_dict(self, list_form):
"""Convert list into a data structure we can query easier"""
compare_dict = {}
for field in list_form:
if field['name'] in compare_dict:
self.pr_dbg("List has duplicate field %s:\n%s" %
(field['name'], compare_dict[field['name']]))
if compare_dict[field['name']] != field:
self.pr_dbg("And values are different:\n%s" % field)
return None
compare_dict[field['name']] = field
for ign_f in self.mappings_ignore:
compare_dict[field['name']][ign_f] = 0
return compare_dict
|
python
|
def list_to_compare_dict(self, list_form):
"""Convert list into a data structure we can query easier"""
compare_dict = {}
for field in list_form:
if field['name'] in compare_dict:
self.pr_dbg("List has duplicate field %s:\n%s" %
(field['name'], compare_dict[field['name']]))
if compare_dict[field['name']] != field:
self.pr_dbg("And values are different:\n%s" % field)
return None
compare_dict[field['name']] = field
for ign_f in self.mappings_ignore:
compare_dict[field['name']][ign_f] = 0
return compare_dict
|
[
"def",
"list_to_compare_dict",
"(",
"self",
",",
"list_form",
")",
":",
"compare_dict",
"=",
"{",
"}",
"for",
"field",
"in",
"list_form",
":",
"if",
"field",
"[",
"'name'",
"]",
"in",
"compare_dict",
":",
"self",
".",
"pr_dbg",
"(",
"\"List has duplicate field %s:\\n%s\"",
"%",
"(",
"field",
"[",
"'name'",
"]",
",",
"compare_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
")",
")",
"if",
"compare_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"!=",
"field",
":",
"self",
".",
"pr_dbg",
"(",
"\"And values are different:\\n%s\"",
"%",
"field",
")",
"return",
"None",
"compare_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"=",
"field",
"for",
"ign_f",
"in",
"self",
".",
"mappings_ignore",
":",
"compare_dict",
"[",
"field",
"[",
"'name'",
"]",
"]",
"[",
"ign_f",
"]",
"=",
"0",
"return",
"compare_dict"
] |
Convert list into a data structure we can query easier
|
[
"Convert",
"list",
"into",
"a",
"data",
"structure",
"we",
"can",
"query",
"easier"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L341-L354
|
train
|
rfarley3/Kibana
|
kibana/mapping.py
|
KibanaMapping.compare_field_caches
|
def compare_field_caches(self, replica, original):
"""Verify original is subset of replica"""
if original is None:
original = []
if replica is None:
replica = []
self.pr_dbg("Comparing orig with %s fields to replica with %s fields" %
(len(original), len(replica)))
# convert list into dict, with each item's ['name'] as key
orig = self.list_to_compare_dict(original)
if orig is None:
self.pr_dbg("Original has duplicate fields")
return 1
repl = self.list_to_compare_dict(replica)
if repl is None:
self.pr_dbg("Replica has duplicate fields")
return 1
# search orig for each item in repl
# if any items in repl not within orig or vice versa, then complain
# make sure contents of each item match
orig_found = {}
for (key, field) in iteritems(repl):
field_name = field['name']
if field_name not in orig:
self.pr_dbg("Replica has field not found in orig %s: %s" %
(field_name, field))
return 1
orig_found[field_name] = True
if orig[field_name] != field:
self.pr_dbg("Field in replica doesn't match orig:")
self.pr_dbg("orig:%s\nrepl:%s" % (orig[field_name], field))
return 1
unfound = set(orig_found.keys()) - set(repl.keys())
if len(unfound) > 0:
self.pr_dbg("Orig contains fields that were not in replica")
self.pr_dbg('%s' % unfound)
return 1
# We don't care about case when replica has more fields than orig
# unfound = set(repl.keys()) - set(orig_found.keys())
# if len(unfound) > 0:
# self.pr_dbg("Replica contains fields that were not in orig")
# self.pr_dbg('%s' % unfound)
# return 1
self.pr_dbg("Original matches replica")
return 0
|
python
|
def compare_field_caches(self, replica, original):
"""Verify original is subset of replica"""
if original is None:
original = []
if replica is None:
replica = []
self.pr_dbg("Comparing orig with %s fields to replica with %s fields" %
(len(original), len(replica)))
# convert list into dict, with each item's ['name'] as key
orig = self.list_to_compare_dict(original)
if orig is None:
self.pr_dbg("Original has duplicate fields")
return 1
repl = self.list_to_compare_dict(replica)
if repl is None:
self.pr_dbg("Replica has duplicate fields")
return 1
# search orig for each item in repl
# if any items in repl not within orig or vice versa, then complain
# make sure contents of each item match
orig_found = {}
for (key, field) in iteritems(repl):
field_name = field['name']
if field_name not in orig:
self.pr_dbg("Replica has field not found in orig %s: %s" %
(field_name, field))
return 1
orig_found[field_name] = True
if orig[field_name] != field:
self.pr_dbg("Field in replica doesn't match orig:")
self.pr_dbg("orig:%s\nrepl:%s" % (orig[field_name], field))
return 1
unfound = set(orig_found.keys()) - set(repl.keys())
if len(unfound) > 0:
self.pr_dbg("Orig contains fields that were not in replica")
self.pr_dbg('%s' % unfound)
return 1
# We don't care about case when replica has more fields than orig
# unfound = set(repl.keys()) - set(orig_found.keys())
# if len(unfound) > 0:
# self.pr_dbg("Replica contains fields that were not in orig")
# self.pr_dbg('%s' % unfound)
# return 1
self.pr_dbg("Original matches replica")
return 0
|
[
"def",
"compare_field_caches",
"(",
"self",
",",
"replica",
",",
"original",
")",
":",
"if",
"original",
"is",
"None",
":",
"original",
"=",
"[",
"]",
"if",
"replica",
"is",
"None",
":",
"replica",
"=",
"[",
"]",
"self",
".",
"pr_dbg",
"(",
"\"Comparing orig with %s fields to replica with %s fields\"",
"%",
"(",
"len",
"(",
"original",
")",
",",
"len",
"(",
"replica",
")",
")",
")",
"# convert list into dict, with each item's ['name'] as key",
"orig",
"=",
"self",
".",
"list_to_compare_dict",
"(",
"original",
")",
"if",
"orig",
"is",
"None",
":",
"self",
".",
"pr_dbg",
"(",
"\"Original has duplicate fields\"",
")",
"return",
"1",
"repl",
"=",
"self",
".",
"list_to_compare_dict",
"(",
"replica",
")",
"if",
"repl",
"is",
"None",
":",
"self",
".",
"pr_dbg",
"(",
"\"Replica has duplicate fields\"",
")",
"return",
"1",
"# search orig for each item in repl",
"# if any items in repl not within orig or vice versa, then complain",
"# make sure contents of each item match",
"orig_found",
"=",
"{",
"}",
"for",
"(",
"key",
",",
"field",
")",
"in",
"iteritems",
"(",
"repl",
")",
":",
"field_name",
"=",
"field",
"[",
"'name'",
"]",
"if",
"field_name",
"not",
"in",
"orig",
":",
"self",
".",
"pr_dbg",
"(",
"\"Replica has field not found in orig %s: %s\"",
"%",
"(",
"field_name",
",",
"field",
")",
")",
"return",
"1",
"orig_found",
"[",
"field_name",
"]",
"=",
"True",
"if",
"orig",
"[",
"field_name",
"]",
"!=",
"field",
":",
"self",
".",
"pr_dbg",
"(",
"\"Field in replica doesn't match orig:\"",
")",
"self",
".",
"pr_dbg",
"(",
"\"orig:%s\\nrepl:%s\"",
"%",
"(",
"orig",
"[",
"field_name",
"]",
",",
"field",
")",
")",
"return",
"1",
"unfound",
"=",
"set",
"(",
"orig_found",
".",
"keys",
"(",
")",
")",
"-",
"set",
"(",
"repl",
".",
"keys",
"(",
")",
")",
"if",
"len",
"(",
"unfound",
")",
">",
"0",
":",
"self",
".",
"pr_dbg",
"(",
"\"Orig contains fields that were not in replica\"",
")",
"self",
".",
"pr_dbg",
"(",
"'%s'",
"%",
"unfound",
")",
"return",
"1",
"# We don't care about case when replica has more fields than orig",
"# unfound = set(repl.keys()) - set(orig_found.keys())",
"# if len(unfound) > 0:",
"# self.pr_dbg(\"Replica contains fields that were not in orig\")",
"# self.pr_dbg('%s' % unfound)",
"# return 1",
"self",
".",
"pr_dbg",
"(",
"\"Original matches replica\"",
")",
"return",
"0"
] |
Verify original is subset of replica
|
[
"Verify",
"original",
"is",
"subset",
"of",
"replica"
] |
3df1e13be18edfb39ec173d8d2bbe9e90be61022
|
https://github.com/rfarley3/Kibana/blob/3df1e13be18edfb39ec173d8d2bbe9e90be61022/kibana/mapping.py#L356-L400
|
train
|
deep-compute/logagg
|
logagg/util.py
|
start_daemon_thread
|
def start_daemon_thread(target, args=()):
"""starts a deamon thread for a given target function and arguments."""
th = Thread(target=target, args=args)
th.daemon = True
th.start()
return th
|
python
|
def start_daemon_thread(target, args=()):
"""starts a deamon thread for a given target function and arguments."""
th = Thread(target=target, args=args)
th.daemon = True
th.start()
return th
|
[
"def",
"start_daemon_thread",
"(",
"target",
",",
"args",
"=",
"(",
")",
")",
":",
"th",
"=",
"Thread",
"(",
"target",
"=",
"target",
",",
"args",
"=",
"args",
")",
"th",
".",
"daemon",
"=",
"True",
"th",
".",
"start",
"(",
")",
"return",
"th"
] |
starts a deamon thread for a given target function and arguments.
|
[
"starts",
"a",
"deamon",
"thread",
"for",
"a",
"given",
"target",
"function",
"and",
"arguments",
"."
] |
7863bc1b5ddf3e67c4d4b55746799304180589a0
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/util.py#L45-L50
|
train
|
deep-compute/logagg
|
logagg/util.py
|
serialize_dict_keys
|
def serialize_dict_keys(d, prefix=""):
"""returns all the keys in a dictionary.
>>> serialize_dict_keys({"a": {"b": {"c": 1, "b": 2} } })
['a', 'a.b', 'a.b.c', 'a.b.b']
"""
keys = []
for k, v in d.iteritems():
fqk = '%s%s' % (prefix, k)
keys.append(fqk)
if isinstance(v, dict):
keys.extend(serialize_dict_keys(v, prefix="%s." % fqk))
return keys
|
python
|
def serialize_dict_keys(d, prefix=""):
"""returns all the keys in a dictionary.
>>> serialize_dict_keys({"a": {"b": {"c": 1, "b": 2} } })
['a', 'a.b', 'a.b.c', 'a.b.b']
"""
keys = []
for k, v in d.iteritems():
fqk = '%s%s' % (prefix, k)
keys.append(fqk)
if isinstance(v, dict):
keys.extend(serialize_dict_keys(v, prefix="%s." % fqk))
return keys
|
[
"def",
"serialize_dict_keys",
"(",
"d",
",",
"prefix",
"=",
"\"\"",
")",
":",
"keys",
"=",
"[",
"]",
"for",
"k",
",",
"v",
"in",
"d",
".",
"iteritems",
"(",
")",
":",
"fqk",
"=",
"'%s%s'",
"%",
"(",
"prefix",
",",
"k",
")",
"keys",
".",
"append",
"(",
"fqk",
")",
"if",
"isinstance",
"(",
"v",
",",
"dict",
")",
":",
"keys",
".",
"extend",
"(",
"serialize_dict_keys",
"(",
"v",
",",
"prefix",
"=",
"\"%s.\"",
"%",
"fqk",
")",
")",
"return",
"keys"
] |
returns all the keys in a dictionary.
>>> serialize_dict_keys({"a": {"b": {"c": 1, "b": 2} } })
['a', 'a.b', 'a.b.c', 'a.b.b']
|
[
"returns",
"all",
"the",
"keys",
"in",
"a",
"dictionary",
"."
] |
7863bc1b5ddf3e67c4d4b55746799304180589a0
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/util.py#L53-L66
|
train
|
zetaops/zengine
|
zengine/auth/auth_backend.py
|
AuthBackend.set_user
|
def set_user(self, user):
"""
Writes user data to session.
Args:
user: User object
"""
self.session['user_id'] = user.key
self.session['user_data'] = user.clean_value()
role = self.get_role()
# TODO: this should be remembered from previous login
# self.session['role_data'] = default_role.clean_value()
self.session['role_id'] = role.key
self.current.role_id = role.key
self.current.user_id = user.key
# self.perm_cache = PermissionCache(role.key)
self.session['permissions'] = role.get_permissions()
|
python
|
def set_user(self, user):
"""
Writes user data to session.
Args:
user: User object
"""
self.session['user_id'] = user.key
self.session['user_data'] = user.clean_value()
role = self.get_role()
# TODO: this should be remembered from previous login
# self.session['role_data'] = default_role.clean_value()
self.session['role_id'] = role.key
self.current.role_id = role.key
self.current.user_id = user.key
# self.perm_cache = PermissionCache(role.key)
self.session['permissions'] = role.get_permissions()
|
[
"def",
"set_user",
"(",
"self",
",",
"user",
")",
":",
"self",
".",
"session",
"[",
"'user_id'",
"]",
"=",
"user",
".",
"key",
"self",
".",
"session",
"[",
"'user_data'",
"]",
"=",
"user",
".",
"clean_value",
"(",
")",
"role",
"=",
"self",
".",
"get_role",
"(",
")",
"# TODO: this should be remembered from previous login",
"# self.session['role_data'] = default_role.clean_value()",
"self",
".",
"session",
"[",
"'role_id'",
"]",
"=",
"role",
".",
"key",
"self",
".",
"current",
".",
"role_id",
"=",
"role",
".",
"key",
"self",
".",
"current",
".",
"user_id",
"=",
"user",
".",
"key",
"# self.perm_cache = PermissionCache(role.key)",
"self",
".",
"session",
"[",
"'permissions'",
"]",
"=",
"role",
".",
"get_permissions",
"(",
")"
] |
Writes user data to session.
Args:
user: User object
|
[
"Writes",
"user",
"data",
"to",
"session",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/auth/auth_backend.py#L33-L50
|
train
|
LordDarkula/chess_py
|
chess_py/pieces/piece.py
|
Piece.contains_opposite_color_piece
|
def contains_opposite_color_piece(self, square, position):
"""
Finds if square on the board is occupied by a ``Piece``
belonging to the opponent.
:type: square: Location
:type: position: Board
:rtype: bool
"""
return not position.is_square_empty(square) and \
position.piece_at_square(square).color != self.color
|
python
|
def contains_opposite_color_piece(self, square, position):
"""
Finds if square on the board is occupied by a ``Piece``
belonging to the opponent.
:type: square: Location
:type: position: Board
:rtype: bool
"""
return not position.is_square_empty(square) and \
position.piece_at_square(square).color != self.color
|
[
"def",
"contains_opposite_color_piece",
"(",
"self",
",",
"square",
",",
"position",
")",
":",
"return",
"not",
"position",
".",
"is_square_empty",
"(",
"square",
")",
"and",
"position",
".",
"piece_at_square",
"(",
"square",
")",
".",
"color",
"!=",
"self",
".",
"color"
] |
Finds if square on the board is occupied by a ``Piece``
belonging to the opponent.
:type: square: Location
:type: position: Board
:rtype: bool
|
[
"Finds",
"if",
"square",
"on",
"the",
"board",
"is",
"occupied",
"by",
"a",
"Piece",
"belonging",
"to",
"the",
"opponent",
"."
] |
14bebc2f8c49ae25c59375cc83d0b38d8ff7281d
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/pieces/piece.py#L79-L89
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
gettext
|
def gettext(message, domain=DEFAULT_DOMAIN):
"""Mark a message as translateable, and translate it.
All messages in the application that are translateable should be wrapped with this function.
When importing this function, it should be renamed to '_'. For example:
.. code-block:: python
from zengine.lib.translation import gettext as _
print(_('Hello, world!'))
'Merhaba, dünya!'
For the messages that will be formatted later on, instead of using the position-based
formatting, key-based formatting should be used. This gives the translator an idea what
the variables in the format are going to be, and makes it possible for the translator
to reorder the variables. For example:
.. code-block:: python
name, number = 'Elizabeth', 'II'
_('Queen %(name)s %(number)s') % {'name': name, 'number': number}
'Kraliçe II. Elizabeth'
The message returned by this function depends on the language of the current user.
If this function is called before a language is installed (which is normally done
by ZEngine when the user connects), this function will simply return the message
without modification.
If there are messages containing unicode characters, in Python 2 these messages must
be marked as unicode. Otherwise, python will not be able to correctly match these
messages with translations. For example:
.. code-block:: python
print(_('Café'))
'Café'
print(_(u'Café'))
'Kahve'
Args:
message (basestring, unicode): The input message.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The translated message.
"""
if six.PY2:
return InstalledLocale._active_catalogs[domain].ugettext(message)
else:
return InstalledLocale._active_catalogs[domain].gettext(message)
|
python
|
def gettext(message, domain=DEFAULT_DOMAIN):
"""Mark a message as translateable, and translate it.
All messages in the application that are translateable should be wrapped with this function.
When importing this function, it should be renamed to '_'. For example:
.. code-block:: python
from zengine.lib.translation import gettext as _
print(_('Hello, world!'))
'Merhaba, dünya!'
For the messages that will be formatted later on, instead of using the position-based
formatting, key-based formatting should be used. This gives the translator an idea what
the variables in the format are going to be, and makes it possible for the translator
to reorder the variables. For example:
.. code-block:: python
name, number = 'Elizabeth', 'II'
_('Queen %(name)s %(number)s') % {'name': name, 'number': number}
'Kraliçe II. Elizabeth'
The message returned by this function depends on the language of the current user.
If this function is called before a language is installed (which is normally done
by ZEngine when the user connects), this function will simply return the message
without modification.
If there are messages containing unicode characters, in Python 2 these messages must
be marked as unicode. Otherwise, python will not be able to correctly match these
messages with translations. For example:
.. code-block:: python
print(_('Café'))
'Café'
print(_(u'Café'))
'Kahve'
Args:
message (basestring, unicode): The input message.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The translated message.
"""
if six.PY2:
return InstalledLocale._active_catalogs[domain].ugettext(message)
else:
return InstalledLocale._active_catalogs[domain].gettext(message)
|
[
"def",
"gettext",
"(",
"message",
",",
"domain",
"=",
"DEFAULT_DOMAIN",
")",
":",
"if",
"six",
".",
"PY2",
":",
"return",
"InstalledLocale",
".",
"_active_catalogs",
"[",
"domain",
"]",
".",
"ugettext",
"(",
"message",
")",
"else",
":",
"return",
"InstalledLocale",
".",
"_active_catalogs",
"[",
"domain",
"]",
".",
"gettext",
"(",
"message",
")"
] |
Mark a message as translateable, and translate it.
All messages in the application that are translateable should be wrapped with this function.
When importing this function, it should be renamed to '_'. For example:
.. code-block:: python
from zengine.lib.translation import gettext as _
print(_('Hello, world!'))
'Merhaba, dünya!'
For the messages that will be formatted later on, instead of using the position-based
formatting, key-based formatting should be used. This gives the translator an idea what
the variables in the format are going to be, and makes it possible for the translator
to reorder the variables. For example:
.. code-block:: python
name, number = 'Elizabeth', 'II'
_('Queen %(name)s %(number)s') % {'name': name, 'number': number}
'Kraliçe II. Elizabeth'
The message returned by this function depends on the language of the current user.
If this function is called before a language is installed (which is normally done
by ZEngine when the user connects), this function will simply return the message
without modification.
If there are messages containing unicode characters, in Python 2 these messages must
be marked as unicode. Otherwise, python will not be able to correctly match these
messages with translations. For example:
.. code-block:: python
print(_('Café'))
'Café'
print(_(u'Café'))
'Kahve'
Args:
message (basestring, unicode): The input message.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The translated message.
|
[
"Mark",
"a",
"message",
"as",
"translateable",
"and",
"translate",
"it",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L133-L184
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
gettext_lazy
|
def gettext_lazy(message, domain=DEFAULT_DOMAIN):
"""Mark a message as translatable, but delay the translation until the message is used.
Sometimes, there are some messages that need to be translated, but the translation
can't be done at the point the message itself is written. For example, the names of
the fields in a Model can't be translated at the point they are written, otherwise
the translation would be done when the file is imported, long before a user even connects.
To avoid this, `gettext_lazy` should be used. For example:
.. code-block:: python
from zengine.lib.translation import gettext_lazy, InstalledLocale
from pyoko import model, fields
class User(model.Model):
name = fields.String(gettext_lazy('User Name'))
print(User.name.title)
'User Name'
InstalledLocale.install_language('tr')
print(User.name.title)
'Kullanıcı Adı'
Args:
message (basestring, unicode): The input message.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The translated message, with the translation itself being delayed until
the text is actually used.
"""
return LazyProxy(gettext, message, domain=domain, enable_cache=False)
|
python
|
def gettext_lazy(message, domain=DEFAULT_DOMAIN):
"""Mark a message as translatable, but delay the translation until the message is used.
Sometimes, there are some messages that need to be translated, but the translation
can't be done at the point the message itself is written. For example, the names of
the fields in a Model can't be translated at the point they are written, otherwise
the translation would be done when the file is imported, long before a user even connects.
To avoid this, `gettext_lazy` should be used. For example:
.. code-block:: python
from zengine.lib.translation import gettext_lazy, InstalledLocale
from pyoko import model, fields
class User(model.Model):
name = fields.String(gettext_lazy('User Name'))
print(User.name.title)
'User Name'
InstalledLocale.install_language('tr')
print(User.name.title)
'Kullanıcı Adı'
Args:
message (basestring, unicode): The input message.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The translated message, with the translation itself being delayed until
the text is actually used.
"""
return LazyProxy(gettext, message, domain=domain, enable_cache=False)
|
[
"def",
"gettext_lazy",
"(",
"message",
",",
"domain",
"=",
"DEFAULT_DOMAIN",
")",
":",
"return",
"LazyProxy",
"(",
"gettext",
",",
"message",
",",
"domain",
"=",
"domain",
",",
"enable_cache",
"=",
"False",
")"
] |
Mark a message as translatable, but delay the translation until the message is used.
Sometimes, there are some messages that need to be translated, but the translation
can't be done at the point the message itself is written. For example, the names of
the fields in a Model can't be translated at the point they are written, otherwise
the translation would be done when the file is imported, long before a user even connects.
To avoid this, `gettext_lazy` should be used. For example:
.. code-block:: python
from zengine.lib.translation import gettext_lazy, InstalledLocale
from pyoko import model, fields
class User(model.Model):
name = fields.String(gettext_lazy('User Name'))
print(User.name.title)
'User Name'
InstalledLocale.install_language('tr')
print(User.name.title)
'Kullanıcı Adı'
Args:
message (basestring, unicode): The input message.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The translated message, with the translation itself being delayed until
the text is actually used.
|
[
"Mark",
"a",
"message",
"as",
"translatable",
"but",
"delay",
"the",
"translation",
"until",
"the",
"message",
"is",
"used",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L187-L219
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
ngettext
|
def ngettext(singular, plural, n, domain=DEFAULT_DOMAIN):
"""Mark a message as translateable, and translate it considering plural forms.
Some messages may need to change based on a number. For example, consider a message
like the following:
.. code-block:: python
def alert_msg(msg_count): print(
'You have %d %s' % (msg_count, 'message' if msg_count == 1 else 'messages'))
alert_msg(1)
'You have 1 message'
alert_msg(5)
'You have 5 messages'
To translate this message, you can use ngettext to consider the plural forms:
.. code-block:: python
from zengine.lib.translation import ngettext
def alert_msg(msg_count): print(ngettext('You have %(count)d message',
'You have %(count)d messages',
msg_count) % {'count': msg_count})
alert_msg(1)
'1 mesajınız var'
alert_msg(5)
'5 mesajlarınız var'
When doing formatting, both singular and plural forms of the message should
have the exactly same variables.
Args:
singular (unicode): The singular form of the message.
plural (unicode): The plural form of the message.
n (int): The number that is used to decide which form should be used.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The correct pluralization, translated.
"""
if six.PY2:
return InstalledLocale._active_catalogs[domain].ungettext(singular, plural, n)
else:
return InstalledLocale._active_catalogs[domain].ngettext(singular, plural, n)
|
python
|
def ngettext(singular, plural, n, domain=DEFAULT_DOMAIN):
"""Mark a message as translateable, and translate it considering plural forms.
Some messages may need to change based on a number. For example, consider a message
like the following:
.. code-block:: python
def alert_msg(msg_count): print(
'You have %d %s' % (msg_count, 'message' if msg_count == 1 else 'messages'))
alert_msg(1)
'You have 1 message'
alert_msg(5)
'You have 5 messages'
To translate this message, you can use ngettext to consider the plural forms:
.. code-block:: python
from zengine.lib.translation import ngettext
def alert_msg(msg_count): print(ngettext('You have %(count)d message',
'You have %(count)d messages',
msg_count) % {'count': msg_count})
alert_msg(1)
'1 mesajınız var'
alert_msg(5)
'5 mesajlarınız var'
When doing formatting, both singular and plural forms of the message should
have the exactly same variables.
Args:
singular (unicode): The singular form of the message.
plural (unicode): The plural form of the message.
n (int): The number that is used to decide which form should be used.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The correct pluralization, translated.
"""
if six.PY2:
return InstalledLocale._active_catalogs[domain].ungettext(singular, plural, n)
else:
return InstalledLocale._active_catalogs[domain].ngettext(singular, plural, n)
|
[
"def",
"ngettext",
"(",
"singular",
",",
"plural",
",",
"n",
",",
"domain",
"=",
"DEFAULT_DOMAIN",
")",
":",
"if",
"six",
".",
"PY2",
":",
"return",
"InstalledLocale",
".",
"_active_catalogs",
"[",
"domain",
"]",
".",
"ungettext",
"(",
"singular",
",",
"plural",
",",
"n",
")",
"else",
":",
"return",
"InstalledLocale",
".",
"_active_catalogs",
"[",
"domain",
"]",
".",
"ngettext",
"(",
"singular",
",",
"plural",
",",
"n",
")"
] |
Mark a message as translateable, and translate it considering plural forms.
Some messages may need to change based on a number. For example, consider a message
like the following:
.. code-block:: python
def alert_msg(msg_count): print(
'You have %d %s' % (msg_count, 'message' if msg_count == 1 else 'messages'))
alert_msg(1)
'You have 1 message'
alert_msg(5)
'You have 5 messages'
To translate this message, you can use ngettext to consider the plural forms:
.. code-block:: python
from zengine.lib.translation import ngettext
def alert_msg(msg_count): print(ngettext('You have %(count)d message',
'You have %(count)d messages',
msg_count) % {'count': msg_count})
alert_msg(1)
'1 mesajınız var'
alert_msg(5)
'5 mesajlarınız var'
When doing formatting, both singular and plural forms of the message should
have the exactly same variables.
Args:
singular (unicode): The singular form of the message.
plural (unicode): The plural form of the message.
n (int): The number that is used to decide which form should be used.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The correct pluralization, translated.
|
[
"Mark",
"a",
"message",
"as",
"translateable",
"and",
"translate",
"it",
"considering",
"plural",
"forms",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L222-L269
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
ngettext_lazy
|
def ngettext_lazy(singular, plural, n, domain=DEFAULT_DOMAIN):
"""Mark a message with plural forms translateable, and delay the translation
until the message is used.
Works the same was a `ngettext`, with a delaying functionality similiar to `gettext_lazy`.
Args:
singular (unicode): The singular form of the message.
plural (unicode): The plural form of the message.
n (int): The number that is used to decide which form should be used.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The correct pluralization, with the translation being
delayed until the message is used.
"""
return LazyProxy(ngettext, singular, plural, n, domain=domain, enable_cache=False)
|
python
|
def ngettext_lazy(singular, plural, n, domain=DEFAULT_DOMAIN):
"""Mark a message with plural forms translateable, and delay the translation
until the message is used.
Works the same was a `ngettext`, with a delaying functionality similiar to `gettext_lazy`.
Args:
singular (unicode): The singular form of the message.
plural (unicode): The plural form of the message.
n (int): The number that is used to decide which form should be used.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The correct pluralization, with the translation being
delayed until the message is used.
"""
return LazyProxy(ngettext, singular, plural, n, domain=domain, enable_cache=False)
|
[
"def",
"ngettext_lazy",
"(",
"singular",
",",
"plural",
",",
"n",
",",
"domain",
"=",
"DEFAULT_DOMAIN",
")",
":",
"return",
"LazyProxy",
"(",
"ngettext",
",",
"singular",
",",
"plural",
",",
"n",
",",
"domain",
"=",
"domain",
",",
"enable_cache",
"=",
"False",
")"
] |
Mark a message with plural forms translateable, and delay the translation
until the message is used.
Works the same was a `ngettext`, with a delaying functionality similiar to `gettext_lazy`.
Args:
singular (unicode): The singular form of the message.
plural (unicode): The plural form of the message.
n (int): The number that is used to decide which form should be used.
domain (basestring): The domain of the message. Defaults to 'messages', which
is the domain where all application messages should be located.
Returns:
unicode: The correct pluralization, with the translation being
delayed until the message is used.
|
[
"Mark",
"a",
"message",
"with",
"plural",
"forms",
"translateable",
"and",
"delay",
"the",
"translation",
"until",
"the",
"message",
"is",
"used",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L272-L288
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
_wrap_locale_formatter
|
def _wrap_locale_formatter(fn, locale_type):
"""Wrap a Babel data formatting function to automatically format
for currently installed locale."""
def wrapped_locale_formatter(*args, **kwargs):
"""A Babel formatting function, wrapped to automatically use the
currently installed language.
The wrapped function will not throw any exceptions for unknown locales,
if Babel doesn't recognise the locale, we will simply fall back to
the default language.
The locale used by the wrapped function can be overriden by passing it a `locale` keyword.
To learn more about this function, check the documentation of Babel for the function of
the same name.
"""
# Get the current locale from the class
kwargs_ = {'locale': getattr(InstalledLocale, locale_type)}
# By creating a dict then updating it, we allow locale to be overridden
kwargs_.update(kwargs)
try:
formatted = fn(*args, **kwargs_)
except UnknownLocaleError:
log.warning(
"""Can\'t do formatting for language code {locale},
falling back to default {default}""".format(
locale=kwargs_['locale'],
default=settings.DEFAULT_LANG)
)
kwargs_['locale'] = settings.DEFAULT_LANG
formatted = fn(*args, **kwargs_)
return formatted
return wrapped_locale_formatter
|
python
|
def _wrap_locale_formatter(fn, locale_type):
"""Wrap a Babel data formatting function to automatically format
for currently installed locale."""
def wrapped_locale_formatter(*args, **kwargs):
"""A Babel formatting function, wrapped to automatically use the
currently installed language.
The wrapped function will not throw any exceptions for unknown locales,
if Babel doesn't recognise the locale, we will simply fall back to
the default language.
The locale used by the wrapped function can be overriden by passing it a `locale` keyword.
To learn more about this function, check the documentation of Babel for the function of
the same name.
"""
# Get the current locale from the class
kwargs_ = {'locale': getattr(InstalledLocale, locale_type)}
# By creating a dict then updating it, we allow locale to be overridden
kwargs_.update(kwargs)
try:
formatted = fn(*args, **kwargs_)
except UnknownLocaleError:
log.warning(
"""Can\'t do formatting for language code {locale},
falling back to default {default}""".format(
locale=kwargs_['locale'],
default=settings.DEFAULT_LANG)
)
kwargs_['locale'] = settings.DEFAULT_LANG
formatted = fn(*args, **kwargs_)
return formatted
return wrapped_locale_formatter
|
[
"def",
"_wrap_locale_formatter",
"(",
"fn",
",",
"locale_type",
")",
":",
"def",
"wrapped_locale_formatter",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"\"\"\"A Babel formatting function, wrapped to automatically use the\n currently installed language.\n\n The wrapped function will not throw any exceptions for unknown locales,\n if Babel doesn't recognise the locale, we will simply fall back to\n the default language.\n\n The locale used by the wrapped function can be overriden by passing it a `locale` keyword.\n To learn more about this function, check the documentation of Babel for the function of\n the same name.\n \"\"\"",
"# Get the current locale from the class",
"kwargs_",
"=",
"{",
"'locale'",
":",
"getattr",
"(",
"InstalledLocale",
",",
"locale_type",
")",
"}",
"# By creating a dict then updating it, we allow locale to be overridden",
"kwargs_",
".",
"update",
"(",
"kwargs",
")",
"try",
":",
"formatted",
"=",
"fn",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs_",
")",
"except",
"UnknownLocaleError",
":",
"log",
".",
"warning",
"(",
"\"\"\"Can\\'t do formatting for language code {locale},\n falling back to default {default}\"\"\"",
".",
"format",
"(",
"locale",
"=",
"kwargs_",
"[",
"'locale'",
"]",
",",
"default",
"=",
"settings",
".",
"DEFAULT_LANG",
")",
")",
"kwargs_",
"[",
"'locale'",
"]",
"=",
"settings",
".",
"DEFAULT_LANG",
"formatted",
"=",
"fn",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs_",
")",
"return",
"formatted",
"return",
"wrapped_locale_formatter"
] |
Wrap a Babel data formatting function to automatically format
for currently installed locale.
|
[
"Wrap",
"a",
"Babel",
"data",
"formatting",
"function",
"to",
"automatically",
"format",
"for",
"currently",
"installed",
"locale",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L328-L361
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
InstalledLocale.install_language
|
def install_language(cls, language_code):
"""Install the translations for language specified by `language_code`.
If we don't have translations for this language, then the default language will be used.
If the language specified is already installed, then this is a no-op.
"""
# Skip if the language is already installed
if language_code == cls.language:
return
try:
cls._active_catalogs = cls._translation_catalogs[language_code]
cls.language = language_code
log.debug('Installed language %s', language_code)
except KeyError:
default = settings.DEFAULT_LANG
log.warning('Unknown language %s, falling back to %s', language_code, default)
cls._active_catalogs = cls._translation_catalogs[default]
cls.language = default
|
python
|
def install_language(cls, language_code):
"""Install the translations for language specified by `language_code`.
If we don't have translations for this language, then the default language will be used.
If the language specified is already installed, then this is a no-op.
"""
# Skip if the language is already installed
if language_code == cls.language:
return
try:
cls._active_catalogs = cls._translation_catalogs[language_code]
cls.language = language_code
log.debug('Installed language %s', language_code)
except KeyError:
default = settings.DEFAULT_LANG
log.warning('Unknown language %s, falling back to %s', language_code, default)
cls._active_catalogs = cls._translation_catalogs[default]
cls.language = default
|
[
"def",
"install_language",
"(",
"cls",
",",
"language_code",
")",
":",
"# Skip if the language is already installed",
"if",
"language_code",
"==",
"cls",
".",
"language",
":",
"return",
"try",
":",
"cls",
".",
"_active_catalogs",
"=",
"cls",
".",
"_translation_catalogs",
"[",
"language_code",
"]",
"cls",
".",
"language",
"=",
"language_code",
"log",
".",
"debug",
"(",
"'Installed language %s'",
",",
"language_code",
")",
"except",
"KeyError",
":",
"default",
"=",
"settings",
".",
"DEFAULT_LANG",
"log",
".",
"warning",
"(",
"'Unknown language %s, falling back to %s'",
",",
"language_code",
",",
"default",
")",
"cls",
".",
"_active_catalogs",
"=",
"cls",
".",
"_translation_catalogs",
"[",
"default",
"]",
"cls",
".",
"language",
"=",
"default"
] |
Install the translations for language specified by `language_code`.
If we don't have translations for this language, then the default language will be used.
If the language specified is already installed, then this is a no-op.
|
[
"Install",
"the",
"translations",
"for",
"language",
"specified",
"by",
"language_code",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L89-L107
|
train
|
zetaops/zengine
|
zengine/lib/translation.py
|
InstalledLocale.install_locale
|
def install_locale(cls, locale_code, locale_type):
"""Install the locale specified by `language_code`, for localizations of type `locale_type`.
If we can't perform localized formatting for the specified locale,
then the default localization format will be used.
If the locale specified is already installed for the selected type, then this is a no-op.
"""
# Skip if the locale is already installed
if locale_code == getattr(cls, locale_type):
return
try:
# We create a Locale instance to see if the locale code is supported
locale = Locale(locale_code)
log.debug('Installed locale %s', locale_code)
except UnknownLocaleError:
default = settings.DEFAULT_LOCALIZATION_FORMAT
log.warning('Unknown locale %s, falling back to %s', locale_code, default)
locale = Locale(default)
setattr(cls, locale_type, locale.language)
|
python
|
def install_locale(cls, locale_code, locale_type):
"""Install the locale specified by `language_code`, for localizations of type `locale_type`.
If we can't perform localized formatting for the specified locale,
then the default localization format will be used.
If the locale specified is already installed for the selected type, then this is a no-op.
"""
# Skip if the locale is already installed
if locale_code == getattr(cls, locale_type):
return
try:
# We create a Locale instance to see if the locale code is supported
locale = Locale(locale_code)
log.debug('Installed locale %s', locale_code)
except UnknownLocaleError:
default = settings.DEFAULT_LOCALIZATION_FORMAT
log.warning('Unknown locale %s, falling back to %s', locale_code, default)
locale = Locale(default)
setattr(cls, locale_type, locale.language)
|
[
"def",
"install_locale",
"(",
"cls",
",",
"locale_code",
",",
"locale_type",
")",
":",
"# Skip if the locale is already installed",
"if",
"locale_code",
"==",
"getattr",
"(",
"cls",
",",
"locale_type",
")",
":",
"return",
"try",
":",
"# We create a Locale instance to see if the locale code is supported",
"locale",
"=",
"Locale",
"(",
"locale_code",
")",
"log",
".",
"debug",
"(",
"'Installed locale %s'",
",",
"locale_code",
")",
"except",
"UnknownLocaleError",
":",
"default",
"=",
"settings",
".",
"DEFAULT_LOCALIZATION_FORMAT",
"log",
".",
"warning",
"(",
"'Unknown locale %s, falling back to %s'",
",",
"locale_code",
",",
"default",
")",
"locale",
"=",
"Locale",
"(",
"default",
")",
"setattr",
"(",
"cls",
",",
"locale_type",
",",
"locale",
".",
"language",
")"
] |
Install the locale specified by `language_code`, for localizations of type `locale_type`.
If we can't perform localized formatting for the specified locale,
then the default localization format will be used.
If the locale specified is already installed for the selected type, then this is a no-op.
|
[
"Install",
"the",
"locale",
"specified",
"by",
"language_code",
"for",
"localizations",
"of",
"type",
"locale_type",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/lib/translation.py#L110-L130
|
train
|
cimm-kzn/CGRtools
|
CGRtools/algorithms/depict.py
|
Depict._rotate_vector
|
def _rotate_vector(x, y, x2, y2, x1, y1):
"""
rotate x,y vector over x2-x1, y2-y1 angle
"""
angle = atan2(y2 - y1, x2 - x1)
cos_rad = cos(angle)
sin_rad = sin(angle)
return cos_rad * x + sin_rad * y, -sin_rad * x + cos_rad * y
|
python
|
def _rotate_vector(x, y, x2, y2, x1, y1):
"""
rotate x,y vector over x2-x1, y2-y1 angle
"""
angle = atan2(y2 - y1, x2 - x1)
cos_rad = cos(angle)
sin_rad = sin(angle)
return cos_rad * x + sin_rad * y, -sin_rad * x + cos_rad * y
|
[
"def",
"_rotate_vector",
"(",
"x",
",",
"y",
",",
"x2",
",",
"y2",
",",
"x1",
",",
"y1",
")",
":",
"angle",
"=",
"atan2",
"(",
"y2",
"-",
"y1",
",",
"x2",
"-",
"x1",
")",
"cos_rad",
"=",
"cos",
"(",
"angle",
")",
"sin_rad",
"=",
"sin",
"(",
"angle",
")",
"return",
"cos_rad",
"*",
"x",
"+",
"sin_rad",
"*",
"y",
",",
"-",
"sin_rad",
"*",
"x",
"+",
"cos_rad",
"*",
"y"
] |
rotate x,y vector over x2-x1, y2-y1 angle
|
[
"rotate",
"x",
"y",
"vector",
"over",
"x2",
"-",
"x1",
"y2",
"-",
"y1",
"angle"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/algorithms/depict.py#L86-L93
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
_build_static_table_mapping
|
def _build_static_table_mapping():
"""
Build static table mapping from header name to tuple with next structure:
(<minimal index of header>, <mapping from header value to it index>).
static_table_mapping used for hash searching.
"""
static_table_mapping = {}
for index, (name, value) in enumerate(CocaineHeaders.STATIC_TABLE, 1):
header_name_search_result = static_table_mapping.setdefault(name, (index, {}))
header_name_search_result[1][value] = index
return static_table_mapping
|
python
|
def _build_static_table_mapping():
"""
Build static table mapping from header name to tuple with next structure:
(<minimal index of header>, <mapping from header value to it index>).
static_table_mapping used for hash searching.
"""
static_table_mapping = {}
for index, (name, value) in enumerate(CocaineHeaders.STATIC_TABLE, 1):
header_name_search_result = static_table_mapping.setdefault(name, (index, {}))
header_name_search_result[1][value] = index
return static_table_mapping
|
[
"def",
"_build_static_table_mapping",
"(",
")",
":",
"static_table_mapping",
"=",
"{",
"}",
"for",
"index",
",",
"(",
"name",
",",
"value",
")",
"in",
"enumerate",
"(",
"CocaineHeaders",
".",
"STATIC_TABLE",
",",
"1",
")",
":",
"header_name_search_result",
"=",
"static_table_mapping",
".",
"setdefault",
"(",
"name",
",",
"(",
"index",
",",
"{",
"}",
")",
")",
"header_name_search_result",
"[",
"1",
"]",
"[",
"value",
"]",
"=",
"index",
"return",
"static_table_mapping"
] |
Build static table mapping from header name to tuple with next structure:
(<minimal index of header>, <mapping from header value to it index>).
static_table_mapping used for hash searching.
|
[
"Build",
"static",
"table",
"mapping",
"from",
"header",
"name",
"to",
"tuple",
"with",
"next",
"structure",
":",
"(",
"<minimal",
"index",
"of",
"header",
">",
"<mapping",
"from",
"header",
"value",
"to",
"it",
"index",
">",
")",
"."
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L268-L279
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
CocaineHeaders.get_by_index
|
def get_by_index(self, index):
"""
Returns the entry specified by index
Note that the table is 1-based ie an index of 0 is
invalid. This is due to the fact that a zero value
index signals that a completely unindexed header
follows.
The entry will either be from the static table or
the dynamic table depending on the value of index.
"""
index -= 1
if 0 <= index < len(CocaineHeaders.STATIC_TABLE):
return CocaineHeaders.STATIC_TABLE[index]
index -= len(CocaineHeaders.STATIC_TABLE)
if 0 <= index < len(self.dynamic_entries):
return self.dynamic_entries[index]
raise InvalidTableIndex("Invalid table index %d" % index)
|
python
|
def get_by_index(self, index):
"""
Returns the entry specified by index
Note that the table is 1-based ie an index of 0 is
invalid. This is due to the fact that a zero value
index signals that a completely unindexed header
follows.
The entry will either be from the static table or
the dynamic table depending on the value of index.
"""
index -= 1
if 0 <= index < len(CocaineHeaders.STATIC_TABLE):
return CocaineHeaders.STATIC_TABLE[index]
index -= len(CocaineHeaders.STATIC_TABLE)
if 0 <= index < len(self.dynamic_entries):
return self.dynamic_entries[index]
raise InvalidTableIndex("Invalid table index %d" % index)
|
[
"def",
"get_by_index",
"(",
"self",
",",
"index",
")",
":",
"index",
"-=",
"1",
"if",
"0",
"<=",
"index",
"<",
"len",
"(",
"CocaineHeaders",
".",
"STATIC_TABLE",
")",
":",
"return",
"CocaineHeaders",
".",
"STATIC_TABLE",
"[",
"index",
"]",
"index",
"-=",
"len",
"(",
"CocaineHeaders",
".",
"STATIC_TABLE",
")",
"if",
"0",
"<=",
"index",
"<",
"len",
"(",
"self",
".",
"dynamic_entries",
")",
":",
"return",
"self",
".",
"dynamic_entries",
"[",
"index",
"]",
"raise",
"InvalidTableIndex",
"(",
"\"Invalid table index %d\"",
"%",
"index",
")"
] |
Returns the entry specified by index
Note that the table is 1-based ie an index of 0 is
invalid. This is due to the fact that a zero value
index signals that a completely unindexed header
follows.
The entry will either be from the static table or
the dynamic table depending on the value of index.
|
[
"Returns",
"the",
"entry",
"specified",
"by",
"index"
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L145-L163
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
CocaineHeaders.add
|
def add(self, name, value):
"""
Adds a new entry to the table
We reduce the table size if the entry will make the
table size greater than maxsize.
"""
# We just clear the table if the entry is too big
size = table_entry_size(name, value)
if size > self._maxsize:
self.dynamic_entries.clear()
self._current_size = 0
# Add new entry if the table actually has a size
elif self._maxsize > 0:
self.dynamic_entries.appendleft((name, value))
self._current_size += size
self._shrink()
|
python
|
def add(self, name, value):
"""
Adds a new entry to the table
We reduce the table size if the entry will make the
table size greater than maxsize.
"""
# We just clear the table if the entry is too big
size = table_entry_size(name, value)
if size > self._maxsize:
self.dynamic_entries.clear()
self._current_size = 0
# Add new entry if the table actually has a size
elif self._maxsize > 0:
self.dynamic_entries.appendleft((name, value))
self._current_size += size
self._shrink()
|
[
"def",
"add",
"(",
"self",
",",
"name",
",",
"value",
")",
":",
"# We just clear the table if the entry is too big",
"size",
"=",
"table_entry_size",
"(",
"name",
",",
"value",
")",
"if",
"size",
">",
"self",
".",
"_maxsize",
":",
"self",
".",
"dynamic_entries",
".",
"clear",
"(",
")",
"self",
".",
"_current_size",
"=",
"0",
"# Add new entry if the table actually has a size",
"elif",
"self",
".",
"_maxsize",
">",
"0",
":",
"self",
".",
"dynamic_entries",
".",
"appendleft",
"(",
"(",
"name",
",",
"value",
")",
")",
"self",
".",
"_current_size",
"+=",
"size",
"self",
".",
"_shrink",
"(",
")"
] |
Adds a new entry to the table
We reduce the table size if the entry will make the
table size greater than maxsize.
|
[
"Adds",
"a",
"new",
"entry",
"to",
"the",
"table"
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L172-L189
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
CocaineHeaders.search
|
def search(self, name, value):
"""
Searches the table for the entry specified by name
and value
Returns one of the following:
- ``None``, no match at all
- ``(index, name, None)`` for partial matches on name only.
- ``(index, name, value)`` for perfect matches.
"""
partial = None
header_name_search_result = CocaineHeaders.STATIC_TABLE_MAPPING.get(name)
if header_name_search_result:
index = header_name_search_result[1].get(value)
if index is not None:
return index, name, value
partial = (header_name_search_result[0], name, None)
offset = len(CocaineHeaders.STATIC_TABLE)
for (i, (n, v)) in enumerate(self.dynamic_entries):
if n == name:
if v == value:
return i + offset + 1, n, v
elif partial is None:
partial = (i + offset + 1, n, None)
return partial
|
python
|
def search(self, name, value):
"""
Searches the table for the entry specified by name
and value
Returns one of the following:
- ``None``, no match at all
- ``(index, name, None)`` for partial matches on name only.
- ``(index, name, value)`` for perfect matches.
"""
partial = None
header_name_search_result = CocaineHeaders.STATIC_TABLE_MAPPING.get(name)
if header_name_search_result:
index = header_name_search_result[1].get(value)
if index is not None:
return index, name, value
partial = (header_name_search_result[0], name, None)
offset = len(CocaineHeaders.STATIC_TABLE)
for (i, (n, v)) in enumerate(self.dynamic_entries):
if n == name:
if v == value:
return i + offset + 1, n, v
elif partial is None:
partial = (i + offset + 1, n, None)
return partial
|
[
"def",
"search",
"(",
"self",
",",
"name",
",",
"value",
")",
":",
"partial",
"=",
"None",
"header_name_search_result",
"=",
"CocaineHeaders",
".",
"STATIC_TABLE_MAPPING",
".",
"get",
"(",
"name",
")",
"if",
"header_name_search_result",
":",
"index",
"=",
"header_name_search_result",
"[",
"1",
"]",
".",
"get",
"(",
"value",
")",
"if",
"index",
"is",
"not",
"None",
":",
"return",
"index",
",",
"name",
",",
"value",
"partial",
"=",
"(",
"header_name_search_result",
"[",
"0",
"]",
",",
"name",
",",
"None",
")",
"offset",
"=",
"len",
"(",
"CocaineHeaders",
".",
"STATIC_TABLE",
")",
"for",
"(",
"i",
",",
"(",
"n",
",",
"v",
")",
")",
"in",
"enumerate",
"(",
"self",
".",
"dynamic_entries",
")",
":",
"if",
"n",
"==",
"name",
":",
"if",
"v",
"==",
"value",
":",
"return",
"i",
"+",
"offset",
"+",
"1",
",",
"n",
",",
"v",
"elif",
"partial",
"is",
"None",
":",
"partial",
"=",
"(",
"i",
"+",
"offset",
"+",
"1",
",",
"n",
",",
"None",
")",
"return",
"partial"
] |
Searches the table for the entry specified by name
and value
Returns one of the following:
- ``None``, no match at all
- ``(index, name, None)`` for partial matches on name only.
- ``(index, name, value)`` for perfect matches.
|
[
"Searches",
"the",
"table",
"for",
"the",
"entry",
"specified",
"by",
"name",
"and",
"value"
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L191-L217
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
CocaineHeaders._shrink
|
def _shrink(self):
"""
Shrinks the dynamic table to be at or below maxsize
"""
cursize = self._current_size
while cursize > self._maxsize:
name, value = self.dynamic_entries.pop()
cursize -= table_entry_size(name, value)
self._current_size = cursize
|
python
|
def _shrink(self):
"""
Shrinks the dynamic table to be at or below maxsize
"""
cursize = self._current_size
while cursize > self._maxsize:
name, value = self.dynamic_entries.pop()
cursize -= table_entry_size(name, value)
self._current_size = cursize
|
[
"def",
"_shrink",
"(",
"self",
")",
":",
"cursize",
"=",
"self",
".",
"_current_size",
"while",
"cursize",
">",
"self",
".",
"_maxsize",
":",
"name",
",",
"value",
"=",
"self",
".",
"dynamic_entries",
".",
"pop",
"(",
")",
"cursize",
"-=",
"table_entry_size",
"(",
"name",
",",
"value",
")",
"self",
".",
"_current_size",
"=",
"cursize"
] |
Shrinks the dynamic table to be at or below maxsize
|
[
"Shrinks",
"the",
"dynamic",
"table",
"to",
"be",
"at",
"or",
"below",
"maxsize"
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L235-L243
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
Headers.add
|
def add(self, name, value):
# type: (str, str) -> None
"""Adds a new value for the given key."""
self._last_key = name
if name in self:
self._dict[name] = value
self._as_list[name].append(value)
else:
self[name] = value
|
python
|
def add(self, name, value):
# type: (str, str) -> None
"""Adds a new value for the given key."""
self._last_key = name
if name in self:
self._dict[name] = value
self._as_list[name].append(value)
else:
self[name] = value
|
[
"def",
"add",
"(",
"self",
",",
"name",
",",
"value",
")",
":",
"# type: (str, str) -> None",
"self",
".",
"_last_key",
"=",
"name",
"if",
"name",
"in",
"self",
":",
"self",
".",
"_dict",
"[",
"name",
"]",
"=",
"value",
"self",
".",
"_as_list",
"[",
"name",
"]",
".",
"append",
"(",
"value",
")",
"else",
":",
"self",
"[",
"name",
"]",
"=",
"value"
] |
Adds a new value for the given key.
|
[
"Adds",
"a",
"new",
"value",
"for",
"the",
"given",
"key",
"."
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L314-L322
|
train
|
cocaine/cocaine-framework-python
|
cocaine/detail/headers.py
|
Headers.get_all
|
def get_all(self):
# type: () -> typing.Iterable[typing.Tuple[str, str]]
"""Returns an iterable of all (name, value) pairs.
If a header has multiple values, multiple pairs will be
returned with the same name.
"""
for name, values in six.iteritems(self._as_list):
for value in values:
yield (name, value)
|
python
|
def get_all(self):
# type: () -> typing.Iterable[typing.Tuple[str, str]]
"""Returns an iterable of all (name, value) pairs.
If a header has multiple values, multiple pairs will be
returned with the same name.
"""
for name, values in six.iteritems(self._as_list):
for value in values:
yield (name, value)
|
[
"def",
"get_all",
"(",
"self",
")",
":",
"# type: () -> typing.Iterable[typing.Tuple[str, str]]",
"for",
"name",
",",
"values",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_as_list",
")",
":",
"for",
"value",
"in",
"values",
":",
"yield",
"(",
"name",
",",
"value",
")"
] |
Returns an iterable of all (name, value) pairs.
If a header has multiple values, multiple pairs will be
returned with the same name.
|
[
"Returns",
"an",
"iterable",
"of",
"all",
"(",
"name",
"value",
")",
"pairs",
"."
] |
d8a30074b6338bac4389eb996e00d404338115e4
|
https://github.com/cocaine/cocaine-framework-python/blob/d8a30074b6338bac4389eb996e00d404338115e4/cocaine/detail/headers.py#L328-L337
|
train
|
camptocamp/marabunta
|
marabunta/output.py
|
safe_print
|
def safe_print(ustring, errors='replace', **kwargs):
""" Safely print a unicode string """
encoding = sys.stdout.encoding or 'utf-8'
if sys.version_info[0] == 3:
print(ustring, **kwargs)
else:
bytestr = ustring.encode(encoding, errors=errors)
print(bytestr, **kwargs)
|
python
|
def safe_print(ustring, errors='replace', **kwargs):
""" Safely print a unicode string """
encoding = sys.stdout.encoding or 'utf-8'
if sys.version_info[0] == 3:
print(ustring, **kwargs)
else:
bytestr = ustring.encode(encoding, errors=errors)
print(bytestr, **kwargs)
|
[
"def",
"safe_print",
"(",
"ustring",
",",
"errors",
"=",
"'replace'",
",",
"*",
"*",
"kwargs",
")",
":",
"encoding",
"=",
"sys",
".",
"stdout",
".",
"encoding",
"or",
"'utf-8'",
"if",
"sys",
".",
"version_info",
"[",
"0",
"]",
"==",
"3",
":",
"print",
"(",
"ustring",
",",
"*",
"*",
"kwargs",
")",
"else",
":",
"bytestr",
"=",
"ustring",
".",
"encode",
"(",
"encoding",
",",
"errors",
"=",
"errors",
")",
"print",
"(",
"bytestr",
",",
"*",
"*",
"kwargs",
")"
] |
Safely print a unicode string
|
[
"Safely",
"print",
"a",
"unicode",
"string"
] |
ec3a7a725c7426d6ed642e0a80119b37880eb91e
|
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/output.py#L26-L33
|
train
|
zetaops/zengine
|
zengine/views/permissions.py
|
Permissions.edit_permissions
|
def edit_permissions(self):
"""Creates the view used to edit permissions.
To create the view, data in the following format is passed to the UI
in the objects field:
.. code-block:: python
{
"type": "tree-toggle",
"action": "set_permission",
"tree": [
{
"checked": true,
"name": "Workflow 1 Name",
"id": "workflow1",
"children":
[
{
"checked": true,
"name": "Task 1 Name",
"id": "workflow1..task1",
"children": []
},
{
"checked": false,
"id": "workflow1..task2",
"name": "Task 2 Name",
"children": []
}
]
},
{
"checked": true,
"name": "Workflow 2 Name",
"id": "workflow2",
"children": [
{
"checked": true,
"name": "Workflow 2 Lane 1 Name",
"id": "workflow2.lane1",
"children": [
{
"checked": true,
"name": "Workflow 2 Task 1 Name",
"id": "workflow2.lane1.task1",
"children": []
},
{
"checked": false,
"name": "Workflow 2 Task 2 Name",
"id": "workflow2.lane1.task2",
"children": []
}
]
}
]
}
]
}
"type" field denotes that the object is a tree view which has elements that can be toggled.
"action" field is the
"name" field is the human readable name.
"id" field is used to make requests to the backend.
"checked" field shows whether the role has the permission or not.
"children" field is the sub-permissions of the permission.
"""
# Get the role that was selected in the CRUD view
key = self.current.input['object_id']
self.current.task_data['role_id'] = key
role = RoleModel.objects.get(key=key)
# Get the cached permission tree, or build a new one if there is none cached
# TODO: Add an extra view in case there was no cache, as in 'please wait calculating permissions'
permission_tree = self._permission_trees(PermissionModel.objects)
# Apply the selected role to the permission tree, setting the 'checked' field
# of the permission the role has
role_tree = self._apply_role_tree(permission_tree, role)
# Apply final formatting, and output the tree to the UI
self.output['objects'] = [
{
'type': 'tree-toggle',
'action': 'apply_change',
'trees': self._format_tree_output(role_tree),
},
]
self.form_out(PermissionForm())
|
python
|
def edit_permissions(self):
"""Creates the view used to edit permissions.
To create the view, data in the following format is passed to the UI
in the objects field:
.. code-block:: python
{
"type": "tree-toggle",
"action": "set_permission",
"tree": [
{
"checked": true,
"name": "Workflow 1 Name",
"id": "workflow1",
"children":
[
{
"checked": true,
"name": "Task 1 Name",
"id": "workflow1..task1",
"children": []
},
{
"checked": false,
"id": "workflow1..task2",
"name": "Task 2 Name",
"children": []
}
]
},
{
"checked": true,
"name": "Workflow 2 Name",
"id": "workflow2",
"children": [
{
"checked": true,
"name": "Workflow 2 Lane 1 Name",
"id": "workflow2.lane1",
"children": [
{
"checked": true,
"name": "Workflow 2 Task 1 Name",
"id": "workflow2.lane1.task1",
"children": []
},
{
"checked": false,
"name": "Workflow 2 Task 2 Name",
"id": "workflow2.lane1.task2",
"children": []
}
]
}
]
}
]
}
"type" field denotes that the object is a tree view which has elements that can be toggled.
"action" field is the
"name" field is the human readable name.
"id" field is used to make requests to the backend.
"checked" field shows whether the role has the permission or not.
"children" field is the sub-permissions of the permission.
"""
# Get the role that was selected in the CRUD view
key = self.current.input['object_id']
self.current.task_data['role_id'] = key
role = RoleModel.objects.get(key=key)
# Get the cached permission tree, or build a new one if there is none cached
# TODO: Add an extra view in case there was no cache, as in 'please wait calculating permissions'
permission_tree = self._permission_trees(PermissionModel.objects)
# Apply the selected role to the permission tree, setting the 'checked' field
# of the permission the role has
role_tree = self._apply_role_tree(permission_tree, role)
# Apply final formatting, and output the tree to the UI
self.output['objects'] = [
{
'type': 'tree-toggle',
'action': 'apply_change',
'trees': self._format_tree_output(role_tree),
},
]
self.form_out(PermissionForm())
|
[
"def",
"edit_permissions",
"(",
"self",
")",
":",
"# Get the role that was selected in the CRUD view",
"key",
"=",
"self",
".",
"current",
".",
"input",
"[",
"'object_id'",
"]",
"self",
".",
"current",
".",
"task_data",
"[",
"'role_id'",
"]",
"=",
"key",
"role",
"=",
"RoleModel",
".",
"objects",
".",
"get",
"(",
"key",
"=",
"key",
")",
"# Get the cached permission tree, or build a new one if there is none cached",
"# TODO: Add an extra view in case there was no cache, as in 'please wait calculating permissions'",
"permission_tree",
"=",
"self",
".",
"_permission_trees",
"(",
"PermissionModel",
".",
"objects",
")",
"# Apply the selected role to the permission tree, setting the 'checked' field",
"# of the permission the role has",
"role_tree",
"=",
"self",
".",
"_apply_role_tree",
"(",
"permission_tree",
",",
"role",
")",
"# Apply final formatting, and output the tree to the UI",
"self",
".",
"output",
"[",
"'objects'",
"]",
"=",
"[",
"{",
"'type'",
":",
"'tree-toggle'",
",",
"'action'",
":",
"'apply_change'",
",",
"'trees'",
":",
"self",
".",
"_format_tree_output",
"(",
"role_tree",
")",
",",
"}",
",",
"]",
"self",
".",
"form_out",
"(",
"PermissionForm",
"(",
")",
")"
] |
Creates the view used to edit permissions.
To create the view, data in the following format is passed to the UI
in the objects field:
.. code-block:: python
{
"type": "tree-toggle",
"action": "set_permission",
"tree": [
{
"checked": true,
"name": "Workflow 1 Name",
"id": "workflow1",
"children":
[
{
"checked": true,
"name": "Task 1 Name",
"id": "workflow1..task1",
"children": []
},
{
"checked": false,
"id": "workflow1..task2",
"name": "Task 2 Name",
"children": []
}
]
},
{
"checked": true,
"name": "Workflow 2 Name",
"id": "workflow2",
"children": [
{
"checked": true,
"name": "Workflow 2 Lane 1 Name",
"id": "workflow2.lane1",
"children": [
{
"checked": true,
"name": "Workflow 2 Task 1 Name",
"id": "workflow2.lane1.task1",
"children": []
},
{
"checked": false,
"name": "Workflow 2 Task 2 Name",
"id": "workflow2.lane1.task2",
"children": []
}
]
}
]
}
]
}
"type" field denotes that the object is a tree view which has elements that can be toggled.
"action" field is the
"name" field is the human readable name.
"id" field is used to make requests to the backend.
"checked" field shows whether the role has the permission or not.
"children" field is the sub-permissions of the permission.
|
[
"Creates",
"the",
"view",
"used",
"to",
"edit",
"permissions",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/permissions.py#L110-L196
|
train
|
zetaops/zengine
|
zengine/views/permissions.py
|
Permissions._permission_trees
|
def _permission_trees(permissions):
"""Get the cached permission tree, or build a new one if necessary."""
treecache = PermissionTreeCache()
cached = treecache.get()
if not cached:
tree = PermissionTreeBuilder()
for permission in permissions:
tree.insert(permission)
result = tree.serialize()
treecache.set(result)
return result
return cached
|
python
|
def _permission_trees(permissions):
"""Get the cached permission tree, or build a new one if necessary."""
treecache = PermissionTreeCache()
cached = treecache.get()
if not cached:
tree = PermissionTreeBuilder()
for permission in permissions:
tree.insert(permission)
result = tree.serialize()
treecache.set(result)
return result
return cached
|
[
"def",
"_permission_trees",
"(",
"permissions",
")",
":",
"treecache",
"=",
"PermissionTreeCache",
"(",
")",
"cached",
"=",
"treecache",
".",
"get",
"(",
")",
"if",
"not",
"cached",
":",
"tree",
"=",
"PermissionTreeBuilder",
"(",
")",
"for",
"permission",
"in",
"permissions",
":",
"tree",
".",
"insert",
"(",
"permission",
")",
"result",
"=",
"tree",
".",
"serialize",
"(",
")",
"treecache",
".",
"set",
"(",
"result",
")",
"return",
"result",
"return",
"cached"
] |
Get the cached permission tree, or build a new one if necessary.
|
[
"Get",
"the",
"cached",
"permission",
"tree",
"or",
"build",
"a",
"new",
"one",
"if",
"necessary",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/permissions.py#L199-L210
|
train
|
zetaops/zengine
|
zengine/views/permissions.py
|
Permissions._apply_role_tree
|
def _apply_role_tree(self, perm_tree, role):
"""In permission tree, sets `'checked': True` for the permissions that the role has."""
role_permissions = role.get_permissions()
for perm in role_permissions:
self._traverse_tree(perm_tree, perm)['checked'] = True
return perm_tree
|
python
|
def _apply_role_tree(self, perm_tree, role):
"""In permission tree, sets `'checked': True` for the permissions that the role has."""
role_permissions = role.get_permissions()
for perm in role_permissions:
self._traverse_tree(perm_tree, perm)['checked'] = True
return perm_tree
|
[
"def",
"_apply_role_tree",
"(",
"self",
",",
"perm_tree",
",",
"role",
")",
":",
"role_permissions",
"=",
"role",
".",
"get_permissions",
"(",
")",
"for",
"perm",
"in",
"role_permissions",
":",
"self",
".",
"_traverse_tree",
"(",
"perm_tree",
",",
"perm",
")",
"[",
"'checked'",
"]",
"=",
"True",
"return",
"perm_tree"
] |
In permission tree, sets `'checked': True` for the permissions that the role has.
|
[
"In",
"permission",
"tree",
"sets",
"checked",
":",
"True",
"for",
"the",
"permissions",
"that",
"the",
"role",
"has",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/permissions.py#L212-L217
|
train
|
zetaops/zengine
|
zengine/views/permissions.py
|
Permissions._traverse_tree
|
def _traverse_tree(tree, path):
"""Traverses the permission tree, returning the permission at given permission path."""
path_steps = (step for step in path.split('.') if step != '')
# Special handling for first step, because the first step isn't under 'objects'
first_step = path_steps.next()
subtree = tree[first_step]
for step in path_steps:
subtree = subtree['children'][step]
return subtree
|
python
|
def _traverse_tree(tree, path):
"""Traverses the permission tree, returning the permission at given permission path."""
path_steps = (step for step in path.split('.') if step != '')
# Special handling for first step, because the first step isn't under 'objects'
first_step = path_steps.next()
subtree = tree[first_step]
for step in path_steps:
subtree = subtree['children'][step]
return subtree
|
[
"def",
"_traverse_tree",
"(",
"tree",
",",
"path",
")",
":",
"path_steps",
"=",
"(",
"step",
"for",
"step",
"in",
"path",
".",
"split",
"(",
"'.'",
")",
"if",
"step",
"!=",
"''",
")",
"# Special handling for first step, because the first step isn't under 'objects'",
"first_step",
"=",
"path_steps",
".",
"next",
"(",
")",
"subtree",
"=",
"tree",
"[",
"first_step",
"]",
"for",
"step",
"in",
"path_steps",
":",
"subtree",
"=",
"subtree",
"[",
"'children'",
"]",
"[",
"step",
"]",
"return",
"subtree"
] |
Traverses the permission tree, returning the permission at given permission path.
|
[
"Traverses",
"the",
"permission",
"tree",
"returning",
"the",
"permission",
"at",
"given",
"permission",
"path",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/permissions.py#L220-L229
|
train
|
zetaops/zengine
|
zengine/views/permissions.py
|
Permissions._format_subtree
|
def _format_subtree(self, subtree):
"""Recursively format all subtrees."""
subtree['children'] = list(subtree['children'].values())
for child in subtree['children']:
self._format_subtree(child)
return subtree
|
python
|
def _format_subtree(self, subtree):
"""Recursively format all subtrees."""
subtree['children'] = list(subtree['children'].values())
for child in subtree['children']:
self._format_subtree(child)
return subtree
|
[
"def",
"_format_subtree",
"(",
"self",
",",
"subtree",
")",
":",
"subtree",
"[",
"'children'",
"]",
"=",
"list",
"(",
"subtree",
"[",
"'children'",
"]",
".",
"values",
"(",
")",
")",
"for",
"child",
"in",
"subtree",
"[",
"'children'",
"]",
":",
"self",
".",
"_format_subtree",
"(",
"child",
")",
"return",
"subtree"
] |
Recursively format all subtrees.
|
[
"Recursively",
"format",
"all",
"subtrees",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/permissions.py#L241-L246
|
train
|
zetaops/zengine
|
zengine/views/permissions.py
|
Permissions.apply_change
|
def apply_change(self):
"""Applies changes to the permissions of the role.
To make a change to the permission of the role, a request
in the following format should be sent:
.. code-block:: python
{
'change':
{
'id': 'workflow2.lane1.task1',
'checked': false
},
}
The 'id' field of the change is the id of the tree element
that was sent to the UI (see `Permissions.edit_permissions`).
'checked' field is the new state of the element.
"""
changes = self.input['change']
key = self.current.task_data['role_id']
role = RoleModel.objects.get(key=key)
for change in changes:
permission = PermissionModel.objects.get(code=change['id'])
if change['checked'] is True:
role.add_permission(permission)
else:
role.remove_permission(permission)
role.save()
|
python
|
def apply_change(self):
"""Applies changes to the permissions of the role.
To make a change to the permission of the role, a request
in the following format should be sent:
.. code-block:: python
{
'change':
{
'id': 'workflow2.lane1.task1',
'checked': false
},
}
The 'id' field of the change is the id of the tree element
that was sent to the UI (see `Permissions.edit_permissions`).
'checked' field is the new state of the element.
"""
changes = self.input['change']
key = self.current.task_data['role_id']
role = RoleModel.objects.get(key=key)
for change in changes:
permission = PermissionModel.objects.get(code=change['id'])
if change['checked'] is True:
role.add_permission(permission)
else:
role.remove_permission(permission)
role.save()
|
[
"def",
"apply_change",
"(",
"self",
")",
":",
"changes",
"=",
"self",
".",
"input",
"[",
"'change'",
"]",
"key",
"=",
"self",
".",
"current",
".",
"task_data",
"[",
"'role_id'",
"]",
"role",
"=",
"RoleModel",
".",
"objects",
".",
"get",
"(",
"key",
"=",
"key",
")",
"for",
"change",
"in",
"changes",
":",
"permission",
"=",
"PermissionModel",
".",
"objects",
".",
"get",
"(",
"code",
"=",
"change",
"[",
"'id'",
"]",
")",
"if",
"change",
"[",
"'checked'",
"]",
"is",
"True",
":",
"role",
".",
"add_permission",
"(",
"permission",
")",
"else",
":",
"role",
".",
"remove_permission",
"(",
"permission",
")",
"role",
".",
"save",
"(",
")"
] |
Applies changes to the permissions of the role.
To make a change to the permission of the role, a request
in the following format should be sent:
.. code-block:: python
{
'change':
{
'id': 'workflow2.lane1.task1',
'checked': false
},
}
The 'id' field of the change is the id of the tree element
that was sent to the UI (see `Permissions.edit_permissions`).
'checked' field is the new state of the element.
|
[
"Applies",
"changes",
"to",
"the",
"permissions",
"of",
"the",
"role",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/permissions.py#L248-L276
|
train
|
cimm-kzn/CGRtools
|
CGRtools/files/SDFrw.py
|
SDFread.seek
|
def seek(self, offset):
"""
shifts on a given number of record in the original file
:param offset: number of record
"""
if self._shifts:
if 0 <= offset < len(self._shifts):
current_pos = self._file.tell()
new_pos = self._shifts[offset]
if current_pos != new_pos:
if current_pos == self._shifts[-1]: # reached the end of the file
self._data = self.__reader()
self.__file = iter(self._file.readline, '')
self._file.seek(new_pos)
else:
raise IndexError('invalid offset')
else:
raise self._implement_error
|
python
|
def seek(self, offset):
"""
shifts on a given number of record in the original file
:param offset: number of record
"""
if self._shifts:
if 0 <= offset < len(self._shifts):
current_pos = self._file.tell()
new_pos = self._shifts[offset]
if current_pos != new_pos:
if current_pos == self._shifts[-1]: # reached the end of the file
self._data = self.__reader()
self.__file = iter(self._file.readline, '')
self._file.seek(new_pos)
else:
raise IndexError('invalid offset')
else:
raise self._implement_error
|
[
"def",
"seek",
"(",
"self",
",",
"offset",
")",
":",
"if",
"self",
".",
"_shifts",
":",
"if",
"0",
"<=",
"offset",
"<",
"len",
"(",
"self",
".",
"_shifts",
")",
":",
"current_pos",
"=",
"self",
".",
"_file",
".",
"tell",
"(",
")",
"new_pos",
"=",
"self",
".",
"_shifts",
"[",
"offset",
"]",
"if",
"current_pos",
"!=",
"new_pos",
":",
"if",
"current_pos",
"==",
"self",
".",
"_shifts",
"[",
"-",
"1",
"]",
":",
"# reached the end of the file",
"self",
".",
"_data",
"=",
"self",
".",
"__reader",
"(",
")",
"self",
".",
"__file",
"=",
"iter",
"(",
"self",
".",
"_file",
".",
"readline",
",",
"''",
")",
"self",
".",
"_file",
".",
"seek",
"(",
"new_pos",
")",
"else",
":",
"raise",
"IndexError",
"(",
"'invalid offset'",
")",
"else",
":",
"raise",
"self",
".",
"_implement_error"
] |
shifts on a given number of record in the original file
:param offset: number of record
|
[
"shifts",
"on",
"a",
"given",
"number",
"of",
"record",
"in",
"the",
"original",
"file",
":",
"param",
"offset",
":",
"number",
"of",
"record"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/files/SDFrw.py#L62-L79
|
train
|
cimm-kzn/CGRtools
|
CGRtools/files/SDFrw.py
|
SDFread.tell
|
def tell(self):
"""
:return: number of records processed from the original file
"""
if self._shifts:
t = self._file.tell()
return bisect_left(self._shifts, t)
raise self._implement_error
|
python
|
def tell(self):
"""
:return: number of records processed from the original file
"""
if self._shifts:
t = self._file.tell()
return bisect_left(self._shifts, t)
raise self._implement_error
|
[
"def",
"tell",
"(",
"self",
")",
":",
"if",
"self",
".",
"_shifts",
":",
"t",
"=",
"self",
".",
"_file",
".",
"tell",
"(",
")",
"return",
"bisect_left",
"(",
"self",
".",
"_shifts",
",",
"t",
")",
"raise",
"self",
".",
"_implement_error"
] |
:return: number of records processed from the original file
|
[
":",
"return",
":",
"number",
"of",
"records",
"processed",
"from",
"the",
"original",
"file"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/files/SDFrw.py#L81-L88
|
train
|
cimm-kzn/CGRtools
|
CGRtools/files/SDFrw.py
|
SDFwrite.write
|
def write(self, data):
"""
write single molecule into file
"""
m = self._convert_structure(data)
self._file.write(self._format_mol(*m))
self._file.write('M END\n')
for k, v in data.meta.items():
self._file.write(f'> <{k}>\n{v}\n')
self._file.write('$$$$\n')
|
python
|
def write(self, data):
"""
write single molecule into file
"""
m = self._convert_structure(data)
self._file.write(self._format_mol(*m))
self._file.write('M END\n')
for k, v in data.meta.items():
self._file.write(f'> <{k}>\n{v}\n')
self._file.write('$$$$\n')
|
[
"def",
"write",
"(",
"self",
",",
"data",
")",
":",
"m",
"=",
"self",
".",
"_convert_structure",
"(",
"data",
")",
"self",
".",
"_file",
".",
"write",
"(",
"self",
".",
"_format_mol",
"(",
"*",
"m",
")",
")",
"self",
".",
"_file",
".",
"write",
"(",
"'M END\\n'",
")",
"for",
"k",
",",
"v",
"in",
"data",
".",
"meta",
".",
"items",
"(",
")",
":",
"self",
".",
"_file",
".",
"write",
"(",
"f'> <{k}>\\n{v}\\n'",
")",
"self",
".",
"_file",
".",
"write",
"(",
"'$$$$\\n'",
")"
] |
write single molecule into file
|
[
"write",
"single",
"molecule",
"into",
"file"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/files/SDFrw.py#L166-L176
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.save_workflow_to_cache
|
def save_workflow_to_cache(self, serialized_wf_instance):
"""
If we aren't come to the end of the wf,
saves the wf state and task_data to cache
Task_data items that starts with underscore "_" are treated as
local and does not passed to subsequent task steps.
"""
# self.current.task_data['flow'] = None
task_data = self.current.task_data.copy()
for k, v in list(task_data.items()):
if k.startswith('_'):
del task_data[k]
if 'cmd' in task_data:
del task_data['cmd']
self.wf_state.update({'step': serialized_wf_instance,
'data': task_data,
'name': self.current.workflow_name,
'wf_id': self.workflow_spec.wf_id
})
if self.current.lane_id:
self.current.pool[self.current.lane_id] = self.current.role.key
self.wf_state['pool'] = self.current.pool
self.current.log.debug("POOL Content before WF Save: %s" % self.current.pool)
self.current.wf_cache.save(self.wf_state)
|
python
|
def save_workflow_to_cache(self, serialized_wf_instance):
"""
If we aren't come to the end of the wf,
saves the wf state and task_data to cache
Task_data items that starts with underscore "_" are treated as
local and does not passed to subsequent task steps.
"""
# self.current.task_data['flow'] = None
task_data = self.current.task_data.copy()
for k, v in list(task_data.items()):
if k.startswith('_'):
del task_data[k]
if 'cmd' in task_data:
del task_data['cmd']
self.wf_state.update({'step': serialized_wf_instance,
'data': task_data,
'name': self.current.workflow_name,
'wf_id': self.workflow_spec.wf_id
})
if self.current.lane_id:
self.current.pool[self.current.lane_id] = self.current.role.key
self.wf_state['pool'] = self.current.pool
self.current.log.debug("POOL Content before WF Save: %s" % self.current.pool)
self.current.wf_cache.save(self.wf_state)
|
[
"def",
"save_workflow_to_cache",
"(",
"self",
",",
"serialized_wf_instance",
")",
":",
"# self.current.task_data['flow'] = None",
"task_data",
"=",
"self",
".",
"current",
".",
"task_data",
".",
"copy",
"(",
")",
"for",
"k",
",",
"v",
"in",
"list",
"(",
"task_data",
".",
"items",
"(",
")",
")",
":",
"if",
"k",
".",
"startswith",
"(",
"'_'",
")",
":",
"del",
"task_data",
"[",
"k",
"]",
"if",
"'cmd'",
"in",
"task_data",
":",
"del",
"task_data",
"[",
"'cmd'",
"]",
"self",
".",
"wf_state",
".",
"update",
"(",
"{",
"'step'",
":",
"serialized_wf_instance",
",",
"'data'",
":",
"task_data",
",",
"'name'",
":",
"self",
".",
"current",
".",
"workflow_name",
",",
"'wf_id'",
":",
"self",
".",
"workflow_spec",
".",
"wf_id",
"}",
")",
"if",
"self",
".",
"current",
".",
"lane_id",
":",
"self",
".",
"current",
".",
"pool",
"[",
"self",
".",
"current",
".",
"lane_id",
"]",
"=",
"self",
".",
"current",
".",
"role",
".",
"key",
"self",
".",
"wf_state",
"[",
"'pool'",
"]",
"=",
"self",
".",
"current",
".",
"pool",
"self",
".",
"current",
".",
"log",
".",
"debug",
"(",
"\"POOL Content before WF Save: %s\"",
"%",
"self",
".",
"current",
".",
"pool",
")",
"self",
".",
"current",
".",
"wf_cache",
".",
"save",
"(",
"self",
".",
"wf_state",
")"
] |
If we aren't come to the end of the wf,
saves the wf state and task_data to cache
Task_data items that starts with underscore "_" are treated as
local and does not passed to subsequent task steps.
|
[
"If",
"we",
"aren",
"t",
"come",
"to",
"the",
"end",
"of",
"the",
"wf",
"saves",
"the",
"wf",
"state",
"and",
"task_data",
"to",
"cache"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L76-L102
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.get_pool_context
|
def get_pool_context(self):
# TODO: Add in-process caching
"""
Builds context for the WF pool.
Returns:
Context dict.
"""
context = {self.current.lane_id: self.current.role, 'self': self.current.role}
for lane_id, role_id in self.current.pool.items():
if role_id:
context[lane_id] = lazy_object_proxy.Proxy(
lambda: self.role_model(super_context).objects.get(role_id))
return context
|
python
|
def get_pool_context(self):
# TODO: Add in-process caching
"""
Builds context for the WF pool.
Returns:
Context dict.
"""
context = {self.current.lane_id: self.current.role, 'self': self.current.role}
for lane_id, role_id in self.current.pool.items():
if role_id:
context[lane_id] = lazy_object_proxy.Proxy(
lambda: self.role_model(super_context).objects.get(role_id))
return context
|
[
"def",
"get_pool_context",
"(",
"self",
")",
":",
"# TODO: Add in-process caching",
"context",
"=",
"{",
"self",
".",
"current",
".",
"lane_id",
":",
"self",
".",
"current",
".",
"role",
",",
"'self'",
":",
"self",
".",
"current",
".",
"role",
"}",
"for",
"lane_id",
",",
"role_id",
"in",
"self",
".",
"current",
".",
"pool",
".",
"items",
"(",
")",
":",
"if",
"role_id",
":",
"context",
"[",
"lane_id",
"]",
"=",
"lazy_object_proxy",
".",
"Proxy",
"(",
"lambda",
":",
"self",
".",
"role_model",
"(",
"super_context",
")",
".",
"objects",
".",
"get",
"(",
"role_id",
")",
")",
"return",
"context"
] |
Builds context for the WF pool.
Returns:
Context dict.
|
[
"Builds",
"context",
"for",
"the",
"WF",
"pool",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L104-L117
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.load_workflow_from_cache
|
def load_workflow_from_cache(self):
"""
loads the serialized wf state and data from cache
updates the self.current.task_data
"""
if not self.current.new_token:
self.wf_state = self.current.wf_cache.get(self.wf_state)
self.current.task_data = self.wf_state['data']
self.current.set_client_cmds()
self.current.pool = self.wf_state['pool']
return self.wf_state['step']
|
python
|
def load_workflow_from_cache(self):
"""
loads the serialized wf state and data from cache
updates the self.current.task_data
"""
if not self.current.new_token:
self.wf_state = self.current.wf_cache.get(self.wf_state)
self.current.task_data = self.wf_state['data']
self.current.set_client_cmds()
self.current.pool = self.wf_state['pool']
return self.wf_state['step']
|
[
"def",
"load_workflow_from_cache",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"current",
".",
"new_token",
":",
"self",
".",
"wf_state",
"=",
"self",
".",
"current",
".",
"wf_cache",
".",
"get",
"(",
"self",
".",
"wf_state",
")",
"self",
".",
"current",
".",
"task_data",
"=",
"self",
".",
"wf_state",
"[",
"'data'",
"]",
"self",
".",
"current",
".",
"set_client_cmds",
"(",
")",
"self",
".",
"current",
".",
"pool",
"=",
"self",
".",
"wf_state",
"[",
"'pool'",
"]",
"return",
"self",
".",
"wf_state",
"[",
"'step'",
"]"
] |
loads the serialized wf state and data from cache
updates the self.current.task_data
|
[
"loads",
"the",
"serialized",
"wf",
"state",
"and",
"data",
"from",
"cache",
"updates",
"the",
"self",
".",
"current",
".",
"task_data"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L119-L129
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.serialize_workflow
|
def serialize_workflow(self):
"""
Serializes the current WF.
Returns:
WF state data.
"""
self.workflow.refresh_waiting_tasks()
return CompactWorkflowSerializer().serialize_workflow(self.workflow,
include_spec=False)
|
python
|
def serialize_workflow(self):
"""
Serializes the current WF.
Returns:
WF state data.
"""
self.workflow.refresh_waiting_tasks()
return CompactWorkflowSerializer().serialize_workflow(self.workflow,
include_spec=False)
|
[
"def",
"serialize_workflow",
"(",
"self",
")",
":",
"self",
".",
"workflow",
".",
"refresh_waiting_tasks",
"(",
")",
"return",
"CompactWorkflowSerializer",
"(",
")",
".",
"serialize_workflow",
"(",
"self",
".",
"workflow",
",",
"include_spec",
"=",
"False",
")"
] |
Serializes the current WF.
Returns:
WF state data.
|
[
"Serializes",
"the",
"current",
"WF",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L150-L159
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.load_or_create_workflow
|
def load_or_create_workflow(self):
"""
Tries to load the previously serialized (and saved) workflow
Creates a new one if it can't
"""
self.workflow_spec = self.get_worfklow_spec()
return self._load_workflow() or self.create_workflow()
|
python
|
def load_or_create_workflow(self):
"""
Tries to load the previously serialized (and saved) workflow
Creates a new one if it can't
"""
self.workflow_spec = self.get_worfklow_spec()
return self._load_workflow() or self.create_workflow()
|
[
"def",
"load_or_create_workflow",
"(",
"self",
")",
":",
"self",
".",
"workflow_spec",
"=",
"self",
".",
"get_worfklow_spec",
"(",
")",
"return",
"self",
".",
"_load_workflow",
"(",
")",
"or",
"self",
".",
"create_workflow",
"(",
")"
] |
Tries to load the previously serialized (and saved) workflow
Creates a new one if it can't
|
[
"Tries",
"to",
"load",
"the",
"previously",
"serialized",
"(",
"and",
"saved",
")",
"workflow",
"Creates",
"a",
"new",
"one",
"if",
"it",
"can",
"t"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L169-L175
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.find_workflow_path
|
def find_workflow_path(self):
"""
Tries to find the path of the workflow diagram file
in `WORKFLOW_PACKAGES_PATHS`.
Returns:
Path of the workflow spec file (BPMN diagram)
"""
for pth in settings.WORKFLOW_PACKAGES_PATHS:
path = "%s/%s.bpmn" % (pth, self.current.workflow_name)
if os.path.exists(path):
return path
err_msg = "BPMN file cannot found: %s" % self.current.workflow_name
log.error(err_msg)
raise RuntimeError(err_msg)
|
python
|
def find_workflow_path(self):
"""
Tries to find the path of the workflow diagram file
in `WORKFLOW_PACKAGES_PATHS`.
Returns:
Path of the workflow spec file (BPMN diagram)
"""
for pth in settings.WORKFLOW_PACKAGES_PATHS:
path = "%s/%s.bpmn" % (pth, self.current.workflow_name)
if os.path.exists(path):
return path
err_msg = "BPMN file cannot found: %s" % self.current.workflow_name
log.error(err_msg)
raise RuntimeError(err_msg)
|
[
"def",
"find_workflow_path",
"(",
"self",
")",
":",
"for",
"pth",
"in",
"settings",
".",
"WORKFLOW_PACKAGES_PATHS",
":",
"path",
"=",
"\"%s/%s.bpmn\"",
"%",
"(",
"pth",
",",
"self",
".",
"current",
".",
"workflow_name",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"path",
")",
":",
"return",
"path",
"err_msg",
"=",
"\"BPMN file cannot found: %s\"",
"%",
"self",
".",
"current",
".",
"workflow_name",
"log",
".",
"error",
"(",
"err_msg",
")",
"raise",
"RuntimeError",
"(",
"err_msg",
")"
] |
Tries to find the path of the workflow diagram file
in `WORKFLOW_PACKAGES_PATHS`.
Returns:
Path of the workflow spec file (BPMN diagram)
|
[
"Tries",
"to",
"find",
"the",
"path",
"of",
"the",
"workflow",
"diagram",
"file",
"in",
"WORKFLOW_PACKAGES_PATHS",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L178-L192
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.get_worfklow_spec
|
def get_worfklow_spec(self):
"""
Generates and caches the workflow spec package from
BPMN diagrams that read from disk
Returns:
SpiffWorkflow Spec object.
"""
# TODO: convert from in-process to redis based caching
if self.current.workflow_name not in self.workflow_spec_cache:
# path = self.find_workflow_path()
# spec_package = InMemoryPackager.package_in_memory(self.current.workflow_name, path)
# spec = BpmnSerializer().deserialize_workflow_spec(spec_package)
try:
self.current.wf_object = BPMNWorkflow.objects.get(name=self.current.workflow_name)
except ObjectDoesNotExist:
self.current.wf_object = BPMNWorkflow.objects.get(name='not_found')
self.current.task_data['non-existent-wf'] = self.current.workflow_name
self.current.workflow_name = 'not_found'
xml_content = self.current.wf_object.xml.body
spec = ZopsSerializer().deserialize_workflow_spec(xml_content, self.current.workflow_name)
spec.wf_id = self.current.wf_object.key
self.workflow_spec_cache[self.current.workflow_name] = spec
return self.workflow_spec_cache[self.current.workflow_name]
|
python
|
def get_worfklow_spec(self):
"""
Generates and caches the workflow spec package from
BPMN diagrams that read from disk
Returns:
SpiffWorkflow Spec object.
"""
# TODO: convert from in-process to redis based caching
if self.current.workflow_name not in self.workflow_spec_cache:
# path = self.find_workflow_path()
# spec_package = InMemoryPackager.package_in_memory(self.current.workflow_name, path)
# spec = BpmnSerializer().deserialize_workflow_spec(spec_package)
try:
self.current.wf_object = BPMNWorkflow.objects.get(name=self.current.workflow_name)
except ObjectDoesNotExist:
self.current.wf_object = BPMNWorkflow.objects.get(name='not_found')
self.current.task_data['non-existent-wf'] = self.current.workflow_name
self.current.workflow_name = 'not_found'
xml_content = self.current.wf_object.xml.body
spec = ZopsSerializer().deserialize_workflow_spec(xml_content, self.current.workflow_name)
spec.wf_id = self.current.wf_object.key
self.workflow_spec_cache[self.current.workflow_name] = spec
return self.workflow_spec_cache[self.current.workflow_name]
|
[
"def",
"get_worfklow_spec",
"(",
"self",
")",
":",
"# TODO: convert from in-process to redis based caching",
"if",
"self",
".",
"current",
".",
"workflow_name",
"not",
"in",
"self",
".",
"workflow_spec_cache",
":",
"# path = self.find_workflow_path()",
"# spec_package = InMemoryPackager.package_in_memory(self.current.workflow_name, path)",
"# spec = BpmnSerializer().deserialize_workflow_spec(spec_package)",
"try",
":",
"self",
".",
"current",
".",
"wf_object",
"=",
"BPMNWorkflow",
".",
"objects",
".",
"get",
"(",
"name",
"=",
"self",
".",
"current",
".",
"workflow_name",
")",
"except",
"ObjectDoesNotExist",
":",
"self",
".",
"current",
".",
"wf_object",
"=",
"BPMNWorkflow",
".",
"objects",
".",
"get",
"(",
"name",
"=",
"'not_found'",
")",
"self",
".",
"current",
".",
"task_data",
"[",
"'non-existent-wf'",
"]",
"=",
"self",
".",
"current",
".",
"workflow_name",
"self",
".",
"current",
".",
"workflow_name",
"=",
"'not_found'",
"xml_content",
"=",
"self",
".",
"current",
".",
"wf_object",
".",
"xml",
".",
"body",
"spec",
"=",
"ZopsSerializer",
"(",
")",
".",
"deserialize_workflow_spec",
"(",
"xml_content",
",",
"self",
".",
"current",
".",
"workflow_name",
")",
"spec",
".",
"wf_id",
"=",
"self",
".",
"current",
".",
"wf_object",
".",
"key",
"self",
".",
"workflow_spec_cache",
"[",
"self",
".",
"current",
".",
"workflow_name",
"]",
"=",
"spec",
"return",
"self",
".",
"workflow_spec_cache",
"[",
"self",
".",
"current",
".",
"workflow_name",
"]"
] |
Generates and caches the workflow spec package from
BPMN diagrams that read from disk
Returns:
SpiffWorkflow Spec object.
|
[
"Generates",
"and",
"caches",
"the",
"workflow",
"spec",
"package",
"from",
"BPMN",
"diagrams",
"that",
"read",
"from",
"disk"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L194-L219
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._save_or_delete_workflow
|
def _save_or_delete_workflow(self):
"""
Calls the real save method if we pass the beggining of the wf
"""
if not self.current.task_type.startswith('Start'):
if self.current.task_name.startswith('End') and not self.are_we_in_subprocess():
self.wf_state['finished'] = True
self.wf_state['finish_date'] = datetime.now().strftime(
settings.DATETIME_DEFAULT_FORMAT)
if self.current.workflow_name not in settings.EPHEMERAL_WORKFLOWS and not \
self.wf_state['in_external']:
wfi = WFCache(self.current).get_instance()
TaskInvitation.objects.filter(instance=wfi, role=self.current.role,
wf_name=wfi.wf.name).delete()
self.current.log.info("Delete WFCache: %s %s" % (self.current.workflow_name,
self.current.token))
self.save_workflow_to_cache(self.serialize_workflow())
|
python
|
def _save_or_delete_workflow(self):
"""
Calls the real save method if we pass the beggining of the wf
"""
if not self.current.task_type.startswith('Start'):
if self.current.task_name.startswith('End') and not self.are_we_in_subprocess():
self.wf_state['finished'] = True
self.wf_state['finish_date'] = datetime.now().strftime(
settings.DATETIME_DEFAULT_FORMAT)
if self.current.workflow_name not in settings.EPHEMERAL_WORKFLOWS and not \
self.wf_state['in_external']:
wfi = WFCache(self.current).get_instance()
TaskInvitation.objects.filter(instance=wfi, role=self.current.role,
wf_name=wfi.wf.name).delete()
self.current.log.info("Delete WFCache: %s %s" % (self.current.workflow_name,
self.current.token))
self.save_workflow_to_cache(self.serialize_workflow())
|
[
"def",
"_save_or_delete_workflow",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"current",
".",
"task_type",
".",
"startswith",
"(",
"'Start'",
")",
":",
"if",
"self",
".",
"current",
".",
"task_name",
".",
"startswith",
"(",
"'End'",
")",
"and",
"not",
"self",
".",
"are_we_in_subprocess",
"(",
")",
":",
"self",
".",
"wf_state",
"[",
"'finished'",
"]",
"=",
"True",
"self",
".",
"wf_state",
"[",
"'finish_date'",
"]",
"=",
"datetime",
".",
"now",
"(",
")",
".",
"strftime",
"(",
"settings",
".",
"DATETIME_DEFAULT_FORMAT",
")",
"if",
"self",
".",
"current",
".",
"workflow_name",
"not",
"in",
"settings",
".",
"EPHEMERAL_WORKFLOWS",
"and",
"not",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
":",
"wfi",
"=",
"WFCache",
"(",
"self",
".",
"current",
")",
".",
"get_instance",
"(",
")",
"TaskInvitation",
".",
"objects",
".",
"filter",
"(",
"instance",
"=",
"wfi",
",",
"role",
"=",
"self",
".",
"current",
".",
"role",
",",
"wf_name",
"=",
"wfi",
".",
"wf",
".",
"name",
")",
".",
"delete",
"(",
")",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"Delete WFCache: %s %s\"",
"%",
"(",
"self",
".",
"current",
".",
"workflow_name",
",",
"self",
".",
"current",
".",
"token",
")",
")",
"self",
".",
"save_workflow_to_cache",
"(",
"self",
".",
"serialize_workflow",
"(",
")",
")"
] |
Calls the real save method if we pass the beggining of the wf
|
[
"Calls",
"the",
"real",
"save",
"method",
"if",
"we",
"pass",
"the",
"beggining",
"of",
"the",
"wf"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L221-L239
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.start_engine
|
def start_engine(self, **kwargs):
"""
Initializes the workflow with given request, response objects and diagram name.
Args:
session:
input:
workflow_name (str): Name of workflow diagram without ".bpmn" suffix.
File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
"""
self.current = WFCurrent(**kwargs)
self.wf_state = {'in_external': False, 'finished': False}
if not self.current.new_token:
self.wf_state = self.current.wf_cache.get(self.wf_state)
self.current.workflow_name = self.wf_state['name']
# if we have a pre-selected object to work with,
# inserting it as current.input['id'] and task_data['object_id']
if 'subject' in self.wf_state:
self.current.input['id'] = self.wf_state['subject']
self.current.task_data['object_id'] = self.wf_state['subject']
self.check_for_authentication()
self.check_for_permission()
self.workflow = self.load_or_create_workflow()
# if form data exists in input (user submitted)
# put form data in wf task_data
if 'form' in self.current.input:
form = self.current.input['form']
if 'form_name' in form:
self.current.task_data[form['form_name']] = form
# in wf diagram, if property is stated as init = True
# demanded initial values are assigned and put to cache
start_init_values = self.workflow_spec.wf_properties.get('init', 'False') == 'True'
if start_init_values:
WFInit = get_object_from_path(settings.WF_INITIAL_VALUES)()
WFInit.assign_wf_initial_values(self.current)
log_msg = ("\n\n::::::::::: ENGINE STARTED :::::::::::\n"
"\tWF: %s (Possible) TASK:%s\n"
"\tCMD:%s\n"
"\tSUBCMD:%s" % (
self.workflow.name,
self.workflow.get_tasks(Task.READY),
self.current.input.get('cmd'), self.current.input.get('subcmd')))
log.debug(log_msg)
sys._zops_wf_state_log = log_msg
self.current.workflow = self.workflow
|
python
|
def start_engine(self, **kwargs):
"""
Initializes the workflow with given request, response objects and diagram name.
Args:
session:
input:
workflow_name (str): Name of workflow diagram without ".bpmn" suffix.
File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
"""
self.current = WFCurrent(**kwargs)
self.wf_state = {'in_external': False, 'finished': False}
if not self.current.new_token:
self.wf_state = self.current.wf_cache.get(self.wf_state)
self.current.workflow_name = self.wf_state['name']
# if we have a pre-selected object to work with,
# inserting it as current.input['id'] and task_data['object_id']
if 'subject' in self.wf_state:
self.current.input['id'] = self.wf_state['subject']
self.current.task_data['object_id'] = self.wf_state['subject']
self.check_for_authentication()
self.check_for_permission()
self.workflow = self.load_or_create_workflow()
# if form data exists in input (user submitted)
# put form data in wf task_data
if 'form' in self.current.input:
form = self.current.input['form']
if 'form_name' in form:
self.current.task_data[form['form_name']] = form
# in wf diagram, if property is stated as init = True
# demanded initial values are assigned and put to cache
start_init_values = self.workflow_spec.wf_properties.get('init', 'False') == 'True'
if start_init_values:
WFInit = get_object_from_path(settings.WF_INITIAL_VALUES)()
WFInit.assign_wf_initial_values(self.current)
log_msg = ("\n\n::::::::::: ENGINE STARTED :::::::::::\n"
"\tWF: %s (Possible) TASK:%s\n"
"\tCMD:%s\n"
"\tSUBCMD:%s" % (
self.workflow.name,
self.workflow.get_tasks(Task.READY),
self.current.input.get('cmd'), self.current.input.get('subcmd')))
log.debug(log_msg)
sys._zops_wf_state_log = log_msg
self.current.workflow = self.workflow
|
[
"def",
"start_engine",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"self",
".",
"current",
"=",
"WFCurrent",
"(",
"*",
"*",
"kwargs",
")",
"self",
".",
"wf_state",
"=",
"{",
"'in_external'",
":",
"False",
",",
"'finished'",
":",
"False",
"}",
"if",
"not",
"self",
".",
"current",
".",
"new_token",
":",
"self",
".",
"wf_state",
"=",
"self",
".",
"current",
".",
"wf_cache",
".",
"get",
"(",
"self",
".",
"wf_state",
")",
"self",
".",
"current",
".",
"workflow_name",
"=",
"self",
".",
"wf_state",
"[",
"'name'",
"]",
"# if we have a pre-selected object to work with,",
"# inserting it as current.input['id'] and task_data['object_id']",
"if",
"'subject'",
"in",
"self",
".",
"wf_state",
":",
"self",
".",
"current",
".",
"input",
"[",
"'id'",
"]",
"=",
"self",
".",
"wf_state",
"[",
"'subject'",
"]",
"self",
".",
"current",
".",
"task_data",
"[",
"'object_id'",
"]",
"=",
"self",
".",
"wf_state",
"[",
"'subject'",
"]",
"self",
".",
"check_for_authentication",
"(",
")",
"self",
".",
"check_for_permission",
"(",
")",
"self",
".",
"workflow",
"=",
"self",
".",
"load_or_create_workflow",
"(",
")",
"# if form data exists in input (user submitted)",
"# put form data in wf task_data",
"if",
"'form'",
"in",
"self",
".",
"current",
".",
"input",
":",
"form",
"=",
"self",
".",
"current",
".",
"input",
"[",
"'form'",
"]",
"if",
"'form_name'",
"in",
"form",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"form",
"[",
"'form_name'",
"]",
"]",
"=",
"form",
"# in wf diagram, if property is stated as init = True",
"# demanded initial values are assigned and put to cache",
"start_init_values",
"=",
"self",
".",
"workflow_spec",
".",
"wf_properties",
".",
"get",
"(",
"'init'",
",",
"'False'",
")",
"==",
"'True'",
"if",
"start_init_values",
":",
"WFInit",
"=",
"get_object_from_path",
"(",
"settings",
".",
"WF_INITIAL_VALUES",
")",
"(",
")",
"WFInit",
".",
"assign_wf_initial_values",
"(",
"self",
".",
"current",
")",
"log_msg",
"=",
"(",
"\"\\n\\n::::::::::: ENGINE STARTED :::::::::::\\n\"",
"\"\\tWF: %s (Possible) TASK:%s\\n\"",
"\"\\tCMD:%s\\n\"",
"\"\\tSUBCMD:%s\"",
"%",
"(",
"self",
".",
"workflow",
".",
"name",
",",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"Task",
".",
"READY",
")",
",",
"self",
".",
"current",
".",
"input",
".",
"get",
"(",
"'cmd'",
")",
",",
"self",
".",
"current",
".",
"input",
".",
"get",
"(",
"'subcmd'",
")",
")",
")",
"log",
".",
"debug",
"(",
"log_msg",
")",
"sys",
".",
"_zops_wf_state_log",
"=",
"log_msg",
"self",
".",
"current",
".",
"workflow",
"=",
"self",
".",
"workflow"
] |
Initializes the workflow with given request, response objects and diagram name.
Args:
session:
input:
workflow_name (str): Name of workflow diagram without ".bpmn" suffix.
File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
|
[
"Initializes",
"the",
"workflow",
"with",
"given",
"request",
"response",
"objects",
"and",
"diagram",
"name",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L241-L289
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.generate_wf_state_log
|
def generate_wf_state_log(self):
"""
Logs the state of workflow and content of task_data.
"""
output = '\n- - - - - -\n'
output += "WORKFLOW: %s ( %s )" % (self.current.workflow_name.upper(),
self.current.workflow.name)
output += "\nTASK: %s ( %s )\n" % (self.current.task_name, self.current.task_type)
output += "DATA:"
for k, v in self.current.task_data.items():
if v:
output += "\n\t%s: %s" % (k, v)
output += "\nCURRENT:"
output += "\n\tACTIVITY: %s" % self.current.activity
output += "\n\tPOOL: %s" % self.current.pool
output += "\n\tIN EXTERNAL: %s" % self.wf_state['in_external']
output += "\n\tLANE: %s" % self.current.lane_name
output += "\n\tTOKEN: %s" % self.current.token
sys._zops_wf_state_log = output
return output
|
python
|
def generate_wf_state_log(self):
"""
Logs the state of workflow and content of task_data.
"""
output = '\n- - - - - -\n'
output += "WORKFLOW: %s ( %s )" % (self.current.workflow_name.upper(),
self.current.workflow.name)
output += "\nTASK: %s ( %s )\n" % (self.current.task_name, self.current.task_type)
output += "DATA:"
for k, v in self.current.task_data.items():
if v:
output += "\n\t%s: %s" % (k, v)
output += "\nCURRENT:"
output += "\n\tACTIVITY: %s" % self.current.activity
output += "\n\tPOOL: %s" % self.current.pool
output += "\n\tIN EXTERNAL: %s" % self.wf_state['in_external']
output += "\n\tLANE: %s" % self.current.lane_name
output += "\n\tTOKEN: %s" % self.current.token
sys._zops_wf_state_log = output
return output
|
[
"def",
"generate_wf_state_log",
"(",
"self",
")",
":",
"output",
"=",
"'\\n- - - - - -\\n'",
"output",
"+=",
"\"WORKFLOW: %s ( %s )\"",
"%",
"(",
"self",
".",
"current",
".",
"workflow_name",
".",
"upper",
"(",
")",
",",
"self",
".",
"current",
".",
"workflow",
".",
"name",
")",
"output",
"+=",
"\"\\nTASK: %s ( %s )\\n\"",
"%",
"(",
"self",
".",
"current",
".",
"task_name",
",",
"self",
".",
"current",
".",
"task_type",
")",
"output",
"+=",
"\"DATA:\"",
"for",
"k",
",",
"v",
"in",
"self",
".",
"current",
".",
"task_data",
".",
"items",
"(",
")",
":",
"if",
"v",
":",
"output",
"+=",
"\"\\n\\t%s: %s\"",
"%",
"(",
"k",
",",
"v",
")",
"output",
"+=",
"\"\\nCURRENT:\"",
"output",
"+=",
"\"\\n\\tACTIVITY: %s\"",
"%",
"self",
".",
"current",
".",
"activity",
"output",
"+=",
"\"\\n\\tPOOL: %s\"",
"%",
"self",
".",
"current",
".",
"pool",
"output",
"+=",
"\"\\n\\tIN EXTERNAL: %s\"",
"%",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
"output",
"+=",
"\"\\n\\tLANE: %s\"",
"%",
"self",
".",
"current",
".",
"lane_name",
"output",
"+=",
"\"\\n\\tTOKEN: %s\"",
"%",
"self",
".",
"current",
".",
"token",
"sys",
".",
"_zops_wf_state_log",
"=",
"output",
"return",
"output"
] |
Logs the state of workflow and content of task_data.
|
[
"Logs",
"the",
"state",
"of",
"workflow",
"and",
"content",
"of",
"task_data",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L291-L311
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.switch_from_external_to_main_wf
|
def switch_from_external_to_main_wf(self):
"""
Main workflow switcher.
This method recreates main workflow from `main wf` dict which
was set by external workflow swicther previously.
"""
# in external assigned as True in switch_to_external_wf.
# external_wf should finish EndEvent and it's name should be
# also EndEvent for switching again to main wf.
if self.wf_state['in_external'] and self.current.task_type == 'EndEvent' and \
self.current.task_name == 'EndEvent':
# main_wf information was copied in switch_to_external_wf and it takes this information.
main_wf = self.wf_state['main_wf']
# main_wf_name is assigned to current workflow name again.
self.current.workflow_name = main_wf['name']
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# WF knowledge is taken for main wf.
self.workflow_spec = self.get_worfklow_spec()
# WF instance is started again where leave off.
self.workflow = self.deserialize_workflow(main_wf['step'])
# Current WF is this WF instance.
self.current.workflow = self.workflow
# in_external is assigned as False
self.wf_state['in_external'] = False
# finished is assigned as False, because still in progress.
self.wf_state['finished'] = False
# pool info of main_wf is assigned.
self.wf_state['pool'] = main_wf['pool']
self.current.pool = self.wf_state['pool']
# With main_wf is executed.
self.run()
|
python
|
def switch_from_external_to_main_wf(self):
"""
Main workflow switcher.
This method recreates main workflow from `main wf` dict which
was set by external workflow swicther previously.
"""
# in external assigned as True in switch_to_external_wf.
# external_wf should finish EndEvent and it's name should be
# also EndEvent for switching again to main wf.
if self.wf_state['in_external'] and self.current.task_type == 'EndEvent' and \
self.current.task_name == 'EndEvent':
# main_wf information was copied in switch_to_external_wf and it takes this information.
main_wf = self.wf_state['main_wf']
# main_wf_name is assigned to current workflow name again.
self.current.workflow_name = main_wf['name']
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# WF knowledge is taken for main wf.
self.workflow_spec = self.get_worfklow_spec()
# WF instance is started again where leave off.
self.workflow = self.deserialize_workflow(main_wf['step'])
# Current WF is this WF instance.
self.current.workflow = self.workflow
# in_external is assigned as False
self.wf_state['in_external'] = False
# finished is assigned as False, because still in progress.
self.wf_state['finished'] = False
# pool info of main_wf is assigned.
self.wf_state['pool'] = main_wf['pool']
self.current.pool = self.wf_state['pool']
# With main_wf is executed.
self.run()
|
[
"def",
"switch_from_external_to_main_wf",
"(",
"self",
")",
":",
"# in external assigned as True in switch_to_external_wf.",
"# external_wf should finish EndEvent and it's name should be",
"# also EndEvent for switching again to main wf.",
"if",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
"and",
"self",
".",
"current",
".",
"task_type",
"==",
"'EndEvent'",
"and",
"self",
".",
"current",
".",
"task_name",
"==",
"'EndEvent'",
":",
"# main_wf information was copied in switch_to_external_wf and it takes this information.",
"main_wf",
"=",
"self",
".",
"wf_state",
"[",
"'main_wf'",
"]",
"# main_wf_name is assigned to current workflow name again.",
"self",
".",
"current",
".",
"workflow_name",
"=",
"main_wf",
"[",
"'name'",
"]",
"# For external WF, check permission and authentication. But after cleaning current task.",
"self",
".",
"_clear_current_task",
"(",
")",
"# check for auth and perm. current task cleared, do against new workflow_name",
"self",
".",
"check_for_authentication",
"(",
")",
"self",
".",
"check_for_permission",
"(",
")",
"# WF knowledge is taken for main wf.",
"self",
".",
"workflow_spec",
"=",
"self",
".",
"get_worfklow_spec",
"(",
")",
"# WF instance is started again where leave off.",
"self",
".",
"workflow",
"=",
"self",
".",
"deserialize_workflow",
"(",
"main_wf",
"[",
"'step'",
"]",
")",
"# Current WF is this WF instance.",
"self",
".",
"current",
".",
"workflow",
"=",
"self",
".",
"workflow",
"# in_external is assigned as False",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
"=",
"False",
"# finished is assigned as False, because still in progress.",
"self",
".",
"wf_state",
"[",
"'finished'",
"]",
"=",
"False",
"# pool info of main_wf is assigned.",
"self",
".",
"wf_state",
"[",
"'pool'",
"]",
"=",
"main_wf",
"[",
"'pool'",
"]",
"self",
".",
"current",
".",
"pool",
"=",
"self",
".",
"wf_state",
"[",
"'pool'",
"]",
"# With main_wf is executed.",
"self",
".",
"run",
"(",
")"
] |
Main workflow switcher.
This method recreates main workflow from `main wf` dict which
was set by external workflow swicther previously.
|
[
"Main",
"workflow",
"switcher",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L316-L365
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.switch_to_external_wf
|
def switch_to_external_wf(self):
"""
External workflow switcher.
This method copies main workflow information into
a temporary dict `main_wf` and makes external workflow
acting as main workflow.
"""
# External WF name should be stated at main wf diagram and type should be service task.
if (self.current.task_type == 'ServiceTask' and
self.current.task.task_spec.type == 'external'):
log.debug("Entering to EXTERNAL WF")
# Main wf information is copied to main_wf.
main_wf = self.wf_state.copy()
# workflow name from main wf diagram is assigned to current workflow name.
# workflow name must be either in task_data with key 'external_wf' or in main diagram's
# topic.
self.current.workflow_name = self.current.task_data.pop('external_wf', False) or self.\
current.task.task_spec.topic
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# wf knowledge is taken for external wf.
self.workflow_spec = self.get_worfklow_spec()
# New WF instance is created for external wf.
self.workflow = self.create_workflow()
# Current WF is this WF instance.
self.current.workflow = self.workflow
# main_wf: main wf information.
# in_external: it states external wf in progress.
# finished: it shows that main wf didn't finish still progress in external wf.
self.wf_state = {'main_wf': main_wf, 'in_external': True, 'finished': False}
|
python
|
def switch_to_external_wf(self):
"""
External workflow switcher.
This method copies main workflow information into
a temporary dict `main_wf` and makes external workflow
acting as main workflow.
"""
# External WF name should be stated at main wf diagram and type should be service task.
if (self.current.task_type == 'ServiceTask' and
self.current.task.task_spec.type == 'external'):
log.debug("Entering to EXTERNAL WF")
# Main wf information is copied to main_wf.
main_wf = self.wf_state.copy()
# workflow name from main wf diagram is assigned to current workflow name.
# workflow name must be either in task_data with key 'external_wf' or in main diagram's
# topic.
self.current.workflow_name = self.current.task_data.pop('external_wf', False) or self.\
current.task.task_spec.topic
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# wf knowledge is taken for external wf.
self.workflow_spec = self.get_worfklow_spec()
# New WF instance is created for external wf.
self.workflow = self.create_workflow()
# Current WF is this WF instance.
self.current.workflow = self.workflow
# main_wf: main wf information.
# in_external: it states external wf in progress.
# finished: it shows that main wf didn't finish still progress in external wf.
self.wf_state = {'main_wf': main_wf, 'in_external': True, 'finished': False}
|
[
"def",
"switch_to_external_wf",
"(",
"self",
")",
":",
"# External WF name should be stated at main wf diagram and type should be service task.",
"if",
"(",
"self",
".",
"current",
".",
"task_type",
"==",
"'ServiceTask'",
"and",
"self",
".",
"current",
".",
"task",
".",
"task_spec",
".",
"type",
"==",
"'external'",
")",
":",
"log",
".",
"debug",
"(",
"\"Entering to EXTERNAL WF\"",
")",
"# Main wf information is copied to main_wf.",
"main_wf",
"=",
"self",
".",
"wf_state",
".",
"copy",
"(",
")",
"# workflow name from main wf diagram is assigned to current workflow name.",
"# workflow name must be either in task_data with key 'external_wf' or in main diagram's",
"# topic.",
"self",
".",
"current",
".",
"workflow_name",
"=",
"self",
".",
"current",
".",
"task_data",
".",
"pop",
"(",
"'external_wf'",
",",
"False",
")",
"or",
"self",
".",
"current",
".",
"task",
".",
"task_spec",
".",
"topic",
"# For external WF, check permission and authentication. But after cleaning current task.",
"self",
".",
"_clear_current_task",
"(",
")",
"# check for auth and perm. current task cleared, do against new workflow_name",
"self",
".",
"check_for_authentication",
"(",
")",
"self",
".",
"check_for_permission",
"(",
")",
"# wf knowledge is taken for external wf.",
"self",
".",
"workflow_spec",
"=",
"self",
".",
"get_worfklow_spec",
"(",
")",
"# New WF instance is created for external wf.",
"self",
".",
"workflow",
"=",
"self",
".",
"create_workflow",
"(",
")",
"# Current WF is this WF instance.",
"self",
".",
"current",
".",
"workflow",
"=",
"self",
".",
"workflow",
"# main_wf: main wf information.",
"# in_external: it states external wf in progress.",
"# finished: it shows that main wf didn't finish still progress in external wf.",
"self",
".",
"wf_state",
"=",
"{",
"'main_wf'",
":",
"main_wf",
",",
"'in_external'",
":",
"True",
",",
"'finished'",
":",
"False",
"}"
] |
External workflow switcher.
This method copies main workflow information into
a temporary dict `main_wf` and makes external workflow
acting as main workflow.
|
[
"External",
"workflow",
"switcher",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L367-L408
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._clear_current_task
|
def _clear_current_task(self):
"""
Clear tasks related attributes, checks permissions
While switching WF to WF, authentication and permissions are checked for new WF.
"""
self.current.task_name = None
self.current.task_type = None
self.current.task = None
|
python
|
def _clear_current_task(self):
"""
Clear tasks related attributes, checks permissions
While switching WF to WF, authentication and permissions are checked for new WF.
"""
self.current.task_name = None
self.current.task_type = None
self.current.task = None
|
[
"def",
"_clear_current_task",
"(",
"self",
")",
":",
"self",
".",
"current",
".",
"task_name",
"=",
"None",
"self",
".",
"current",
".",
"task_type",
"=",
"None",
"self",
".",
"current",
".",
"task",
"=",
"None"
] |
Clear tasks related attributes, checks permissions
While switching WF to WF, authentication and permissions are checked for new WF.
|
[
"Clear",
"tasks",
"related",
"attributes",
"checks",
"permissions",
"While",
"switching",
"WF",
"to",
"WF",
"authentication",
"and",
"permissions",
"are",
"checked",
"for",
"new",
"WF",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L410-L418
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.run
|
def run(self):
"""
Main loop of the workflow engine
- Updates ::class:`~WFCurrent` object.
- Checks for Permissions.
- Activates all READY tasks.
- Runs referenced activities (method calls).
- Saves WF states.
- Stops if current task is a UserTask or EndTask.
- Deletes state object if we finish the WF.
"""
# FIXME: raise if first task after line change isn't a UserTask
# FIXME: raise if last task of a workflow is a UserTask
# actually this check should be done at parser
is_lane_changed = False
while self._should_we_run():
self.check_for_rerun_user_task()
task = None
for task in self.workflow.get_tasks(state=Task.READY):
self.current.old_lane = self.current.lane_name
self.current._update_task(task)
if self.catch_lane_change():
return
self.check_for_permission()
self.check_for_lane_permission()
self.log_wf_state()
self.switch_lang()
self.run_activity()
self.parse_workflow_messages()
self.workflow.complete_task_from_id(self.current.task.id)
self._save_or_delete_workflow()
self.switch_to_external_wf()
if task is None:
break
self.switch_from_external_to_main_wf()
self.current.output['token'] = self.current.token
# look for incoming ready task(s)
for task in self.workflow.get_tasks(state=Task.READY):
self.current._update_task(task)
self.catch_lane_change()
self.handle_wf_finalization()
|
python
|
def run(self):
"""
Main loop of the workflow engine
- Updates ::class:`~WFCurrent` object.
- Checks for Permissions.
- Activates all READY tasks.
- Runs referenced activities (method calls).
- Saves WF states.
- Stops if current task is a UserTask or EndTask.
- Deletes state object if we finish the WF.
"""
# FIXME: raise if first task after line change isn't a UserTask
# FIXME: raise if last task of a workflow is a UserTask
# actually this check should be done at parser
is_lane_changed = False
while self._should_we_run():
self.check_for_rerun_user_task()
task = None
for task in self.workflow.get_tasks(state=Task.READY):
self.current.old_lane = self.current.lane_name
self.current._update_task(task)
if self.catch_lane_change():
return
self.check_for_permission()
self.check_for_lane_permission()
self.log_wf_state()
self.switch_lang()
self.run_activity()
self.parse_workflow_messages()
self.workflow.complete_task_from_id(self.current.task.id)
self._save_or_delete_workflow()
self.switch_to_external_wf()
if task is None:
break
self.switch_from_external_to_main_wf()
self.current.output['token'] = self.current.token
# look for incoming ready task(s)
for task in self.workflow.get_tasks(state=Task.READY):
self.current._update_task(task)
self.catch_lane_change()
self.handle_wf_finalization()
|
[
"def",
"run",
"(",
"self",
")",
":",
"# FIXME: raise if first task after line change isn't a UserTask",
"# FIXME: raise if last task of a workflow is a UserTask",
"# actually this check should be done at parser",
"is_lane_changed",
"=",
"False",
"while",
"self",
".",
"_should_we_run",
"(",
")",
":",
"self",
".",
"check_for_rerun_user_task",
"(",
")",
"task",
"=",
"None",
"for",
"task",
"in",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"state",
"=",
"Task",
".",
"READY",
")",
":",
"self",
".",
"current",
".",
"old_lane",
"=",
"self",
".",
"current",
".",
"lane_name",
"self",
".",
"current",
".",
"_update_task",
"(",
"task",
")",
"if",
"self",
".",
"catch_lane_change",
"(",
")",
":",
"return",
"self",
".",
"check_for_permission",
"(",
")",
"self",
".",
"check_for_lane_permission",
"(",
")",
"self",
".",
"log_wf_state",
"(",
")",
"self",
".",
"switch_lang",
"(",
")",
"self",
".",
"run_activity",
"(",
")",
"self",
".",
"parse_workflow_messages",
"(",
")",
"self",
".",
"workflow",
".",
"complete_task_from_id",
"(",
"self",
".",
"current",
".",
"task",
".",
"id",
")",
"self",
".",
"_save_or_delete_workflow",
"(",
")",
"self",
".",
"switch_to_external_wf",
"(",
")",
"if",
"task",
"is",
"None",
":",
"break",
"self",
".",
"switch_from_external_to_main_wf",
"(",
")",
"self",
".",
"current",
".",
"output",
"[",
"'token'",
"]",
"=",
"self",
".",
"current",
".",
"token",
"# look for incoming ready task(s)",
"for",
"task",
"in",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"state",
"=",
"Task",
".",
"READY",
")",
":",
"self",
".",
"current",
".",
"_update_task",
"(",
"task",
")",
"self",
".",
"catch_lane_change",
"(",
")",
"self",
".",
"handle_wf_finalization",
"(",
")"
] |
Main loop of the workflow engine
- Updates ::class:`~WFCurrent` object.
- Checks for Permissions.
- Activates all READY tasks.
- Runs referenced activities (method calls).
- Saves WF states.
- Stops if current task is a UserTask or EndTask.
- Deletes state object if we finish the WF.
|
[
"Main",
"loop",
"of",
"the",
"workflow",
"engine"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L432-L477
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_rerun_user_task
|
def check_for_rerun_user_task(self):
"""
Checks that the user task needs to re-run.
If necessary, current task and pre task's states are changed and re-run.
If wf_meta not in data(there is no user interaction from pre-task) and last completed task
type is user task and current step is not EndEvent and there is no lane change,
this user task is rerun.
"""
data = self.current.input
if 'wf_meta' in data:
return
current_task = self.workflow.get_tasks(Task.READY)[0]
current_task_type = current_task.task_spec.__class__.__name__
pre_task = current_task.parent
pre_task_type = pre_task.task_spec.__class__.__name__
if pre_task_type != 'UserTask':
return
if current_task_type == 'EndEvent':
return
pre_lane = pre_task.task_spec.lane
current_lane = current_task.task_spec.lane
if pre_lane == current_lane:
pre_task._set_state(Task.READY)
current_task._set_state(Task.MAYBE)
|
python
|
def check_for_rerun_user_task(self):
"""
Checks that the user task needs to re-run.
If necessary, current task and pre task's states are changed and re-run.
If wf_meta not in data(there is no user interaction from pre-task) and last completed task
type is user task and current step is not EndEvent and there is no lane change,
this user task is rerun.
"""
data = self.current.input
if 'wf_meta' in data:
return
current_task = self.workflow.get_tasks(Task.READY)[0]
current_task_type = current_task.task_spec.__class__.__name__
pre_task = current_task.parent
pre_task_type = pre_task.task_spec.__class__.__name__
if pre_task_type != 'UserTask':
return
if current_task_type == 'EndEvent':
return
pre_lane = pre_task.task_spec.lane
current_lane = current_task.task_spec.lane
if pre_lane == current_lane:
pre_task._set_state(Task.READY)
current_task._set_state(Task.MAYBE)
|
[
"def",
"check_for_rerun_user_task",
"(",
"self",
")",
":",
"data",
"=",
"self",
".",
"current",
".",
"input",
"if",
"'wf_meta'",
"in",
"data",
":",
"return",
"current_task",
"=",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"Task",
".",
"READY",
")",
"[",
"0",
"]",
"current_task_type",
"=",
"current_task",
".",
"task_spec",
".",
"__class__",
".",
"__name__",
"pre_task",
"=",
"current_task",
".",
"parent",
"pre_task_type",
"=",
"pre_task",
".",
"task_spec",
".",
"__class__",
".",
"__name__",
"if",
"pre_task_type",
"!=",
"'UserTask'",
":",
"return",
"if",
"current_task_type",
"==",
"'EndEvent'",
":",
"return",
"pre_lane",
"=",
"pre_task",
".",
"task_spec",
".",
"lane",
"current_lane",
"=",
"current_task",
".",
"task_spec",
".",
"lane",
"if",
"pre_lane",
"==",
"current_lane",
":",
"pre_task",
".",
"_set_state",
"(",
"Task",
".",
"READY",
")",
"current_task",
".",
"_set_state",
"(",
"Task",
".",
"MAYBE",
")"
] |
Checks that the user task needs to re-run.
If necessary, current task and pre task's states are changed and re-run.
If wf_meta not in data(there is no user interaction from pre-task) and last completed task
type is user task and current step is not EndEvent and there is no lane change,
this user task is rerun.
|
[
"Checks",
"that",
"the",
"user",
"task",
"needs",
"to",
"re",
"-",
"run",
".",
"If",
"necessary",
"current",
"task",
"and",
"pre",
"task",
"s",
"states",
"are",
"changed",
"and",
"re",
"-",
"run",
".",
"If",
"wf_meta",
"not",
"in",
"data",
"(",
"there",
"is",
"no",
"user",
"interaction",
"from",
"pre",
"-",
"task",
")",
"and",
"last",
"completed",
"task",
"type",
"is",
"user",
"task",
"and",
"current",
"step",
"is",
"not",
"EndEvent",
"and",
"there",
"is",
"no",
"lane",
"change",
"this",
"user",
"task",
"is",
"rerun",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L479-L506
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.switch_lang
|
def switch_lang(self):
"""Switch to the language of the current user.
If the current language is already the specified one, nothing will be done.
"""
locale = self.current.locale
translation.InstalledLocale.install_language(locale['locale_language'])
translation.InstalledLocale.install_locale(locale['locale_datetime'], 'datetime')
translation.InstalledLocale.install_locale(locale['locale_number'], 'number')
|
python
|
def switch_lang(self):
"""Switch to the language of the current user.
If the current language is already the specified one, nothing will be done.
"""
locale = self.current.locale
translation.InstalledLocale.install_language(locale['locale_language'])
translation.InstalledLocale.install_locale(locale['locale_datetime'], 'datetime')
translation.InstalledLocale.install_locale(locale['locale_number'], 'number')
|
[
"def",
"switch_lang",
"(",
"self",
")",
":",
"locale",
"=",
"self",
".",
"current",
".",
"locale",
"translation",
".",
"InstalledLocale",
".",
"install_language",
"(",
"locale",
"[",
"'locale_language'",
"]",
")",
"translation",
".",
"InstalledLocale",
".",
"install_locale",
"(",
"locale",
"[",
"'locale_datetime'",
"]",
",",
"'datetime'",
")",
"translation",
".",
"InstalledLocale",
".",
"install_locale",
"(",
"locale",
"[",
"'locale_number'",
"]",
",",
"'number'",
")"
] |
Switch to the language of the current user.
If the current language is already the specified one, nothing will be done.
|
[
"Switch",
"to",
"the",
"language",
"of",
"the",
"current",
"user",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L508-L516
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.catch_lane_change
|
def catch_lane_change(self):
"""
trigger a lane_user_change signal if we switched to a new lane
and new lane's user is different from current one
"""
if self.current.lane_name:
if self.current.old_lane and self.current.lane_name != self.current.old_lane:
# if lane_name not found in pool or it's user different from the current(old) user
if (self.current.lane_id not in self.current.pool or
self.current.pool[self.current.lane_id] != self.current.user_id):
self.current.log.info("LANE CHANGE : %s >> %s" % (self.current.old_lane,
self.current.lane_name))
if self.current.lane_auto_sendoff:
self.current.sendoff_current_user()
self.current.flow_enabled = False
if self.current.lane_auto_invite:
self.current.invite_other_parties(self._get_possible_lane_owners())
return True
|
python
|
def catch_lane_change(self):
"""
trigger a lane_user_change signal if we switched to a new lane
and new lane's user is different from current one
"""
if self.current.lane_name:
if self.current.old_lane and self.current.lane_name != self.current.old_lane:
# if lane_name not found in pool or it's user different from the current(old) user
if (self.current.lane_id not in self.current.pool or
self.current.pool[self.current.lane_id] != self.current.user_id):
self.current.log.info("LANE CHANGE : %s >> %s" % (self.current.old_lane,
self.current.lane_name))
if self.current.lane_auto_sendoff:
self.current.sendoff_current_user()
self.current.flow_enabled = False
if self.current.lane_auto_invite:
self.current.invite_other_parties(self._get_possible_lane_owners())
return True
|
[
"def",
"catch_lane_change",
"(",
"self",
")",
":",
"if",
"self",
".",
"current",
".",
"lane_name",
":",
"if",
"self",
".",
"current",
".",
"old_lane",
"and",
"self",
".",
"current",
".",
"lane_name",
"!=",
"self",
".",
"current",
".",
"old_lane",
":",
"# if lane_name not found in pool or it's user different from the current(old) user",
"if",
"(",
"self",
".",
"current",
".",
"lane_id",
"not",
"in",
"self",
".",
"current",
".",
"pool",
"or",
"self",
".",
"current",
".",
"pool",
"[",
"self",
".",
"current",
".",
"lane_id",
"]",
"!=",
"self",
".",
"current",
".",
"user_id",
")",
":",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"LANE CHANGE : %s >> %s\"",
"%",
"(",
"self",
".",
"current",
".",
"old_lane",
",",
"self",
".",
"current",
".",
"lane_name",
")",
")",
"if",
"self",
".",
"current",
".",
"lane_auto_sendoff",
":",
"self",
".",
"current",
".",
"sendoff_current_user",
"(",
")",
"self",
".",
"current",
".",
"flow_enabled",
"=",
"False",
"if",
"self",
".",
"current",
".",
"lane_auto_invite",
":",
"self",
".",
"current",
".",
"invite_other_parties",
"(",
"self",
".",
"_get_possible_lane_owners",
"(",
")",
")",
"return",
"True"
] |
trigger a lane_user_change signal if we switched to a new lane
and new lane's user is different from current one
|
[
"trigger",
"a",
"lane_user_change",
"signal",
"if",
"we",
"switched",
"to",
"a",
"new",
"lane",
"and",
"new",
"lane",
"s",
"user",
"is",
"different",
"from",
"current",
"one"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L518-L535
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.parse_workflow_messages
|
def parse_workflow_messages(self):
"""
Transmits client message that defined in
a workflow task's inputOutput extension
.. code-block:: xml
<bpmn2:extensionElements>
<camunda:inputOutput>
<camunda:inputParameter name="client_message">
<camunda:map>
<camunda:entry key="title">Teşekkürler</camunda:entry>
<camunda:entry key="body">İşlem Başarılı</camunda:entry>
<camunda:entry key="type">info</camunda:entry>
</camunda:map>
</camunda:inputParameter>
</camunda:inputOutput>
</bpmn2:extensionElements>
"""
if 'client_message' in self.current.spec.data:
m = self.current.spec.data['client_message']
self.current.msg_box(title=m.get('title'),
msg=m.get('body'),
typ=m.get('type', 'info'))
|
python
|
def parse_workflow_messages(self):
"""
Transmits client message that defined in
a workflow task's inputOutput extension
.. code-block:: xml
<bpmn2:extensionElements>
<camunda:inputOutput>
<camunda:inputParameter name="client_message">
<camunda:map>
<camunda:entry key="title">Teşekkürler</camunda:entry>
<camunda:entry key="body">İşlem Başarılı</camunda:entry>
<camunda:entry key="type">info</camunda:entry>
</camunda:map>
</camunda:inputParameter>
</camunda:inputOutput>
</bpmn2:extensionElements>
"""
if 'client_message' in self.current.spec.data:
m = self.current.spec.data['client_message']
self.current.msg_box(title=m.get('title'),
msg=m.get('body'),
typ=m.get('type', 'info'))
|
[
"def",
"parse_workflow_messages",
"(",
"self",
")",
":",
"if",
"'client_message'",
"in",
"self",
".",
"current",
".",
"spec",
".",
"data",
":",
"m",
"=",
"self",
".",
"current",
".",
"spec",
".",
"data",
"[",
"'client_message'",
"]",
"self",
".",
"current",
".",
"msg_box",
"(",
"title",
"=",
"m",
".",
"get",
"(",
"'title'",
")",
",",
"msg",
"=",
"m",
".",
"get",
"(",
"'body'",
")",
",",
"typ",
"=",
"m",
".",
"get",
"(",
"'type'",
",",
"'info'",
")",
")"
] |
Transmits client message that defined in
a workflow task's inputOutput extension
.. code-block:: xml
<bpmn2:extensionElements>
<camunda:inputOutput>
<camunda:inputParameter name="client_message">
<camunda:map>
<camunda:entry key="title">Teşekkürler</camunda:entry>
<camunda:entry key="body">İşlem Başarılı</camunda:entry>
<camunda:entry key="type">info</camunda:entry>
</camunda:map>
</camunda:inputParameter>
</camunda:inputOutput>
</bpmn2:extensionElements>
|
[
"Transmits",
"client",
"message",
"that",
"defined",
"in",
"a",
"workflow",
"task",
"s",
"inputOutput",
"extension"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L538-L562
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.run_activity
|
def run_activity(self):
"""
runs the method that referenced from current task
"""
activity = self.current.activity
if activity:
if activity not in self.wf_activities:
self._load_activity(activity)
self.current.log.debug(
"Calling Activity %s from %s" % (activity, self.wf_activities[activity]))
self.wf_activities[self.current.activity](self.current)
|
python
|
def run_activity(self):
"""
runs the method that referenced from current task
"""
activity = self.current.activity
if activity:
if activity not in self.wf_activities:
self._load_activity(activity)
self.current.log.debug(
"Calling Activity %s from %s" % (activity, self.wf_activities[activity]))
self.wf_activities[self.current.activity](self.current)
|
[
"def",
"run_activity",
"(",
"self",
")",
":",
"activity",
"=",
"self",
".",
"current",
".",
"activity",
"if",
"activity",
":",
"if",
"activity",
"not",
"in",
"self",
".",
"wf_activities",
":",
"self",
".",
"_load_activity",
"(",
"activity",
")",
"self",
".",
"current",
".",
"log",
".",
"debug",
"(",
"\"Calling Activity %s from %s\"",
"%",
"(",
"activity",
",",
"self",
".",
"wf_activities",
"[",
"activity",
"]",
")",
")",
"self",
".",
"wf_activities",
"[",
"self",
".",
"current",
".",
"activity",
"]",
"(",
"self",
".",
"current",
")"
] |
runs the method that referenced from current task
|
[
"runs",
"the",
"method",
"that",
"referenced",
"from",
"current",
"task"
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L573-L583
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._import_object
|
def _import_object(self, path, look_for_cls_method):
"""
Imports the module that contains the referenced method.
Args:
path: python path of class/function
look_for_cls_method (bool): If True, treat the last part of path as class method.
Returns:
Tuple. (class object, class name, method to be called)
"""
last_nth = 2 if look_for_cls_method else 1
path = path.split('.')
module_path = '.'.join(path[:-last_nth])
class_name = path[-last_nth]
module = importlib.import_module(module_path)
if look_for_cls_method and path[-last_nth:][0] == path[-last_nth]:
class_method = path[-last_nth:][1]
else:
class_method = None
return getattr(module, class_name), class_name, class_method
|
python
|
def _import_object(self, path, look_for_cls_method):
"""
Imports the module that contains the referenced method.
Args:
path: python path of class/function
look_for_cls_method (bool): If True, treat the last part of path as class method.
Returns:
Tuple. (class object, class name, method to be called)
"""
last_nth = 2 if look_for_cls_method else 1
path = path.split('.')
module_path = '.'.join(path[:-last_nth])
class_name = path[-last_nth]
module = importlib.import_module(module_path)
if look_for_cls_method and path[-last_nth:][0] == path[-last_nth]:
class_method = path[-last_nth:][1]
else:
class_method = None
return getattr(module, class_name), class_name, class_method
|
[
"def",
"_import_object",
"(",
"self",
",",
"path",
",",
"look_for_cls_method",
")",
":",
"last_nth",
"=",
"2",
"if",
"look_for_cls_method",
"else",
"1",
"path",
"=",
"path",
".",
"split",
"(",
"'.'",
")",
"module_path",
"=",
"'.'",
".",
"join",
"(",
"path",
"[",
":",
"-",
"last_nth",
"]",
")",
"class_name",
"=",
"path",
"[",
"-",
"last_nth",
"]",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"module_path",
")",
"if",
"look_for_cls_method",
"and",
"path",
"[",
"-",
"last_nth",
":",
"]",
"[",
"0",
"]",
"==",
"path",
"[",
"-",
"last_nth",
"]",
":",
"class_method",
"=",
"path",
"[",
"-",
"last_nth",
":",
"]",
"[",
"1",
"]",
"else",
":",
"class_method",
"=",
"None",
"return",
"getattr",
"(",
"module",
",",
"class_name",
")",
",",
"class_name",
",",
"class_method"
] |
Imports the module that contains the referenced method.
Args:
path: python path of class/function
look_for_cls_method (bool): If True, treat the last part of path as class method.
Returns:
Tuple. (class object, class name, method to be called)
|
[
"Imports",
"the",
"module",
"that",
"contains",
"the",
"referenced",
"method",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L585-L606
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._load_activity
|
def _load_activity(self, activity):
"""
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
"""
fpths = []
full_path = ''
errors = []
paths = settings.ACTIVITY_MODULES_IMPORT_PATHS
number_of_paths = len(paths)
for index_no in range(number_of_paths):
full_path = "%s.%s" % (paths[index_no], activity)
for look4kls in (0, 1):
try:
self.current.log.info("try to load from %s[%s]" % (full_path, look4kls))
kls, cls_name, cls_method = self._import_object(full_path, look4kls)
if cls_method:
self.current.log.info("WILLCall %s(current).%s()" % (kls, cls_method))
self.wf_activities[activity] = lambda crnt: getattr(kls(crnt), cls_method)()
else:
self.wf_activities[activity] = kls
return
except (ImportError, AttributeError):
fpths.append(full_path)
errmsg = "{activity} not found under these paths:\n\n >>> {paths} \n\n" \
"Error Messages:\n {errors}"
errors.append("\n========================================================>\n"
"| PATH | %s"
"\n========================================================>\n\n"
"%s" % (full_path, traceback.format_exc()))
assert index_no != number_of_paths - 1, errmsg.format(activity=activity,
paths='\n >>> '.join(
set(fpths)),
errors='\n\n'.join(errors)
)
except:
self.current.log.exception("Cannot found the %s" % activity)
|
python
|
def _load_activity(self, activity):
"""
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
"""
fpths = []
full_path = ''
errors = []
paths = settings.ACTIVITY_MODULES_IMPORT_PATHS
number_of_paths = len(paths)
for index_no in range(number_of_paths):
full_path = "%s.%s" % (paths[index_no], activity)
for look4kls in (0, 1):
try:
self.current.log.info("try to load from %s[%s]" % (full_path, look4kls))
kls, cls_name, cls_method = self._import_object(full_path, look4kls)
if cls_method:
self.current.log.info("WILLCall %s(current).%s()" % (kls, cls_method))
self.wf_activities[activity] = lambda crnt: getattr(kls(crnt), cls_method)()
else:
self.wf_activities[activity] = kls
return
except (ImportError, AttributeError):
fpths.append(full_path)
errmsg = "{activity} not found under these paths:\n\n >>> {paths} \n\n" \
"Error Messages:\n {errors}"
errors.append("\n========================================================>\n"
"| PATH | %s"
"\n========================================================>\n\n"
"%s" % (full_path, traceback.format_exc()))
assert index_no != number_of_paths - 1, errmsg.format(activity=activity,
paths='\n >>> '.join(
set(fpths)),
errors='\n\n'.join(errors)
)
except:
self.current.log.exception("Cannot found the %s" % activity)
|
[
"def",
"_load_activity",
"(",
"self",
",",
"activity",
")",
":",
"fpths",
"=",
"[",
"]",
"full_path",
"=",
"''",
"errors",
"=",
"[",
"]",
"paths",
"=",
"settings",
".",
"ACTIVITY_MODULES_IMPORT_PATHS",
"number_of_paths",
"=",
"len",
"(",
"paths",
")",
"for",
"index_no",
"in",
"range",
"(",
"number_of_paths",
")",
":",
"full_path",
"=",
"\"%s.%s\"",
"%",
"(",
"paths",
"[",
"index_no",
"]",
",",
"activity",
")",
"for",
"look4kls",
"in",
"(",
"0",
",",
"1",
")",
":",
"try",
":",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"try to load from %s[%s]\"",
"%",
"(",
"full_path",
",",
"look4kls",
")",
")",
"kls",
",",
"cls_name",
",",
"cls_method",
"=",
"self",
".",
"_import_object",
"(",
"full_path",
",",
"look4kls",
")",
"if",
"cls_method",
":",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"WILLCall %s(current).%s()\"",
"%",
"(",
"kls",
",",
"cls_method",
")",
")",
"self",
".",
"wf_activities",
"[",
"activity",
"]",
"=",
"lambda",
"crnt",
":",
"getattr",
"(",
"kls",
"(",
"crnt",
")",
",",
"cls_method",
")",
"(",
")",
"else",
":",
"self",
".",
"wf_activities",
"[",
"activity",
"]",
"=",
"kls",
"return",
"except",
"(",
"ImportError",
",",
"AttributeError",
")",
":",
"fpths",
".",
"append",
"(",
"full_path",
")",
"errmsg",
"=",
"\"{activity} not found under these paths:\\n\\n >>> {paths} \\n\\n\"",
"\"Error Messages:\\n {errors}\"",
"errors",
".",
"append",
"(",
"\"\\n========================================================>\\n\"",
"\"| PATH | %s\"",
"\"\\n========================================================>\\n\\n\"",
"\"%s\"",
"%",
"(",
"full_path",
",",
"traceback",
".",
"format_exc",
"(",
")",
")",
")",
"assert",
"index_no",
"!=",
"number_of_paths",
"-",
"1",
",",
"errmsg",
".",
"format",
"(",
"activity",
"=",
"activity",
",",
"paths",
"=",
"'\\n >>> '",
".",
"join",
"(",
"set",
"(",
"fpths",
")",
")",
",",
"errors",
"=",
"'\\n\\n'",
".",
"join",
"(",
"errors",
")",
")",
"except",
":",
"self",
".",
"current",
".",
"log",
".",
"exception",
"(",
"\"Cannot found the %s\"",
"%",
"activity",
")"
] |
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
|
[
"Iterates",
"trough",
"the",
"all",
"enabled",
"~zengine",
".",
"settings",
".",
"ACTIVITY_MODULES_IMPORT_PATHS",
"to",
"find",
"the",
"given",
"path",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L608-L643
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_authentication
|
def check_for_authentication(self):
"""
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list.
Raises:
HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
"""
auth_required = self.current.workflow_name not in settings.ANONYMOUS_WORKFLOWS
if auth_required and not self.current.is_auth:
self.current.log.debug("LOGIN REQUIRED:::: %s" % self.current.workflow_name)
raise HTTPError(401, "Login required for %s" % self.current.workflow_name)
|
python
|
def check_for_authentication(self):
"""
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list.
Raises:
HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
"""
auth_required = self.current.workflow_name not in settings.ANONYMOUS_WORKFLOWS
if auth_required and not self.current.is_auth:
self.current.log.debug("LOGIN REQUIRED:::: %s" % self.current.workflow_name)
raise HTTPError(401, "Login required for %s" % self.current.workflow_name)
|
[
"def",
"check_for_authentication",
"(",
"self",
")",
":",
"auth_required",
"=",
"self",
".",
"current",
".",
"workflow_name",
"not",
"in",
"settings",
".",
"ANONYMOUS_WORKFLOWS",
"if",
"auth_required",
"and",
"not",
"self",
".",
"current",
".",
"is_auth",
":",
"self",
".",
"current",
".",
"log",
".",
"debug",
"(",
"\"LOGIN REQUIRED:::: %s\"",
"%",
"self",
".",
"current",
".",
"workflow_name",
")",
"raise",
"HTTPError",
"(",
"401",
",",
"\"Login required for %s\"",
"%",
"self",
".",
"current",
".",
"workflow_name",
")"
] |
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list.
Raises:
HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
|
[
"Checks",
"current",
"workflow",
"against",
":",
"py",
":",
"data",
":",
"~zengine",
".",
"settings",
".",
"ANONYMOUS_WORKFLOWS",
"list",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L645-L655
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_lane_permission
|
def check_for_lane_permission(self):
"""
One or more permissions can be associated with a lane
of a workflow. In a similar way, a lane can be
restricted with relation to other lanes of the workflow.
This method called on lane changes and checks user has
required permissions and relations.
Raises:
HTTPForbidden: if the current user hasn't got the
required permissions and proper relations
"""
# TODO: Cache lane_data in app memory
if self.current.lane_permission:
log.debug("HAS LANE PERM: %s" % self.current.lane_permission)
perm = self.current.lane_permission
if not self.current.has_permission(perm):
raise HTTPError(403, "You don't have required lane permission: %s" % perm)
if self.current.lane_relations:
context = self.get_pool_context()
log.debug("HAS LANE RELS: %s" % self.current.lane_relations)
try:
cond_result = eval(self.current.lane_relations, context)
except:
log.exception("CONDITION EVAL ERROR : %s || %s" % (
self.current.lane_relations, context))
raise
if not cond_result:
log.debug("LANE RELATION ERR: %s %s" % (self.current.lane_relations, context))
raise HTTPError(403, "You aren't qualified for this lane: %s" %
self.current.lane_relations)
|
python
|
def check_for_lane_permission(self):
"""
One or more permissions can be associated with a lane
of a workflow. In a similar way, a lane can be
restricted with relation to other lanes of the workflow.
This method called on lane changes and checks user has
required permissions and relations.
Raises:
HTTPForbidden: if the current user hasn't got the
required permissions and proper relations
"""
# TODO: Cache lane_data in app memory
if self.current.lane_permission:
log.debug("HAS LANE PERM: %s" % self.current.lane_permission)
perm = self.current.lane_permission
if not self.current.has_permission(perm):
raise HTTPError(403, "You don't have required lane permission: %s" % perm)
if self.current.lane_relations:
context = self.get_pool_context()
log.debug("HAS LANE RELS: %s" % self.current.lane_relations)
try:
cond_result = eval(self.current.lane_relations, context)
except:
log.exception("CONDITION EVAL ERROR : %s || %s" % (
self.current.lane_relations, context))
raise
if not cond_result:
log.debug("LANE RELATION ERR: %s %s" % (self.current.lane_relations, context))
raise HTTPError(403, "You aren't qualified for this lane: %s" %
self.current.lane_relations)
|
[
"def",
"check_for_lane_permission",
"(",
"self",
")",
":",
"# TODO: Cache lane_data in app memory",
"if",
"self",
".",
"current",
".",
"lane_permission",
":",
"log",
".",
"debug",
"(",
"\"HAS LANE PERM: %s\"",
"%",
"self",
".",
"current",
".",
"lane_permission",
")",
"perm",
"=",
"self",
".",
"current",
".",
"lane_permission",
"if",
"not",
"self",
".",
"current",
".",
"has_permission",
"(",
"perm",
")",
":",
"raise",
"HTTPError",
"(",
"403",
",",
"\"You don't have required lane permission: %s\"",
"%",
"perm",
")",
"if",
"self",
".",
"current",
".",
"lane_relations",
":",
"context",
"=",
"self",
".",
"get_pool_context",
"(",
")",
"log",
".",
"debug",
"(",
"\"HAS LANE RELS: %s\"",
"%",
"self",
".",
"current",
".",
"lane_relations",
")",
"try",
":",
"cond_result",
"=",
"eval",
"(",
"self",
".",
"current",
".",
"lane_relations",
",",
"context",
")",
"except",
":",
"log",
".",
"exception",
"(",
"\"CONDITION EVAL ERROR : %s || %s\"",
"%",
"(",
"self",
".",
"current",
".",
"lane_relations",
",",
"context",
")",
")",
"raise",
"if",
"not",
"cond_result",
":",
"log",
".",
"debug",
"(",
"\"LANE RELATION ERR: %s %s\"",
"%",
"(",
"self",
".",
"current",
".",
"lane_relations",
",",
"context",
")",
")",
"raise",
"HTTPError",
"(",
"403",
",",
"\"You aren't qualified for this lane: %s\"",
"%",
"self",
".",
"current",
".",
"lane_relations",
")"
] |
One or more permissions can be associated with a lane
of a workflow. In a similar way, a lane can be
restricted with relation to other lanes of the workflow.
This method called on lane changes and checks user has
required permissions and relations.
Raises:
HTTPForbidden: if the current user hasn't got the
required permissions and proper relations
|
[
"One",
"or",
"more",
"permissions",
"can",
"be",
"associated",
"with",
"a",
"lane",
"of",
"a",
"workflow",
".",
"In",
"a",
"similar",
"way",
"a",
"lane",
"can",
"be",
"restricted",
"with",
"relation",
"to",
"other",
"lanes",
"of",
"the",
"workflow",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L657-L690
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_permission
|
def check_for_permission(self):
# TODO: Works but not beautiful, needs review!
"""
Checks if current user (or role) has the required permission
for current workflow step.
Raises:
HTTPError: if user doesn't have required permissions.
"""
if self.current.task:
lane = self.current.lane_id
permission = "%s.%s.%s" % (self.current.workflow_name, lane, self.current.task_name)
else:
permission = self.current.workflow_name
log.debug("CHECK PERM: %s" % permission)
if (self.current.task_type not in PERM_REQ_TASK_TYPES or
permission.startswith(tuple(settings.ANONYMOUS_WORKFLOWS)) or
(self.current.is_auth and permission.startswith(tuple(settings.COMMON_WORKFLOWS)))):
return
# FIXME:needs hardening
log.debug("REQUIRE PERM: %s" % permission)
if not self.current.has_permission(permission):
raise HTTPError(403, "You don't have required permission: %s" % permission)
|
python
|
def check_for_permission(self):
# TODO: Works but not beautiful, needs review!
"""
Checks if current user (or role) has the required permission
for current workflow step.
Raises:
HTTPError: if user doesn't have required permissions.
"""
if self.current.task:
lane = self.current.lane_id
permission = "%s.%s.%s" % (self.current.workflow_name, lane, self.current.task_name)
else:
permission = self.current.workflow_name
log.debug("CHECK PERM: %s" % permission)
if (self.current.task_type not in PERM_REQ_TASK_TYPES or
permission.startswith(tuple(settings.ANONYMOUS_WORKFLOWS)) or
(self.current.is_auth and permission.startswith(tuple(settings.COMMON_WORKFLOWS)))):
return
# FIXME:needs hardening
log.debug("REQUIRE PERM: %s" % permission)
if not self.current.has_permission(permission):
raise HTTPError(403, "You don't have required permission: %s" % permission)
|
[
"def",
"check_for_permission",
"(",
"self",
")",
":",
"# TODO: Works but not beautiful, needs review!",
"if",
"self",
".",
"current",
".",
"task",
":",
"lane",
"=",
"self",
".",
"current",
".",
"lane_id",
"permission",
"=",
"\"%s.%s.%s\"",
"%",
"(",
"self",
".",
"current",
".",
"workflow_name",
",",
"lane",
",",
"self",
".",
"current",
".",
"task_name",
")",
"else",
":",
"permission",
"=",
"self",
".",
"current",
".",
"workflow_name",
"log",
".",
"debug",
"(",
"\"CHECK PERM: %s\"",
"%",
"permission",
")",
"if",
"(",
"self",
".",
"current",
".",
"task_type",
"not",
"in",
"PERM_REQ_TASK_TYPES",
"or",
"permission",
".",
"startswith",
"(",
"tuple",
"(",
"settings",
".",
"ANONYMOUS_WORKFLOWS",
")",
")",
"or",
"(",
"self",
".",
"current",
".",
"is_auth",
"and",
"permission",
".",
"startswith",
"(",
"tuple",
"(",
"settings",
".",
"COMMON_WORKFLOWS",
")",
")",
")",
")",
":",
"return",
"# FIXME:needs hardening",
"log",
".",
"debug",
"(",
"\"REQUIRE PERM: %s\"",
"%",
"permission",
")",
"if",
"not",
"self",
".",
"current",
".",
"has_permission",
"(",
"permission",
")",
":",
"raise",
"HTTPError",
"(",
"403",
",",
"\"You don't have required permission: %s\"",
"%",
"permission",
")"
] |
Checks if current user (or role) has the required permission
for current workflow step.
Raises:
HTTPError: if user doesn't have required permissions.
|
[
"Checks",
"if",
"current",
"user",
"(",
"or",
"role",
")",
"has",
"the",
"required",
"permission",
"for",
"current",
"workflow",
"step",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L692-L716
|
train
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.handle_wf_finalization
|
def handle_wf_finalization(self):
"""
Removes the ``token`` key from ``current.output`` if WF is over.
"""
if ((not self.current.flow_enabled or (
self.current.task_type.startswith('End') and not self.are_we_in_subprocess())) and
'token' in self.current.output):
del self.current.output['token']
|
python
|
def handle_wf_finalization(self):
"""
Removes the ``token`` key from ``current.output`` if WF is over.
"""
if ((not self.current.flow_enabled or (
self.current.task_type.startswith('End') and not self.are_we_in_subprocess())) and
'token' in self.current.output):
del self.current.output['token']
|
[
"def",
"handle_wf_finalization",
"(",
"self",
")",
":",
"if",
"(",
"(",
"not",
"self",
".",
"current",
".",
"flow_enabled",
"or",
"(",
"self",
".",
"current",
".",
"task_type",
".",
"startswith",
"(",
"'End'",
")",
"and",
"not",
"self",
".",
"are_we_in_subprocess",
"(",
")",
")",
")",
"and",
"'token'",
"in",
"self",
".",
"current",
".",
"output",
")",
":",
"del",
"self",
".",
"current",
".",
"output",
"[",
"'token'",
"]"
] |
Removes the ``token`` key from ``current.output`` if WF is over.
|
[
"Removes",
"the",
"token",
"key",
"from",
"current",
".",
"output",
"if",
"WF",
"is",
"over",
"."
] |
b5bc32d3b37bca799f8985be916f04528ac79e4a
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L718-L725
|
train
|
cimm-kzn/CGRtools
|
CGRtools/utils/rdkit.py
|
from_rdkit_molecule
|
def from_rdkit_molecule(data):
"""
RDKit molecule object to MoleculeContainer converter
"""
m = MoleculeContainer()
atoms, mapping = [], []
for a in data.GetAtoms():
atom = {'element': a.GetSymbol(), 'charge': a.GetFormalCharge()}
atoms.append(atom)
mapping.append(a.GetAtomMapNum())
isotope = a.GetIsotope()
if isotope:
atom['isotope'] = isotope
radical = a.GetNumRadicalElectrons()
if radical:
atom['multiplicity'] = radical + 1
conformers = data.GetConformers()
if conformers:
for atom, (x, y, z) in zip(atoms, conformers[0].GetPositions()):
atom['x'] = x
atom['y'] = y
atom['z'] = z
for atom, mapping in zip(atoms, mapping):
a = m.add_atom(atom)
if mapping:
m.atom(a)._parsed_mapping = mapping
for bond in data.GetBonds():
m.add_bond(bond.GetBeginAtomIdx() + 1, bond.GetEndAtomIdx() + 1, _rdkit_bond_map[bond.GetBondType()])
return m
|
python
|
def from_rdkit_molecule(data):
"""
RDKit molecule object to MoleculeContainer converter
"""
m = MoleculeContainer()
atoms, mapping = [], []
for a in data.GetAtoms():
atom = {'element': a.GetSymbol(), 'charge': a.GetFormalCharge()}
atoms.append(atom)
mapping.append(a.GetAtomMapNum())
isotope = a.GetIsotope()
if isotope:
atom['isotope'] = isotope
radical = a.GetNumRadicalElectrons()
if radical:
atom['multiplicity'] = radical + 1
conformers = data.GetConformers()
if conformers:
for atom, (x, y, z) in zip(atoms, conformers[0].GetPositions()):
atom['x'] = x
atom['y'] = y
atom['z'] = z
for atom, mapping in zip(atoms, mapping):
a = m.add_atom(atom)
if mapping:
m.atom(a)._parsed_mapping = mapping
for bond in data.GetBonds():
m.add_bond(bond.GetBeginAtomIdx() + 1, bond.GetEndAtomIdx() + 1, _rdkit_bond_map[bond.GetBondType()])
return m
|
[
"def",
"from_rdkit_molecule",
"(",
"data",
")",
":",
"m",
"=",
"MoleculeContainer",
"(",
")",
"atoms",
",",
"mapping",
"=",
"[",
"]",
",",
"[",
"]",
"for",
"a",
"in",
"data",
".",
"GetAtoms",
"(",
")",
":",
"atom",
"=",
"{",
"'element'",
":",
"a",
".",
"GetSymbol",
"(",
")",
",",
"'charge'",
":",
"a",
".",
"GetFormalCharge",
"(",
")",
"}",
"atoms",
".",
"append",
"(",
"atom",
")",
"mapping",
".",
"append",
"(",
"a",
".",
"GetAtomMapNum",
"(",
")",
")",
"isotope",
"=",
"a",
".",
"GetIsotope",
"(",
")",
"if",
"isotope",
":",
"atom",
"[",
"'isotope'",
"]",
"=",
"isotope",
"radical",
"=",
"a",
".",
"GetNumRadicalElectrons",
"(",
")",
"if",
"radical",
":",
"atom",
"[",
"'multiplicity'",
"]",
"=",
"radical",
"+",
"1",
"conformers",
"=",
"data",
".",
"GetConformers",
"(",
")",
"if",
"conformers",
":",
"for",
"atom",
",",
"(",
"x",
",",
"y",
",",
"z",
")",
"in",
"zip",
"(",
"atoms",
",",
"conformers",
"[",
"0",
"]",
".",
"GetPositions",
"(",
")",
")",
":",
"atom",
"[",
"'x'",
"]",
"=",
"x",
"atom",
"[",
"'y'",
"]",
"=",
"y",
"atom",
"[",
"'z'",
"]",
"=",
"z",
"for",
"atom",
",",
"mapping",
"in",
"zip",
"(",
"atoms",
",",
"mapping",
")",
":",
"a",
"=",
"m",
".",
"add_atom",
"(",
"atom",
")",
"if",
"mapping",
":",
"m",
".",
"atom",
"(",
"a",
")",
".",
"_parsed_mapping",
"=",
"mapping",
"for",
"bond",
"in",
"data",
".",
"GetBonds",
"(",
")",
":",
"m",
".",
"add_bond",
"(",
"bond",
".",
"GetBeginAtomIdx",
"(",
")",
"+",
"1",
",",
"bond",
".",
"GetEndAtomIdx",
"(",
")",
"+",
"1",
",",
"_rdkit_bond_map",
"[",
"bond",
".",
"GetBondType",
"(",
")",
"]",
")",
"return",
"m"
] |
RDKit molecule object to MoleculeContainer converter
|
[
"RDKit",
"molecule",
"object",
"to",
"MoleculeContainer",
"converter"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/utils/rdkit.py#L23-L56
|
train
|
cimm-kzn/CGRtools
|
CGRtools/utils/rdkit.py
|
to_rdkit_molecule
|
def to_rdkit_molecule(data):
"""
MoleculeContainer to RDKit molecule object converter
"""
mol = RWMol()
conf = Conformer()
mapping = {}
is_3d = False
for n, a in data.atoms():
ra = Atom(a.number)
ra.SetAtomMapNum(n)
if a.charge:
ra.SetFormalCharge(a.charge)
if a.isotope != a.common_isotope:
ra.SetIsotope(a.isotope)
if a.radical:
ra.SetNumRadicalElectrons(a.radical)
mapping[n] = m = mol.AddAtom(ra)
conf.SetAtomPosition(m, (a.x, a.y, a.z))
if a.z:
is_3d = True
if not is_3d:
conf.Set3D(False)
for n, m, b in data.bonds():
mol.AddBond(mapping[n], mapping[m], _bond_map[b.order])
mol.AddConformer(conf)
SanitizeMol(mol)
return mol
|
python
|
def to_rdkit_molecule(data):
"""
MoleculeContainer to RDKit molecule object converter
"""
mol = RWMol()
conf = Conformer()
mapping = {}
is_3d = False
for n, a in data.atoms():
ra = Atom(a.number)
ra.SetAtomMapNum(n)
if a.charge:
ra.SetFormalCharge(a.charge)
if a.isotope != a.common_isotope:
ra.SetIsotope(a.isotope)
if a.radical:
ra.SetNumRadicalElectrons(a.radical)
mapping[n] = m = mol.AddAtom(ra)
conf.SetAtomPosition(m, (a.x, a.y, a.z))
if a.z:
is_3d = True
if not is_3d:
conf.Set3D(False)
for n, m, b in data.bonds():
mol.AddBond(mapping[n], mapping[m], _bond_map[b.order])
mol.AddConformer(conf)
SanitizeMol(mol)
return mol
|
[
"def",
"to_rdkit_molecule",
"(",
"data",
")",
":",
"mol",
"=",
"RWMol",
"(",
")",
"conf",
"=",
"Conformer",
"(",
")",
"mapping",
"=",
"{",
"}",
"is_3d",
"=",
"False",
"for",
"n",
",",
"a",
"in",
"data",
".",
"atoms",
"(",
")",
":",
"ra",
"=",
"Atom",
"(",
"a",
".",
"number",
")",
"ra",
".",
"SetAtomMapNum",
"(",
"n",
")",
"if",
"a",
".",
"charge",
":",
"ra",
".",
"SetFormalCharge",
"(",
"a",
".",
"charge",
")",
"if",
"a",
".",
"isotope",
"!=",
"a",
".",
"common_isotope",
":",
"ra",
".",
"SetIsotope",
"(",
"a",
".",
"isotope",
")",
"if",
"a",
".",
"radical",
":",
"ra",
".",
"SetNumRadicalElectrons",
"(",
"a",
".",
"radical",
")",
"mapping",
"[",
"n",
"]",
"=",
"m",
"=",
"mol",
".",
"AddAtom",
"(",
"ra",
")",
"conf",
".",
"SetAtomPosition",
"(",
"m",
",",
"(",
"a",
".",
"x",
",",
"a",
".",
"y",
",",
"a",
".",
"z",
")",
")",
"if",
"a",
".",
"z",
":",
"is_3d",
"=",
"True",
"if",
"not",
"is_3d",
":",
"conf",
".",
"Set3D",
"(",
"False",
")",
"for",
"n",
",",
"m",
",",
"b",
"in",
"data",
".",
"bonds",
"(",
")",
":",
"mol",
".",
"AddBond",
"(",
"mapping",
"[",
"n",
"]",
",",
"mapping",
"[",
"m",
"]",
",",
"_bond_map",
"[",
"b",
".",
"order",
"]",
")",
"mol",
".",
"AddConformer",
"(",
"conf",
")",
"SanitizeMol",
"(",
"mol",
")",
"return",
"mol"
] |
MoleculeContainer to RDKit molecule object converter
|
[
"MoleculeContainer",
"to",
"RDKit",
"molecule",
"object",
"converter"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/utils/rdkit.py#L59-L88
|
train
|
cimm-kzn/CGRtools
|
CGRtools/algorithms/strings.py
|
StringCommon.__dfs
|
def __dfs(self, start, weights, depth_limit):
"""
modified NX dfs
"""
adj = self._adj
stack = [(start, depth_limit, iter(sorted(adj[start], key=weights)))]
visited = {start}
disconnected = defaultdict(list)
edges = defaultdict(list)
while stack:
parent, depth_now, children = stack[-1]
try:
child = next(children)
except StopIteration:
stack.pop()
else:
if child not in visited:
edges[parent].append(child)
visited.add(child)
if depth_now > 1:
front = adj[child].keys() - {parent}
if front:
stack.append((child, depth_now - 1, iter(sorted(front, key=weights))))
elif child not in disconnected:
disconnected[parent].append(child)
return visited, edges, disconnected
|
python
|
def __dfs(self, start, weights, depth_limit):
"""
modified NX dfs
"""
adj = self._adj
stack = [(start, depth_limit, iter(sorted(adj[start], key=weights)))]
visited = {start}
disconnected = defaultdict(list)
edges = defaultdict(list)
while stack:
parent, depth_now, children = stack[-1]
try:
child = next(children)
except StopIteration:
stack.pop()
else:
if child not in visited:
edges[parent].append(child)
visited.add(child)
if depth_now > 1:
front = adj[child].keys() - {parent}
if front:
stack.append((child, depth_now - 1, iter(sorted(front, key=weights))))
elif child not in disconnected:
disconnected[parent].append(child)
return visited, edges, disconnected
|
[
"def",
"__dfs",
"(",
"self",
",",
"start",
",",
"weights",
",",
"depth_limit",
")",
":",
"adj",
"=",
"self",
".",
"_adj",
"stack",
"=",
"[",
"(",
"start",
",",
"depth_limit",
",",
"iter",
"(",
"sorted",
"(",
"adj",
"[",
"start",
"]",
",",
"key",
"=",
"weights",
")",
")",
")",
"]",
"visited",
"=",
"{",
"start",
"}",
"disconnected",
"=",
"defaultdict",
"(",
"list",
")",
"edges",
"=",
"defaultdict",
"(",
"list",
")",
"while",
"stack",
":",
"parent",
",",
"depth_now",
",",
"children",
"=",
"stack",
"[",
"-",
"1",
"]",
"try",
":",
"child",
"=",
"next",
"(",
"children",
")",
"except",
"StopIteration",
":",
"stack",
".",
"pop",
"(",
")",
"else",
":",
"if",
"child",
"not",
"in",
"visited",
":",
"edges",
"[",
"parent",
"]",
".",
"append",
"(",
"child",
")",
"visited",
".",
"add",
"(",
"child",
")",
"if",
"depth_now",
">",
"1",
":",
"front",
"=",
"adj",
"[",
"child",
"]",
".",
"keys",
"(",
")",
"-",
"{",
"parent",
"}",
"if",
"front",
":",
"stack",
".",
"append",
"(",
"(",
"child",
",",
"depth_now",
"-",
"1",
",",
"iter",
"(",
"sorted",
"(",
"front",
",",
"key",
"=",
"weights",
")",
")",
")",
")",
"elif",
"child",
"not",
"in",
"disconnected",
":",
"disconnected",
"[",
"parent",
"]",
".",
"append",
"(",
"child",
")",
"return",
"visited",
",",
"edges",
",",
"disconnected"
] |
modified NX dfs
|
[
"modified",
"NX",
"dfs"
] |
15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/algorithms/strings.py#L130-L158
|
train
|
camptocamp/marabunta
|
marabunta/config.py
|
get_args_parser
|
def get_args_parser():
"""Return a parser for command line options."""
parser = argparse.ArgumentParser(
description='Marabunta: Migrating ants for Odoo')
parser.add_argument('--migration-file', '-f',
action=EnvDefault,
envvar='MARABUNTA_MIGRATION_FILE',
required=True,
help='The yaml file containing the migration steps')
parser.add_argument('--database', '-d',
action=EnvDefault,
envvar='MARABUNTA_DATABASE',
required=True,
help="Odoo's database")
parser.add_argument('--db-user', '-u',
action=EnvDefault,
envvar='MARABUNTA_DB_USER',
required=True,
help="Odoo's database user")
parser.add_argument('--db-password', '-w',
action=EnvDefault,
envvar='MARABUNTA_DB_PASSWORD',
required=True,
help="Odoo's database password")
parser.add_argument('--db-port', '-p',
default=os.environ.get('MARABUNTA_DB_PORT', 5432),
help="Odoo's database port")
parser.add_argument('--db-host', '-H',
default=os.environ.get('MARABUNTA_DB_HOST',
'localhost'),
help="Odoo's database host")
parser.add_argument('--mode',
action=EnvDefault,
envvar='MARABUNTA_MODE',
required=False,
help="Specify the mode in which we run the migration,"
"such as 'demo' or 'prod'. Additional operations "
"of this mode will be executed after the main "
"operations and the addons list of this mode "
"will be merged with the main addons list.")
parser.add_argument('--allow-serie',
action=BoolEnvDefault,
required=False,
envvar='MARABUNTA_ALLOW_SERIE',
help='Allow to run more than 1 version upgrade at a '
'time.')
parser.add_argument('--force-version',
required=False,
default=os.environ.get('MARABUNTA_FORCE_VERSION'),
help='Force upgrade of a version, even if it has '
'already been applied.')
group = parser.add_argument_group(
title='Web',
description='Configuration related to the internal web server, '
'used to publish a maintenance page during the migration.',
)
group.add_argument('--web-host',
required=False,
default=os.environ.get('MARABUNTA_WEB_HOST', '0.0.0.0'),
help='Host for the web server')
group.add_argument('--web-port',
required=False,
default=os.environ.get('MARABUNTA_WEB_PORT', 8069),
help='Port for the web server')
group.add_argument('--web-custom-html',
required=False,
default=os.environ.get(
'MARABUNTA_WEB_CUSTOM_HTML'
),
help='Path to a custom html file to publish')
return parser
|
python
|
def get_args_parser():
"""Return a parser for command line options."""
parser = argparse.ArgumentParser(
description='Marabunta: Migrating ants for Odoo')
parser.add_argument('--migration-file', '-f',
action=EnvDefault,
envvar='MARABUNTA_MIGRATION_FILE',
required=True,
help='The yaml file containing the migration steps')
parser.add_argument('--database', '-d',
action=EnvDefault,
envvar='MARABUNTA_DATABASE',
required=True,
help="Odoo's database")
parser.add_argument('--db-user', '-u',
action=EnvDefault,
envvar='MARABUNTA_DB_USER',
required=True,
help="Odoo's database user")
parser.add_argument('--db-password', '-w',
action=EnvDefault,
envvar='MARABUNTA_DB_PASSWORD',
required=True,
help="Odoo's database password")
parser.add_argument('--db-port', '-p',
default=os.environ.get('MARABUNTA_DB_PORT', 5432),
help="Odoo's database port")
parser.add_argument('--db-host', '-H',
default=os.environ.get('MARABUNTA_DB_HOST',
'localhost'),
help="Odoo's database host")
parser.add_argument('--mode',
action=EnvDefault,
envvar='MARABUNTA_MODE',
required=False,
help="Specify the mode in which we run the migration,"
"such as 'demo' or 'prod'. Additional operations "
"of this mode will be executed after the main "
"operations and the addons list of this mode "
"will be merged with the main addons list.")
parser.add_argument('--allow-serie',
action=BoolEnvDefault,
required=False,
envvar='MARABUNTA_ALLOW_SERIE',
help='Allow to run more than 1 version upgrade at a '
'time.')
parser.add_argument('--force-version',
required=False,
default=os.environ.get('MARABUNTA_FORCE_VERSION'),
help='Force upgrade of a version, even if it has '
'already been applied.')
group = parser.add_argument_group(
title='Web',
description='Configuration related to the internal web server, '
'used to publish a maintenance page during the migration.',
)
group.add_argument('--web-host',
required=False,
default=os.environ.get('MARABUNTA_WEB_HOST', '0.0.0.0'),
help='Host for the web server')
group.add_argument('--web-port',
required=False,
default=os.environ.get('MARABUNTA_WEB_PORT', 8069),
help='Port for the web server')
group.add_argument('--web-custom-html',
required=False,
default=os.environ.get(
'MARABUNTA_WEB_CUSTOM_HTML'
),
help='Path to a custom html file to publish')
return parser
|
[
"def",
"get_args_parser",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"'Marabunta: Migrating ants for Odoo'",
")",
"parser",
".",
"add_argument",
"(",
"'--migration-file'",
",",
"'-f'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_MIGRATION_FILE'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"'The yaml file containing the migration steps'",
")",
"parser",
".",
"add_argument",
"(",
"'--database'",
",",
"'-d'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_DATABASE'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"\"Odoo's database\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-user'",
",",
"'-u'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_DB_USER'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"\"Odoo's database user\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-password'",
",",
"'-w'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_DB_PASSWORD'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"\"Odoo's database password\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-port'",
",",
"'-p'",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_DB_PORT'",
",",
"5432",
")",
",",
"help",
"=",
"\"Odoo's database port\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-host'",
",",
"'-H'",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_DB_HOST'",
",",
"'localhost'",
")",
",",
"help",
"=",
"\"Odoo's database host\"",
")",
"parser",
".",
"add_argument",
"(",
"'--mode'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_MODE'",
",",
"required",
"=",
"False",
",",
"help",
"=",
"\"Specify the mode in which we run the migration,\"",
"\"such as 'demo' or 'prod'. Additional operations \"",
"\"of this mode will be executed after the main \"",
"\"operations and the addons list of this mode \"",
"\"will be merged with the main addons list.\"",
")",
"parser",
".",
"add_argument",
"(",
"'--allow-serie'",
",",
"action",
"=",
"BoolEnvDefault",
",",
"required",
"=",
"False",
",",
"envvar",
"=",
"'MARABUNTA_ALLOW_SERIE'",
",",
"help",
"=",
"'Allow to run more than 1 version upgrade at a '",
"'time.'",
")",
"parser",
".",
"add_argument",
"(",
"'--force-version'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_FORCE_VERSION'",
")",
",",
"help",
"=",
"'Force upgrade of a version, even if it has '",
"'already been applied.'",
")",
"group",
"=",
"parser",
".",
"add_argument_group",
"(",
"title",
"=",
"'Web'",
",",
"description",
"=",
"'Configuration related to the internal web server, '",
"'used to publish a maintenance page during the migration.'",
",",
")",
"group",
".",
"add_argument",
"(",
"'--web-host'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_WEB_HOST'",
",",
"'0.0.0.0'",
")",
",",
"help",
"=",
"'Host for the web server'",
")",
"group",
".",
"add_argument",
"(",
"'--web-port'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_WEB_PORT'",
",",
"8069",
")",
",",
"help",
"=",
"'Port for the web server'",
")",
"group",
".",
"add_argument",
"(",
"'--web-custom-html'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_WEB_CUSTOM_HTML'",
")",
",",
"help",
"=",
"'Path to a custom html file to publish'",
")",
"return",
"parser"
] |
Return a parser for command line options.
|
[
"Return",
"a",
"parser",
"for",
"command",
"line",
"options",
"."
] |
ec3a7a725c7426d6ed642e0a80119b37880eb91e
|
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/config.py#L90-L161
|
train
|
camptocamp/marabunta
|
marabunta/config.py
|
Config.from_parse_args
|
def from_parse_args(cls, args):
"""Constructor from command line args.
:param args: parse command line arguments
:type args: argparse.ArgumentParser
"""
return cls(args.migration_file,
args.database,
db_user=args.db_user,
db_password=args.db_password,
db_port=args.db_port,
db_host=args.db_host,
mode=args.mode,
allow_serie=args.allow_serie,
force_version=args.force_version,
web_host=args.web_host,
web_port=args.web_port,
web_custom_html=args.web_custom_html,
)
|
python
|
def from_parse_args(cls, args):
"""Constructor from command line args.
:param args: parse command line arguments
:type args: argparse.ArgumentParser
"""
return cls(args.migration_file,
args.database,
db_user=args.db_user,
db_password=args.db_password,
db_port=args.db_port,
db_host=args.db_host,
mode=args.mode,
allow_serie=args.allow_serie,
force_version=args.force_version,
web_host=args.web_host,
web_port=args.web_port,
web_custom_html=args.web_custom_html,
)
|
[
"def",
"from_parse_args",
"(",
"cls",
",",
"args",
")",
":",
"return",
"cls",
"(",
"args",
".",
"migration_file",
",",
"args",
".",
"database",
",",
"db_user",
"=",
"args",
".",
"db_user",
",",
"db_password",
"=",
"args",
".",
"db_password",
",",
"db_port",
"=",
"args",
".",
"db_port",
",",
"db_host",
"=",
"args",
".",
"db_host",
",",
"mode",
"=",
"args",
".",
"mode",
",",
"allow_serie",
"=",
"args",
".",
"allow_serie",
",",
"force_version",
"=",
"args",
".",
"force_version",
",",
"web_host",
"=",
"args",
".",
"web_host",
",",
"web_port",
"=",
"args",
".",
"web_port",
",",
"web_custom_html",
"=",
"args",
".",
"web_custom_html",
",",
")"
] |
Constructor from command line args.
:param args: parse command line arguments
:type args: argparse.ArgumentParser
|
[
"Constructor",
"from",
"command",
"line",
"args",
"."
] |
ec3a7a725c7426d6ed642e0a80119b37880eb91e
|
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/config.py#L40-L60
|
train
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.