repo
stringlengths
7
55
path
stringlengths
4
223
func_name
stringlengths
1
134
original_string
stringlengths
75
104k
language
stringclasses
1 value
code
stringlengths
75
104k
code_tokens
listlengths
19
28.4k
docstring
stringlengths
1
46.9k
docstring_tokens
listlengths
1
1.97k
sha
stringlengths
40
40
url
stringlengths
87
315
partition
stringclasses
1 value
ergo/ziggurat_foundations
ziggurat_foundations/models/user.py
UserMixin.id
def id(self): """ Unique identifier of user object""" return sa.Column(sa.Integer, primary_key=True, autoincrement=True)
python
def id(self): """ Unique identifier of user object""" return sa.Column(sa.Integer, primary_key=True, autoincrement=True)
[ "def", "id", "(", "self", ")", ":", "return", "sa", ".", "Column", "(", "sa", ".", "Integer", ",", "primary_key", "=", "True", ",", "autoincrement", "=", "True", ")" ]
Unique identifier of user object
[ "Unique", "identifier", "of", "user", "object" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/user.py#L29-L31
train
ergo/ziggurat_foundations
ziggurat_foundations/models/user.py
UserMixin.last_login_date
def last_login_date(self): """ Date of user's last login """ return sa.Column( sa.TIMESTAMP(timezone=False), default=lambda x: datetime.utcnow(), server_default=sa.func.now(), )
python
def last_login_date(self): """ Date of user's last login """ return sa.Column( sa.TIMESTAMP(timezone=False), default=lambda x: datetime.utcnow(), server_default=sa.func.now(), )
[ "def", "last_login_date", "(", "self", ")", ":", "return", "sa", ".", "Column", "(", "sa", ".", "TIMESTAMP", "(", "timezone", "=", "False", ")", ",", "default", "=", "lambda", "x", ":", "datetime", ".", "utcnow", "(", ")", ",", "server_default", "=", "sa", ".", "func", ".", "now", "(", ")", ",", ")" ]
Date of user's last login
[ "Date", "of", "user", "s", "last", "login" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/user.py#L59-L65
train
ergo/ziggurat_foundations
ziggurat_foundations/models/user.py
UserMixin.security_code_date
def security_code_date(self): """ Date of user's security code update """ return sa.Column( sa.TIMESTAMP(timezone=False), default=datetime(2000, 1, 1), server_default="2000-01-01 01:01", )
python
def security_code_date(self): """ Date of user's security code update """ return sa.Column( sa.TIMESTAMP(timezone=False), default=datetime(2000, 1, 1), server_default="2000-01-01 01:01", )
[ "def", "security_code_date", "(", "self", ")", ":", "return", "sa", ".", "Column", "(", "sa", ".", "TIMESTAMP", "(", "timezone", "=", "False", ")", ",", "default", "=", "datetime", "(", "2000", ",", "1", ",", "1", ")", ",", "server_default", "=", "\"2000-01-01 01:01\"", ",", ")" ]
Date of user's security code update
[ "Date", "of", "user", "s", "security", "code", "update" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/user.py#L77-L83
train
ergo/ziggurat_foundations
ziggurat_foundations/models/user.py
UserMixin.groups_dynamic
def groups_dynamic(self): """ returns dynamic relationship for groups - allowing for filtering of data """ return sa.orm.relationship( "Group", secondary="users_groups", lazy="dynamic", passive_deletes=True, passive_updates=True, )
python
def groups_dynamic(self): """ returns dynamic relationship for groups - allowing for filtering of data """ return sa.orm.relationship( "Group", secondary="users_groups", lazy="dynamic", passive_deletes=True, passive_updates=True, )
[ "def", "groups_dynamic", "(", "self", ")", ":", "return", "sa", ".", "orm", ".", "relationship", "(", "\"Group\"", ",", "secondary", "=", "\"users_groups\"", ",", "lazy", "=", "\"dynamic\"", ",", "passive_deletes", "=", "True", ",", "passive_updates", "=", "True", ",", ")" ]
returns dynamic relationship for groups - allowing for filtering of data
[ "returns", "dynamic", "relationship", "for", "groups", "-", "allowing", "for", "filtering", "of", "data" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/user.py#L89-L98
train
ergo/ziggurat_foundations
ziggurat_foundations/models/user.py
UserMixin.resources
def resources(self): """ Returns all resources directly owned by user, can be used to assign ownership of new resources:: user.resources.append(resource) """ return sa.orm.relationship( "Resource", cascade="all", passive_deletes=True, passive_updates=True, backref="owner", lazy="dynamic", )
python
def resources(self): """ Returns all resources directly owned by user, can be used to assign ownership of new resources:: user.resources.append(resource) """ return sa.orm.relationship( "Resource", cascade="all", passive_deletes=True, passive_updates=True, backref="owner", lazy="dynamic", )
[ "def", "resources", "(", "self", ")", ":", "return", "sa", ".", "orm", ".", "relationship", "(", "\"Resource\"", ",", "cascade", "=", "\"all\"", ",", "passive_deletes", "=", "True", ",", "passive_updates", "=", "True", ",", "backref", "=", "\"owner\"", ",", "lazy", "=", "\"dynamic\"", ",", ")" ]
Returns all resources directly owned by user, can be used to assign ownership of new resources:: user.resources.append(resource)
[ "Returns", "all", "resources", "directly", "owned", "by", "user", "can", "be", "used", "to", "assign", "ownership", "of", "new", "resources", "::" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/user.py#L126-L138
train
ergo/ziggurat_foundations
ziggurat_foundations/models/group.py
GroupMixin.resources_dynamic
def resources_dynamic(self): """ Returns all resources directly owned by group, can be used to assign ownership of new resources:: user.resources.append(resource) """ return sa.orm.relationship( "Resource", cascade="all", passive_deletes=True, passive_updates=True, lazy="dynamic", )
python
def resources_dynamic(self): """ Returns all resources directly owned by group, can be used to assign ownership of new resources:: user.resources.append(resource) """ return sa.orm.relationship( "Resource", cascade="all", passive_deletes=True, passive_updates=True, lazy="dynamic", )
[ "def", "resources_dynamic", "(", "self", ")", ":", "return", "sa", ".", "orm", ".", "relationship", "(", "\"Resource\"", ",", "cascade", "=", "\"all\"", ",", "passive_deletes", "=", "True", ",", "passive_updates", "=", "True", ",", "lazy", "=", "\"dynamic\"", ",", ")" ]
Returns all resources directly owned by group, can be used to assign ownership of new resources:: user.resources.append(resource)
[ "Returns", "all", "resources", "directly", "owned", "by", "group", "can", "be", "used", "to", "assign", "ownership", "of", "new", "resources", "::" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/group.py#L99-L110
train
ergo/ziggurat_foundations
ziggurat_foundations/models/group.py
GroupMixin.validate_permission
def validate_permission(self, key, permission): """ validates if group can get assigned with permission""" if permission.perm_name not in self.__possible_permissions__: raise AssertionError( "perm_name is not one of {}".format(self.__possible_permissions__) ) return permission
python
def validate_permission(self, key, permission): """ validates if group can get assigned with permission""" if permission.perm_name not in self.__possible_permissions__: raise AssertionError( "perm_name is not one of {}".format(self.__possible_permissions__) ) return permission
[ "def", "validate_permission", "(", "self", ",", "key", ",", "permission", ")", ":", "if", "permission", ".", "perm_name", "not", "in", "self", ".", "__possible_permissions__", ":", "raise", "AssertionError", "(", "\"perm_name is not one of {}\"", ".", "format", "(", "self", ".", "__possible_permissions__", ")", ")", "return", "permission" ]
validates if group can get assigned with permission
[ "validates", "if", "group", "can", "get", "assigned", "with", "permission" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/group.py#L113-L119
train
stephenmcd/gnotty
gnotty/bots/rss.py
RSSMixin.parse_feeds
def parse_feeds(self, message_channel=True): """ Iterates through each of the feed URLs, parses their items, and sends any items to the channel that have not been previously been parsed. """ if parse: for feed_url in self.feeds: feed = parse(feed_url) for item in feed.entries: if item["id"] not in self.feed_items: self.feed_items.add(item["id"]) if message_channel: message = self.format_item_message(feed, item) self.message_channel(message) return
python
def parse_feeds(self, message_channel=True): """ Iterates through each of the feed URLs, parses their items, and sends any items to the channel that have not been previously been parsed. """ if parse: for feed_url in self.feeds: feed = parse(feed_url) for item in feed.entries: if item["id"] not in self.feed_items: self.feed_items.add(item["id"]) if message_channel: message = self.format_item_message(feed, item) self.message_channel(message) return
[ "def", "parse_feeds", "(", "self", ",", "message_channel", "=", "True", ")", ":", "if", "parse", ":", "for", "feed_url", "in", "self", ".", "feeds", ":", "feed", "=", "parse", "(", "feed_url", ")", "for", "item", "in", "feed", ".", "entries", ":", "if", "item", "[", "\"id\"", "]", "not", "in", "self", ".", "feed_items", ":", "self", ".", "feed_items", ".", "add", "(", "item", "[", "\"id\"", "]", ")", "if", "message_channel", ":", "message", "=", "self", ".", "format_item_message", "(", "feed", ",", "item", ")", "self", ".", "message_channel", "(", "message", ")", "return" ]
Iterates through each of the feed URLs, parses their items, and sends any items to the channel that have not been previously been parsed.
[ "Iterates", "through", "each", "of", "the", "feed", "URLs", "parses", "their", "items", "and", "sends", "any", "items", "to", "the", "channel", "that", "have", "not", "been", "previously", "been", "parsed", "." ]
bea3762dc9cbc3cb21a5ae7224091cf027273c40
https://github.com/stephenmcd/gnotty/blob/bea3762dc9cbc3cb21a5ae7224091cf027273c40/gnotty/bots/rss.py#L30-L45
train
Xython/Linq.py
linq/standard/dict.py
ChunkBy
def ChunkBy(self: dict, f=None): """ [ { 'self': [1, 1, 3, 3, 1, 1], 'f': lambda x: x%2, 'assert': lambda ret: ret == [[1, 1], [3, 3], [1, 1]] } ] """ if f is None: return _chunk(self.items()) if is_to_destruct(f): f = destruct_func(f) return _chunk(self.items(), f)
python
def ChunkBy(self: dict, f=None): """ [ { 'self': [1, 1, 3, 3, 1, 1], 'f': lambda x: x%2, 'assert': lambda ret: ret == [[1, 1], [3, 3], [1, 1]] } ] """ if f is None: return _chunk(self.items()) if is_to_destruct(f): f = destruct_func(f) return _chunk(self.items(), f)
[ "def", "ChunkBy", "(", "self", ":", "dict", ",", "f", "=", "None", ")", ":", "if", "f", "is", "None", ":", "return", "_chunk", "(", "self", ".", "items", "(", ")", ")", "if", "is_to_destruct", "(", "f", ")", ":", "f", "=", "destruct_func", "(", "f", ")", "return", "_chunk", "(", "self", ".", "items", "(", ")", ",", "f", ")" ]
[ { 'self': [1, 1, 3, 3, 1, 1], 'f': lambda x: x%2, 'assert': lambda ret: ret == [[1, 1], [3, 3], [1, 1]] } ]
[ "[", "{", "self", ":", "[", "1", "1", "3", "3", "1", "1", "]", "f", ":", "lambda", "x", ":", "x%2", "assert", ":", "lambda", "ret", ":", "ret", "==", "[[", "1", "1", "]", "[", "3", "3", "]", "[", "1", "1", "]]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L74-L88
train
Xython/Linq.py
linq/standard/dict.py
GroupBy
def GroupBy(self: dict, f=None): """ [ { 'self': [1, 2, 3], 'f': lambda x: x%2, 'assert': lambda ret: ret[0] == [2] and ret[1] == [1, 3] } ] """ if f and is_to_destruct(f): f = destruct_func(f) return _group_by(self.items(), f)
python
def GroupBy(self: dict, f=None): """ [ { 'self': [1, 2, 3], 'f': lambda x: x%2, 'assert': lambda ret: ret[0] == [2] and ret[1] == [1, 3] } ] """ if f and is_to_destruct(f): f = destruct_func(f) return _group_by(self.items(), f)
[ "def", "GroupBy", "(", "self", ":", "dict", ",", "f", "=", "None", ")", ":", "if", "f", "and", "is_to_destruct", "(", "f", ")", ":", "f", "=", "destruct_func", "(", "f", ")", "return", "_group_by", "(", "self", ".", "items", "(", ")", ",", "f", ")" ]
[ { 'self': [1, 2, 3], 'f': lambda x: x%2, 'assert': lambda ret: ret[0] == [2] and ret[1] == [1, 3] } ]
[ "[", "{", "self", ":", "[", "1", "2", "3", "]", "f", ":", "lambda", "x", ":", "x%2", "assert", ":", "lambda", "ret", ":", "ret", "[", "0", "]", "==", "[", "2", "]", "and", "ret", "[", "1", "]", "==", "[", "1", "3", "]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L92-L104
train
Xython/Linq.py
linq/standard/dict.py
Take
def Take(self: dict, n): """ [ { 'self': [1, 2, 3], 'n': 2, 'assert': lambda ret: list(ret) == [1, 2] } ] """ for i, e in enumerate(self.items()): if i == n: break yield e
python
def Take(self: dict, n): """ [ { 'self': [1, 2, 3], 'n': 2, 'assert': lambda ret: list(ret) == [1, 2] } ] """ for i, e in enumerate(self.items()): if i == n: break yield e
[ "def", "Take", "(", "self", ":", "dict", ",", "n", ")", ":", "for", "i", ",", "e", "in", "enumerate", "(", "self", ".", "items", "(", ")", ")", ":", "if", "i", "==", "n", ":", "break", "yield", "e" ]
[ { 'self': [1, 2, 3], 'n': 2, 'assert': lambda ret: list(ret) == [1, 2] } ]
[ "[", "{", "self", ":", "[", "1", "2", "3", "]", "n", ":", "2", "assert", ":", "lambda", "ret", ":", "list", "(", "ret", ")", "==", "[", "1", "2", "]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L108-L122
train
Xython/Linq.py
linq/standard/dict.py
TakeIf
def TakeIf(self: dict, f): """ [ { 'self': [1, 2, 3], 'f': lambda e: e%2, 'assert': lambda ret: list(ret) == [1, 3] } ] """ if is_to_destruct(f): f = destruct_func(f) return (e for e in self.items() if f(e))
python
def TakeIf(self: dict, f): """ [ { 'self': [1, 2, 3], 'f': lambda e: e%2, 'assert': lambda ret: list(ret) == [1, 3] } ] """ if is_to_destruct(f): f = destruct_func(f) return (e for e in self.items() if f(e))
[ "def", "TakeIf", "(", "self", ":", "dict", ",", "f", ")", ":", "if", "is_to_destruct", "(", "f", ")", ":", "f", "=", "destruct_func", "(", "f", ")", "return", "(", "e", "for", "e", "in", "self", ".", "items", "(", ")", "if", "f", "(", "e", ")", ")" ]
[ { 'self': [1, 2, 3], 'f': lambda e: e%2, 'assert': lambda ret: list(ret) == [1, 3] } ]
[ "[", "{", "self", ":", "[", "1", "2", "3", "]", "f", ":", "lambda", "e", ":", "e%2", "assert", ":", "lambda", "ret", ":", "list", "(", "ret", ")", "==", "[", "1", "3", "]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L126-L139
train
Xython/Linq.py
linq/standard/dict.py
TakeWhile
def TakeWhile(self: dict, f): """ [ { 'self': [1, 2, 3, 4, 5], 'f': lambda x: x < 4, 'assert': lambda ret: list(ret) == [1, 2, 3] } ] """ if is_to_destruct(f): f = destruct_func(f) for e in self.items(): if not f(e): break yield e
python
def TakeWhile(self: dict, f): """ [ { 'self': [1, 2, 3, 4, 5], 'f': lambda x: x < 4, 'assert': lambda ret: list(ret) == [1, 2, 3] } ] """ if is_to_destruct(f): f = destruct_func(f) for e in self.items(): if not f(e): break yield e
[ "def", "TakeWhile", "(", "self", ":", "dict", ",", "f", ")", ":", "if", "is_to_destruct", "(", "f", ")", ":", "f", "=", "destruct_func", "(", "f", ")", "for", "e", "in", "self", ".", "items", "(", ")", ":", "if", "not", "f", "(", "e", ")", ":", "break", "yield", "e" ]
[ { 'self': [1, 2, 3, 4, 5], 'f': lambda x: x < 4, 'assert': lambda ret: list(ret) == [1, 2, 3] } ]
[ "[", "{", "self", ":", "[", "1", "2", "3", "4", "5", "]", "f", ":", "lambda", "x", ":", "x", "<", "4", "assert", ":", "lambda", "ret", ":", "list", "(", "ret", ")", "==", "[", "1", "2", "3", "]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L143-L159
train
Xython/Linq.py
linq/standard/dict.py
Drop
def Drop(self: dict, n): """ [ { 'self': [1, 2, 3, 4, 5], 'n': 3, 'assert': lambda ret: list(ret) == [1, 2] } ] """ n = len(self) - n if n <= 0: yield from self.items() else: for i, e in enumerate(self.items()): if i == n: break yield e
python
def Drop(self: dict, n): """ [ { 'self': [1, 2, 3, 4, 5], 'n': 3, 'assert': lambda ret: list(ret) == [1, 2] } ] """ n = len(self) - n if n <= 0: yield from self.items() else: for i, e in enumerate(self.items()): if i == n: break yield e
[ "def", "Drop", "(", "self", ":", "dict", ",", "n", ")", ":", "n", "=", "len", "(", "self", ")", "-", "n", "if", "n", "<=", "0", ":", "yield", "from", "self", ".", "items", "(", ")", "else", ":", "for", "i", ",", "e", "in", "enumerate", "(", "self", ".", "items", "(", ")", ")", ":", "if", "i", "==", "n", ":", "break", "yield", "e" ]
[ { 'self': [1, 2, 3, 4, 5], 'n': 3, 'assert': lambda ret: list(ret) == [1, 2] } ]
[ "[", "{", "self", ":", "[", "1", "2", "3", "4", "5", "]", "n", ":", "3", "assert", ":", "lambda", "ret", ":", "list", "(", "ret", ")", "==", "[", "1", "2", "]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L163-L180
train
Xython/Linq.py
linq/standard/dict.py
Skip
def Skip(self: dict, n): """ [ { 'self': [1, 2, 3, 4, 5], 'n': 3, 'assert': lambda ret: list(ret) == [4, 5] } ] """ con = self.items() for i, _ in enumerate(con): if i == n: break return con
python
def Skip(self: dict, n): """ [ { 'self': [1, 2, 3, 4, 5], 'n': 3, 'assert': lambda ret: list(ret) == [4, 5] } ] """ con = self.items() for i, _ in enumerate(con): if i == n: break return con
[ "def", "Skip", "(", "self", ":", "dict", ",", "n", ")", ":", "con", "=", "self", ".", "items", "(", ")", "for", "i", ",", "_", "in", "enumerate", "(", "con", ")", ":", "if", "i", "==", "n", ":", "break", "return", "con" ]
[ { 'self': [1, 2, 3, 4, 5], 'n': 3, 'assert': lambda ret: list(ret) == [4, 5] } ]
[ "[", "{", "self", ":", "[", "1", "2", "3", "4", "5", "]", "n", ":", "3", "assert", ":", "lambda", "ret", ":", "list", "(", "ret", ")", "==", "[", "4", "5", "]", "}", "]" ]
ffb65f92f1df0d8161d5f835f5947554f6f33d72
https://github.com/Xython/Linq.py/blob/ffb65f92f1df0d8161d5f835f5947554f6f33d72/linq/standard/dict.py#L184-L199
train
aguinane/nem-reader
print_examples.py
print_meter_record
def print_meter_record(file_path, rows=5): """ Output readings for specified number of rows to console """ m = nr.read_nem_file(file_path) print('Header:', m.header) print('Transactions:', m.transactions) for nmi in m.readings: for channel in m.readings[nmi]: print(nmi, 'Channel', channel) for reading in m.readings[nmi][channel][-rows:]: print('', reading)
python
def print_meter_record(file_path, rows=5): """ Output readings for specified number of rows to console """ m = nr.read_nem_file(file_path) print('Header:', m.header) print('Transactions:', m.transactions) for nmi in m.readings: for channel in m.readings[nmi]: print(nmi, 'Channel', channel) for reading in m.readings[nmi][channel][-rows:]: print('', reading)
[ "def", "print_meter_record", "(", "file_path", ",", "rows", "=", "5", ")", ":", "m", "=", "nr", ".", "read_nem_file", "(", "file_path", ")", "print", "(", "'Header:'", ",", "m", ".", "header", ")", "print", "(", "'Transactions:'", ",", "m", ".", "transactions", ")", "for", "nmi", "in", "m", ".", "readings", ":", "for", "channel", "in", "m", ".", "readings", "[", "nmi", "]", ":", "print", "(", "nmi", ",", "'Channel'", ",", "channel", ")", "for", "reading", "in", "m", ".", "readings", "[", "nmi", "]", "[", "channel", "]", "[", "-", "rows", ":", "]", ":", "print", "(", "''", ",", "reading", ")" ]
Output readings for specified number of rows to console
[ "Output", "readings", "for", "specified", "number", "of", "rows", "to", "console" ]
5405a5cba4bb8ebdad05c28455d12bb34a6d3ce5
https://github.com/aguinane/nem-reader/blob/5405a5cba4bb8ebdad05c28455d12bb34a6d3ce5/print_examples.py#L4-L13
train
ergo/ziggurat_foundations
ziggurat_foundations/models/resource.py
ResourceMixin.users
def users(self): """ returns all users that have permissions for this resource""" return sa.orm.relationship( "User", secondary="users_resources_permissions", passive_deletes=True, passive_updates=True, )
python
def users(self): """ returns all users that have permissions for this resource""" return sa.orm.relationship( "User", secondary="users_resources_permissions", passive_deletes=True, passive_updates=True, )
[ "def", "users", "(", "self", ")", ":", "return", "sa", ".", "orm", ".", "relationship", "(", "\"User\"", ",", "secondary", "=", "\"users_resources_permissions\"", ",", "passive_deletes", "=", "True", ",", "passive_updates", "=", "True", ",", ")" ]
returns all users that have permissions for this resource
[ "returns", "all", "users", "that", "have", "permissions", "for", "this", "resource" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/resource.py#L98-L105
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.from_resource_deeper
def from_resource_deeper( cls, resource_id=None, limit_depth=1000000, db_session=None, *args, **kwargs ): """ This returns you subtree of ordered objects relative to the start resource_id (currently only implemented in postgresql) :param resource_id: :param limit_depth: :param db_session: :return: """ tablename = cls.model.__table__.name raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.*, 1 AS depth, LPAD(res.ordering::CHARACTER VARYING, 7, '0') AS sorting, res.resource_id::CHARACTER VARYING AS path FROM {tablename} AS res WHERE res.resource_id = :resource_id UNION ALL SELECT res_u.*, depth+1 AS depth, (st.sorting::CHARACTER VARYING || '/' || LPAD(res_u.ordering::CHARACTER VARYING, 7, '0') ) AS sorting, (st.path::CHARACTER VARYING || '/' || res_u.resource_id::CHARACTER VARYING ) AS path FROM {tablename} res_u, subtree st WHERE res_u.parent_id = st.resource_id ) SELECT * FROM subtree WHERE depth<=:depth ORDER BY sorting; """.format( tablename=tablename ) # noqa db_session = get_db_session(db_session) text_obj = sa.text(raw_q) query = db_session.query(cls.model, "depth", "sorting", "path") query = query.from_statement(text_obj) query = query.params(resource_id=resource_id, depth=limit_depth) return query
python
def from_resource_deeper( cls, resource_id=None, limit_depth=1000000, db_session=None, *args, **kwargs ): """ This returns you subtree of ordered objects relative to the start resource_id (currently only implemented in postgresql) :param resource_id: :param limit_depth: :param db_session: :return: """ tablename = cls.model.__table__.name raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.*, 1 AS depth, LPAD(res.ordering::CHARACTER VARYING, 7, '0') AS sorting, res.resource_id::CHARACTER VARYING AS path FROM {tablename} AS res WHERE res.resource_id = :resource_id UNION ALL SELECT res_u.*, depth+1 AS depth, (st.sorting::CHARACTER VARYING || '/' || LPAD(res_u.ordering::CHARACTER VARYING, 7, '0') ) AS sorting, (st.path::CHARACTER VARYING || '/' || res_u.resource_id::CHARACTER VARYING ) AS path FROM {tablename} res_u, subtree st WHERE res_u.parent_id = st.resource_id ) SELECT * FROM subtree WHERE depth<=:depth ORDER BY sorting; """.format( tablename=tablename ) # noqa db_session = get_db_session(db_session) text_obj = sa.text(raw_q) query = db_session.query(cls.model, "depth", "sorting", "path") query = query.from_statement(text_obj) query = query.params(resource_id=resource_id, depth=limit_depth) return query
[ "def", "from_resource_deeper", "(", "cls", ",", "resource_id", "=", "None", ",", "limit_depth", "=", "1000000", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "tablename", "=", "cls", ".", "model", ".", "__table__", ".", "name", "raw_q", "=", "\"\"\"\n WITH RECURSIVE subtree AS (\n SELECT res.*, 1 AS depth, LPAD(res.ordering::CHARACTER VARYING, 7, '0') AS sorting,\n res.resource_id::CHARACTER VARYING AS path\n FROM {tablename} AS res WHERE res.resource_id = :resource_id\n UNION ALL\n SELECT res_u.*, depth+1 AS depth,\n (st.sorting::CHARACTER VARYING || '/' || LPAD(res_u.ordering::CHARACTER VARYING, 7, '0') ) AS sorting,\n (st.path::CHARACTER VARYING || '/' || res_u.resource_id::CHARACTER VARYING ) AS path\n FROM {tablename} res_u, subtree st\n WHERE res_u.parent_id = st.resource_id\n )\n SELECT * FROM subtree WHERE depth<=:depth ORDER BY sorting;\n \"\"\"", ".", "format", "(", "tablename", "=", "tablename", ")", "# noqa", "db_session", "=", "get_db_session", "(", "db_session", ")", "text_obj", "=", "sa", ".", "text", "(", "raw_q", ")", "query", "=", "db_session", ".", "query", "(", "cls", ".", "model", ",", "\"depth\"", ",", "\"sorting\"", ",", "\"path\"", ")", "query", "=", "query", ".", "from_statement", "(", "text_obj", ")", "query", "=", "query", ".", "params", "(", "resource_id", "=", "resource_id", ",", "depth", "=", "limit_depth", ")", "return", "query" ]
This returns you subtree of ordered objects relative to the start resource_id (currently only implemented in postgresql) :param resource_id: :param limit_depth: :param db_session: :return:
[ "This", "returns", "you", "subtree", "of", "ordered", "objects", "relative", "to", "the", "start", "resource_id", "(", "currently", "only", "implemented", "in", "postgresql", ")" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L24-L58
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.delete_branch
def delete_branch(cls, resource_id=None, db_session=None, *args, **kwargs): """ This deletes whole branch with children starting from resource_id :param resource_id: :param db_session: :return: """ tablename = cls.model.__table__.name # lets lock rows to prevent bad tree states resource = ResourceService.lock_resource_for_update( resource_id=resource_id, db_session=db_session ) parent_id = resource.parent_id ordering = resource.ordering raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.resource_id FROM {tablename} AS res WHERE res.resource_id = :resource_id UNION ALL SELECT res_u.resource_id FROM {tablename} res_u, subtree st WHERE res_u.parent_id = st.resource_id ) DELETE FROM resources where resource_id in (select * from subtree); """.format( tablename=tablename ) # noqa db_session = get_db_session(db_session) text_obj = sa.text(raw_q) db_session.execute(text_obj, params={"resource_id": resource_id}) cls.shift_ordering_down(parent_id, ordering, db_session=db_session) return True
python
def delete_branch(cls, resource_id=None, db_session=None, *args, **kwargs): """ This deletes whole branch with children starting from resource_id :param resource_id: :param db_session: :return: """ tablename = cls.model.__table__.name # lets lock rows to prevent bad tree states resource = ResourceService.lock_resource_for_update( resource_id=resource_id, db_session=db_session ) parent_id = resource.parent_id ordering = resource.ordering raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.resource_id FROM {tablename} AS res WHERE res.resource_id = :resource_id UNION ALL SELECT res_u.resource_id FROM {tablename} res_u, subtree st WHERE res_u.parent_id = st.resource_id ) DELETE FROM resources where resource_id in (select * from subtree); """.format( tablename=tablename ) # noqa db_session = get_db_session(db_session) text_obj = sa.text(raw_q) db_session.execute(text_obj, params={"resource_id": resource_id}) cls.shift_ordering_down(parent_id, ordering, db_session=db_session) return True
[ "def", "delete_branch", "(", "cls", ",", "resource_id", "=", "None", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "tablename", "=", "cls", ".", "model", ".", "__table__", ".", "name", "# lets lock rows to prevent bad tree states", "resource", "=", "ResourceService", ".", "lock_resource_for_update", "(", "resource_id", "=", "resource_id", ",", "db_session", "=", "db_session", ")", "parent_id", "=", "resource", ".", "parent_id", "ordering", "=", "resource", ".", "ordering", "raw_q", "=", "\"\"\"\n WITH RECURSIVE subtree AS (\n SELECT res.resource_id\n FROM {tablename} AS res WHERE res.resource_id = :resource_id\n UNION ALL\n SELECT res_u.resource_id\n FROM {tablename} res_u, subtree st\n WHERE res_u.parent_id = st.resource_id\n )\n DELETE FROM resources where resource_id in (select * from subtree);\n \"\"\"", ".", "format", "(", "tablename", "=", "tablename", ")", "# noqa", "db_session", "=", "get_db_session", "(", "db_session", ")", "text_obj", "=", "sa", ".", "text", "(", "raw_q", ")", "db_session", ".", "execute", "(", "text_obj", ",", "params", "=", "{", "\"resource_id\"", ":", "resource_id", "}", ")", "cls", ".", "shift_ordering_down", "(", "parent_id", ",", "ordering", ",", "db_session", "=", "db_session", ")", "return", "True" ]
This deletes whole branch with children starting from resource_id :param resource_id: :param db_session: :return:
[ "This", "deletes", "whole", "branch", "with", "children", "starting", "from", "resource_id" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L61-L94
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.from_parent_deeper
def from_parent_deeper( cls, parent_id=None, limit_depth=1000000, db_session=None, *args, **kwargs ): """ This returns you subtree of ordered objects relative to the start parent_id (currently only implemented in postgresql) :param resource_id: :param limit_depth: :param db_session: :return: """ if parent_id: limiting_clause = "res.parent_id = :parent_id" else: limiting_clause = "res.parent_id is null" tablename = cls.model.__table__.name raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.*, 1 AS depth, LPAD(res.ordering::CHARACTER VARYING, 7, '0') AS sorting, res.resource_id::CHARACTER VARYING AS path FROM {tablename} AS res WHERE {limiting_clause} UNION ALL SELECT res_u.*, depth+1 AS depth, (st.sorting::CHARACTER VARYING || '/' || LPAD(res_u.ordering::CHARACTER VARYING, 7, '0') ) AS sorting, (st.path::CHARACTER VARYING || '/' || res_u.resource_id::CHARACTER VARYING ) AS path FROM {tablename} res_u, subtree st WHERE res_u.parent_id = st.resource_id ) SELECT * FROM subtree WHERE depth<=:depth ORDER BY sorting; """.format( tablename=tablename, limiting_clause=limiting_clause ) # noqa db_session = get_db_session(db_session) text_obj = sa.text(raw_q) query = db_session.query(cls.model, "depth", "sorting", "path") query = query.from_statement(text_obj) query = query.params(parent_id=parent_id, depth=limit_depth) return query
python
def from_parent_deeper( cls, parent_id=None, limit_depth=1000000, db_session=None, *args, **kwargs ): """ This returns you subtree of ordered objects relative to the start parent_id (currently only implemented in postgresql) :param resource_id: :param limit_depth: :param db_session: :return: """ if parent_id: limiting_clause = "res.parent_id = :parent_id" else: limiting_clause = "res.parent_id is null" tablename = cls.model.__table__.name raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.*, 1 AS depth, LPAD(res.ordering::CHARACTER VARYING, 7, '0') AS sorting, res.resource_id::CHARACTER VARYING AS path FROM {tablename} AS res WHERE {limiting_clause} UNION ALL SELECT res_u.*, depth+1 AS depth, (st.sorting::CHARACTER VARYING || '/' || LPAD(res_u.ordering::CHARACTER VARYING, 7, '0') ) AS sorting, (st.path::CHARACTER VARYING || '/' || res_u.resource_id::CHARACTER VARYING ) AS path FROM {tablename} res_u, subtree st WHERE res_u.parent_id = st.resource_id ) SELECT * FROM subtree WHERE depth<=:depth ORDER BY sorting; """.format( tablename=tablename, limiting_clause=limiting_clause ) # noqa db_session = get_db_session(db_session) text_obj = sa.text(raw_q) query = db_session.query(cls.model, "depth", "sorting", "path") query = query.from_statement(text_obj) query = query.params(parent_id=parent_id, depth=limit_depth) return query
[ "def", "from_parent_deeper", "(", "cls", ",", "parent_id", "=", "None", ",", "limit_depth", "=", "1000000", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "parent_id", ":", "limiting_clause", "=", "\"res.parent_id = :parent_id\"", "else", ":", "limiting_clause", "=", "\"res.parent_id is null\"", "tablename", "=", "cls", ".", "model", ".", "__table__", ".", "name", "raw_q", "=", "\"\"\"\n WITH RECURSIVE subtree AS (\n SELECT res.*, 1 AS depth, LPAD(res.ordering::CHARACTER VARYING, 7, '0') AS sorting,\n res.resource_id::CHARACTER VARYING AS path\n FROM {tablename} AS res WHERE {limiting_clause}\n UNION ALL\n SELECT res_u.*, depth+1 AS depth,\n (st.sorting::CHARACTER VARYING || '/' || LPAD(res_u.ordering::CHARACTER VARYING, 7, '0') ) AS sorting,\n (st.path::CHARACTER VARYING || '/' || res_u.resource_id::CHARACTER VARYING ) AS path\n FROM {tablename} res_u, subtree st\n WHERE res_u.parent_id = st.resource_id\n )\n SELECT * FROM subtree WHERE depth<=:depth ORDER BY sorting;\n \"\"\"", ".", "format", "(", "tablename", "=", "tablename", ",", "limiting_clause", "=", "limiting_clause", ")", "# noqa", "db_session", "=", "get_db_session", "(", "db_session", ")", "text_obj", "=", "sa", ".", "text", "(", "raw_q", ")", "query", "=", "db_session", ".", "query", "(", "cls", ".", "model", ",", "\"depth\"", ",", "\"sorting\"", ",", "\"path\"", ")", "query", "=", "query", ".", "from_statement", "(", "text_obj", ")", "query", "=", "query", ".", "params", "(", "parent_id", "=", "parent_id", ",", "depth", "=", "limit_depth", ")", "return", "query" ]
This returns you subtree of ordered objects relative to the start parent_id (currently only implemented in postgresql) :param resource_id: :param limit_depth: :param db_session: :return:
[ "This", "returns", "you", "subtree", "of", "ordered", "objects", "relative", "to", "the", "start", "parent_id", "(", "currently", "only", "implemented", "in", "postgresql", ")" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L97-L136
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.build_subtree_strut
def build_subtree_strut(self, result, *args, **kwargs): """ Returns a dictionary in form of {node:Resource, children:{node_id: Resource}} :param result: :return: """ items = list(result) root_elem = {"node": None, "children": OrderedDict()} if len(items) == 0: return root_elem for _, node in enumerate(items): new_elem = {"node": node.Resource, "children": OrderedDict()} path = list(map(int, node.path.split("/"))) parent_node = root_elem normalized_path = path[:-1] if normalized_path: for path_part in normalized_path: parent_node = parent_node["children"][path_part] parent_node["children"][new_elem["node"].resource_id] = new_elem return root_elem
python
def build_subtree_strut(self, result, *args, **kwargs): """ Returns a dictionary in form of {node:Resource, children:{node_id: Resource}} :param result: :return: """ items = list(result) root_elem = {"node": None, "children": OrderedDict()} if len(items) == 0: return root_elem for _, node in enumerate(items): new_elem = {"node": node.Resource, "children": OrderedDict()} path = list(map(int, node.path.split("/"))) parent_node = root_elem normalized_path = path[:-1] if normalized_path: for path_part in normalized_path: parent_node = parent_node["children"][path_part] parent_node["children"][new_elem["node"].resource_id] = new_elem return root_elem
[ "def", "build_subtree_strut", "(", "self", ",", "result", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "items", "=", "list", "(", "result", ")", "root_elem", "=", "{", "\"node\"", ":", "None", ",", "\"children\"", ":", "OrderedDict", "(", ")", "}", "if", "len", "(", "items", ")", "==", "0", ":", "return", "root_elem", "for", "_", ",", "node", "in", "enumerate", "(", "items", ")", ":", "new_elem", "=", "{", "\"node\"", ":", "node", ".", "Resource", ",", "\"children\"", ":", "OrderedDict", "(", ")", "}", "path", "=", "list", "(", "map", "(", "int", ",", "node", ".", "path", ".", "split", "(", "\"/\"", ")", ")", ")", "parent_node", "=", "root_elem", "normalized_path", "=", "path", "[", ":", "-", "1", "]", "if", "normalized_path", ":", "for", "path_part", "in", "normalized_path", ":", "parent_node", "=", "parent_node", "[", "\"children\"", "]", "[", "path_part", "]", "parent_node", "[", "\"children\"", "]", "[", "new_elem", "[", "\"node\"", "]", ".", "resource_id", "]", "=", "new_elem", "return", "root_elem" ]
Returns a dictionary in form of {node:Resource, children:{node_id: Resource}} :param result: :return:
[ "Returns", "a", "dictionary", "in", "form", "of", "{", "node", ":", "Resource", "children", ":", "{", "node_id", ":", "Resource", "}}" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L139-L160
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.path_upper
def path_upper( cls, object_id, limit_depth=1000000, db_session=None, *args, **kwargs ): """ This returns you path to root node starting from object_id currently only for postgresql :param object_id: :param limit_depth: :param db_session: :return: """ tablename = cls.model.__table__.name raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.*, 1 as depth FROM {tablename} res WHERE res.resource_id = :resource_id UNION ALL SELECT res_u.*, depth+1 as depth FROM {tablename} res_u, subtree st WHERE res_u.resource_id = st.parent_id ) SELECT * FROM subtree WHERE depth<=:depth; """.format( tablename=tablename ) db_session = get_db_session(db_session) q = ( db_session.query(cls.model) .from_statement(sa.text(raw_q)) .params(resource_id=object_id, depth=limit_depth) ) return q
python
def path_upper( cls, object_id, limit_depth=1000000, db_session=None, *args, **kwargs ): """ This returns you path to root node starting from object_id currently only for postgresql :param object_id: :param limit_depth: :param db_session: :return: """ tablename = cls.model.__table__.name raw_q = """ WITH RECURSIVE subtree AS ( SELECT res.*, 1 as depth FROM {tablename} res WHERE res.resource_id = :resource_id UNION ALL SELECT res_u.*, depth+1 as depth FROM {tablename} res_u, subtree st WHERE res_u.resource_id = st.parent_id ) SELECT * FROM subtree WHERE depth<=:depth; """.format( tablename=tablename ) db_session = get_db_session(db_session) q = ( db_session.query(cls.model) .from_statement(sa.text(raw_q)) .params(resource_id=object_id, depth=limit_depth) ) return q
[ "def", "path_upper", "(", "cls", ",", "object_id", ",", "limit_depth", "=", "1000000", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "tablename", "=", "cls", ".", "model", ".", "__table__", ".", "name", "raw_q", "=", "\"\"\"\n WITH RECURSIVE subtree AS (\n SELECT res.*, 1 as depth FROM {tablename} res\n WHERE res.resource_id = :resource_id\n UNION ALL\n SELECT res_u.*, depth+1 as depth\n FROM {tablename} res_u, subtree st\n WHERE res_u.resource_id = st.parent_id\n )\n SELECT * FROM subtree WHERE depth<=:depth;\n \"\"\"", ".", "format", "(", "tablename", "=", "tablename", ")", "db_session", "=", "get_db_session", "(", "db_session", ")", "q", "=", "(", "db_session", ".", "query", "(", "cls", ".", "model", ")", ".", "from_statement", "(", "sa", ".", "text", "(", "raw_q", ")", ")", ".", "params", "(", "resource_id", "=", "object_id", ",", "depth", "=", "limit_depth", ")", ")", "return", "q" ]
This returns you path to root node starting from object_id currently only for postgresql :param object_id: :param limit_depth: :param db_session: :return:
[ "This", "returns", "you", "path", "to", "root", "node", "starting", "from", "object_id", "currently", "only", "for", "postgresql" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L163-L195
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.move_to_position
def move_to_position( cls, resource_id, to_position, new_parent_id=noop, db_session=None, *args, **kwargs ): """ Moves node to new location in the tree :param resource_id: resource to move :param to_position: new position :param new_parent_id: new parent id :param db_session: :return: """ db_session = get_db_session(db_session) # lets lock rows to prevent bad tree states resource = ResourceService.lock_resource_for_update( resource_id=resource_id, db_session=db_session ) ResourceService.lock_resource_for_update( resource_id=resource.parent_id, db_session=db_session ) same_branch = False # reset if parent is same as old if new_parent_id == resource.parent_id: new_parent_id = noop if new_parent_id is not noop: cls.check_node_parent(resource_id, new_parent_id, db_session=db_session) else: same_branch = True if new_parent_id is noop: # it is not guaranteed that parent exists parent_id = resource.parent_id if resource else None else: parent_id = new_parent_id cls.check_node_position( parent_id, to_position, on_same_branch=same_branch, db_session=db_session ) # move on same branch if new_parent_id is noop: order_range = list(sorted((resource.ordering, to_position))) move_down = resource.ordering > to_position query = db_session.query(cls.model) query = query.filter(cls.model.parent_id == parent_id) query = query.filter(cls.model.ordering.between(*order_range)) if move_down: query.update( {cls.model.ordering: cls.model.ordering + 1}, synchronize_session=False, ) else: query.update( {cls.model.ordering: cls.model.ordering - 1}, synchronize_session=False, ) db_session.flush() db_session.expire(resource) resource.ordering = to_position # move between branches else: cls.shift_ordering_down( resource.parent_id, resource.ordering, db_session=db_session ) cls.shift_ordering_up(new_parent_id, to_position, db_session=db_session) db_session.expire(resource) resource.parent_id = new_parent_id resource.ordering = to_position db_session.flush() return True
python
def move_to_position( cls, resource_id, to_position, new_parent_id=noop, db_session=None, *args, **kwargs ): """ Moves node to new location in the tree :param resource_id: resource to move :param to_position: new position :param new_parent_id: new parent id :param db_session: :return: """ db_session = get_db_session(db_session) # lets lock rows to prevent bad tree states resource = ResourceService.lock_resource_for_update( resource_id=resource_id, db_session=db_session ) ResourceService.lock_resource_for_update( resource_id=resource.parent_id, db_session=db_session ) same_branch = False # reset if parent is same as old if new_parent_id == resource.parent_id: new_parent_id = noop if new_parent_id is not noop: cls.check_node_parent(resource_id, new_parent_id, db_session=db_session) else: same_branch = True if new_parent_id is noop: # it is not guaranteed that parent exists parent_id = resource.parent_id if resource else None else: parent_id = new_parent_id cls.check_node_position( parent_id, to_position, on_same_branch=same_branch, db_session=db_session ) # move on same branch if new_parent_id is noop: order_range = list(sorted((resource.ordering, to_position))) move_down = resource.ordering > to_position query = db_session.query(cls.model) query = query.filter(cls.model.parent_id == parent_id) query = query.filter(cls.model.ordering.between(*order_range)) if move_down: query.update( {cls.model.ordering: cls.model.ordering + 1}, synchronize_session=False, ) else: query.update( {cls.model.ordering: cls.model.ordering - 1}, synchronize_session=False, ) db_session.flush() db_session.expire(resource) resource.ordering = to_position # move between branches else: cls.shift_ordering_down( resource.parent_id, resource.ordering, db_session=db_session ) cls.shift_ordering_up(new_parent_id, to_position, db_session=db_session) db_session.expire(resource) resource.parent_id = new_parent_id resource.ordering = to_position db_session.flush() return True
[ "def", "move_to_position", "(", "cls", ",", "resource_id", ",", "to_position", ",", "new_parent_id", "=", "noop", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "# lets lock rows to prevent bad tree states", "resource", "=", "ResourceService", ".", "lock_resource_for_update", "(", "resource_id", "=", "resource_id", ",", "db_session", "=", "db_session", ")", "ResourceService", ".", "lock_resource_for_update", "(", "resource_id", "=", "resource", ".", "parent_id", ",", "db_session", "=", "db_session", ")", "same_branch", "=", "False", "# reset if parent is same as old", "if", "new_parent_id", "==", "resource", ".", "parent_id", ":", "new_parent_id", "=", "noop", "if", "new_parent_id", "is", "not", "noop", ":", "cls", ".", "check_node_parent", "(", "resource_id", ",", "new_parent_id", ",", "db_session", "=", "db_session", ")", "else", ":", "same_branch", "=", "True", "if", "new_parent_id", "is", "noop", ":", "# it is not guaranteed that parent exists", "parent_id", "=", "resource", ".", "parent_id", "if", "resource", "else", "None", "else", ":", "parent_id", "=", "new_parent_id", "cls", ".", "check_node_position", "(", "parent_id", ",", "to_position", ",", "on_same_branch", "=", "same_branch", ",", "db_session", "=", "db_session", ")", "# move on same branch", "if", "new_parent_id", "is", "noop", ":", "order_range", "=", "list", "(", "sorted", "(", "(", "resource", ".", "ordering", ",", "to_position", ")", ")", ")", "move_down", "=", "resource", ".", "ordering", ">", "to_position", "query", "=", "db_session", ".", "query", "(", "cls", ".", "model", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "model", ".", "parent_id", "==", "parent_id", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "model", ".", "ordering", ".", "between", "(", "*", "order_range", ")", ")", "if", "move_down", ":", "query", ".", "update", "(", "{", "cls", ".", "model", ".", "ordering", ":", "cls", ".", "model", ".", "ordering", "+", "1", "}", ",", "synchronize_session", "=", "False", ",", ")", "else", ":", "query", ".", "update", "(", "{", "cls", ".", "model", ".", "ordering", ":", "cls", ".", "model", ".", "ordering", "-", "1", "}", ",", "synchronize_session", "=", "False", ",", ")", "db_session", ".", "flush", "(", ")", "db_session", ".", "expire", "(", "resource", ")", "resource", ".", "ordering", "=", "to_position", "# move between branches", "else", ":", "cls", ".", "shift_ordering_down", "(", "resource", ".", "parent_id", ",", "resource", ".", "ordering", ",", "db_session", "=", "db_session", ")", "cls", ".", "shift_ordering_up", "(", "new_parent_id", ",", "to_position", ",", "db_session", "=", "db_session", ")", "db_session", ".", "expire", "(", "resource", ")", "resource", ".", "parent_id", "=", "new_parent_id", "resource", ".", "ordering", "=", "to_position", "db_session", ".", "flush", "(", ")", "return", "True" ]
Moves node to new location in the tree :param resource_id: resource to move :param to_position: new position :param new_parent_id: new parent id :param db_session: :return:
[ "Moves", "node", "to", "new", "location", "in", "the", "tree" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L198-L275
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.shift_ordering_up
def shift_ordering_up(cls, parent_id, position, db_session=None, *args, **kwargs): """ Shifts ordering to "open a gap" for node insertion, begins the shift from given position :param parent_id: :param position: :param db_session: :return: """ db_session = get_db_session(db_session) query = db_session.query(cls.model) query = query.filter(cls.model.parent_id == parent_id) query = query.filter(cls.model.ordering >= position) query.update( {cls.model.ordering: cls.model.ordering + 1}, synchronize_session=False ) db_session.flush()
python
def shift_ordering_up(cls, parent_id, position, db_session=None, *args, **kwargs): """ Shifts ordering to "open a gap" for node insertion, begins the shift from given position :param parent_id: :param position: :param db_session: :return: """ db_session = get_db_session(db_session) query = db_session.query(cls.model) query = query.filter(cls.model.parent_id == parent_id) query = query.filter(cls.model.ordering >= position) query.update( {cls.model.ordering: cls.model.ordering + 1}, synchronize_session=False ) db_session.flush()
[ "def", "shift_ordering_up", "(", "cls", ",", "parent_id", ",", "position", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "query", "=", "db_session", ".", "query", "(", "cls", ".", "model", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "model", ".", "parent_id", "==", "parent_id", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "model", ".", "ordering", ">=", "position", ")", "query", ".", "update", "(", "{", "cls", ".", "model", ".", "ordering", ":", "cls", ".", "model", ".", "ordering", "+", "1", "}", ",", "synchronize_session", "=", "False", ")", "db_session", ".", "flush", "(", ")" ]
Shifts ordering to "open a gap" for node insertion, begins the shift from given position :param parent_id: :param position: :param db_session: :return:
[ "Shifts", "ordering", "to", "open", "a", "gap", "for", "node", "insertion", "begins", "the", "shift", "from", "given", "position" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L298-L315
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.set_position
def set_position(cls, resource_id, to_position, db_session=None, *args, **kwargs): """ Sets node position for new node in the tree :param resource_id: resource to move :param to_position: new position :param db_session: :return:def count_children(cls, resource_id, db_session=None): """ db_session = get_db_session(db_session) # lets lock rows to prevent bad tree states resource = ResourceService.lock_resource_for_update( resource_id=resource_id, db_session=db_session ) cls.check_node_position( resource.parent_id, to_position, on_same_branch=True, db_session=db_session ) cls.shift_ordering_up(resource.parent_id, to_position, db_session=db_session) db_session.flush() db_session.expire(resource) resource.ordering = to_position return True
python
def set_position(cls, resource_id, to_position, db_session=None, *args, **kwargs): """ Sets node position for new node in the tree :param resource_id: resource to move :param to_position: new position :param db_session: :return:def count_children(cls, resource_id, db_session=None): """ db_session = get_db_session(db_session) # lets lock rows to prevent bad tree states resource = ResourceService.lock_resource_for_update( resource_id=resource_id, db_session=db_session ) cls.check_node_position( resource.parent_id, to_position, on_same_branch=True, db_session=db_session ) cls.shift_ordering_up(resource.parent_id, to_position, db_session=db_session) db_session.flush() db_session.expire(resource) resource.ordering = to_position return True
[ "def", "set_position", "(", "cls", ",", "resource_id", ",", "to_position", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "# lets lock rows to prevent bad tree states", "resource", "=", "ResourceService", ".", "lock_resource_for_update", "(", "resource_id", "=", "resource_id", ",", "db_session", "=", "db_session", ")", "cls", ".", "check_node_position", "(", "resource", ".", "parent_id", ",", "to_position", ",", "on_same_branch", "=", "True", ",", "db_session", "=", "db_session", ")", "cls", ".", "shift_ordering_up", "(", "resource", ".", "parent_id", ",", "to_position", ",", "db_session", "=", "db_session", ")", "db_session", ".", "flush", "(", ")", "db_session", ".", "expire", "(", "resource", ")", "resource", ".", "ordering", "=", "to_position", "return", "True" ]
Sets node position for new node in the tree :param resource_id: resource to move :param to_position: new position :param db_session: :return:def count_children(cls, resource_id, db_session=None):
[ "Sets", "node", "position", "for", "new", "node", "in", "the", "tree" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L318-L339
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.check_node_parent
def check_node_parent( cls, resource_id, new_parent_id, db_session=None, *args, **kwargs ): """ Checks if parent destination is valid for node :param resource_id: :param new_parent_id: :param db_session: :return: """ db_session = get_db_session(db_session) new_parent = ResourceService.lock_resource_for_update( resource_id=new_parent_id, db_session=db_session ) # we are not moving to "root" so parent should be found if not new_parent and new_parent_id is not None: raise ZigguratResourceTreeMissingException("New parent node not found") else: result = cls.path_upper(new_parent_id, db_session=db_session) path_ids = [r.resource_id for r in result] if resource_id in path_ids: raise ZigguratResourceTreePathException( "Trying to insert node into itself" )
python
def check_node_parent( cls, resource_id, new_parent_id, db_session=None, *args, **kwargs ): """ Checks if parent destination is valid for node :param resource_id: :param new_parent_id: :param db_session: :return: """ db_session = get_db_session(db_session) new_parent = ResourceService.lock_resource_for_update( resource_id=new_parent_id, db_session=db_session ) # we are not moving to "root" so parent should be found if not new_parent and new_parent_id is not None: raise ZigguratResourceTreeMissingException("New parent node not found") else: result = cls.path_upper(new_parent_id, db_session=db_session) path_ids = [r.resource_id for r in result] if resource_id in path_ids: raise ZigguratResourceTreePathException( "Trying to insert node into itself" )
[ "def", "check_node_parent", "(", "cls", ",", "resource_id", ",", "new_parent_id", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "new_parent", "=", "ResourceService", ".", "lock_resource_for_update", "(", "resource_id", "=", "new_parent_id", ",", "db_session", "=", "db_session", ")", "# we are not moving to \"root\" so parent should be found", "if", "not", "new_parent", "and", "new_parent_id", "is", "not", "None", ":", "raise", "ZigguratResourceTreeMissingException", "(", "\"New parent node not found\"", ")", "else", ":", "result", "=", "cls", ".", "path_upper", "(", "new_parent_id", ",", "db_session", "=", "db_session", ")", "path_ids", "=", "[", "r", ".", "resource_id", "for", "r", "in", "result", "]", "if", "resource_id", "in", "path_ids", ":", "raise", "ZigguratResourceTreePathException", "(", "\"Trying to insert node into itself\"", ")" ]
Checks if parent destination is valid for node :param resource_id: :param new_parent_id: :param db_session: :return:
[ "Checks", "if", "parent", "destination", "is", "valid", "for", "node" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L342-L366
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.count_children
def count_children(cls, resource_id, db_session=None, *args, **kwargs): """ Counts children of resource node :param resource_id: :param db_session: :return: """ query = db_session.query(cls.model.resource_id) query = query.filter(cls.model.parent_id == resource_id) return query.count()
python
def count_children(cls, resource_id, db_session=None, *args, **kwargs): """ Counts children of resource node :param resource_id: :param db_session: :return: """ query = db_session.query(cls.model.resource_id) query = query.filter(cls.model.parent_id == resource_id) return query.count()
[ "def", "count_children", "(", "cls", ",", "resource_id", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "query", "=", "db_session", ".", "query", "(", "cls", ".", "model", ".", "resource_id", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "model", ".", "parent_id", "==", "resource_id", ")", "return", "query", ".", "count", "(", ")" ]
Counts children of resource node :param resource_id: :param db_session: :return:
[ "Counts", "children", "of", "resource", "node" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L369-L379
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/resource_tree_postgres.py
ResourceTreeServicePostgreSQL.check_node_position
def check_node_position( cls, parent_id, position, on_same_branch, db_session=None, *args, **kwargs ): """ Checks if node position for given parent is valid, raises exception if this is not the case :param parent_id: :param position: :param on_same_branch: indicates that we are checking same branch :param db_session: :return: """ db_session = get_db_session(db_session) if not position or position < 1: raise ZigguratResourceOutOfBoundaryException( "Position is lower than {}", value=1 ) item_count = cls.count_children(parent_id, db_session=db_session) max_value = item_count if on_same_branch else item_count + 1 if position > max_value: raise ZigguratResourceOutOfBoundaryException( "Maximum resource ordering is {}", value=max_value )
python
def check_node_position( cls, parent_id, position, on_same_branch, db_session=None, *args, **kwargs ): """ Checks if node position for given parent is valid, raises exception if this is not the case :param parent_id: :param position: :param on_same_branch: indicates that we are checking same branch :param db_session: :return: """ db_session = get_db_session(db_session) if not position or position < 1: raise ZigguratResourceOutOfBoundaryException( "Position is lower than {}", value=1 ) item_count = cls.count_children(parent_id, db_session=db_session) max_value = item_count if on_same_branch else item_count + 1 if position > max_value: raise ZigguratResourceOutOfBoundaryException( "Maximum resource ordering is {}", value=max_value )
[ "def", "check_node_position", "(", "cls", ",", "parent_id", ",", "position", ",", "on_same_branch", ",", "db_session", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "if", "not", "position", "or", "position", "<", "1", ":", "raise", "ZigguratResourceOutOfBoundaryException", "(", "\"Position is lower than {}\"", ",", "value", "=", "1", ")", "item_count", "=", "cls", ".", "count_children", "(", "parent_id", ",", "db_session", "=", "db_session", ")", "max_value", "=", "item_count", "if", "on_same_branch", "else", "item_count", "+", "1", "if", "position", ">", "max_value", ":", "raise", "ZigguratResourceOutOfBoundaryException", "(", "\"Maximum resource ordering is {}\"", ",", "value", "=", "max_value", ")" ]
Checks if node position for given parent is valid, raises exception if this is not the case :param parent_id: :param position: :param on_same_branch: indicates that we are checking same branch :param db_session: :return:
[ "Checks", "if", "node", "position", "for", "given", "parent", "is", "valid", "raises", "exception", "if", "this", "is", "not", "the", "case" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/resource_tree_postgres.py#L382-L405
train
lapets/egcd
egcd/egcd.py
egcd
def egcd(b, n): ''' Given two integers (b, n), returns (gcd(b, n), a, m) such that a*b + n*m = gcd(b, n). Adapted from several sources: https://brilliant.org/wiki/extended-euclidean-algorithm/ https://rosettacode.org/wiki/Modular_inverse https://en.wikibooks.org/wiki/Algorithm_Implementation/Mathematics/Extended_Euclidean_algorithm https://en.wikipedia.org/wiki/Euclidean_algorithm >>> egcd(1, 1) (1, 0, 1) >>> egcd(12, 8) (4, 1, -1) >>> egcd(23894798501898, 23948178468116) (2, 2437250447493, -2431817869532) >>> egcd(pow(2, 50), pow(3, 50)) (1, -260414429242905345185687, 408415383037561) ''' (x0, x1, y0, y1) = (1, 0, 0, 1) while n != 0: (q, b, n) = (b // n, n, b % n) (x0, x1) = (x1, x0 - q * x1) (y0, y1) = (y1, y0 - q * y1) return (b, x0, y0)
python
def egcd(b, n): ''' Given two integers (b, n), returns (gcd(b, n), a, m) such that a*b + n*m = gcd(b, n). Adapted from several sources: https://brilliant.org/wiki/extended-euclidean-algorithm/ https://rosettacode.org/wiki/Modular_inverse https://en.wikibooks.org/wiki/Algorithm_Implementation/Mathematics/Extended_Euclidean_algorithm https://en.wikipedia.org/wiki/Euclidean_algorithm >>> egcd(1, 1) (1, 0, 1) >>> egcd(12, 8) (4, 1, -1) >>> egcd(23894798501898, 23948178468116) (2, 2437250447493, -2431817869532) >>> egcd(pow(2, 50), pow(3, 50)) (1, -260414429242905345185687, 408415383037561) ''' (x0, x1, y0, y1) = (1, 0, 0, 1) while n != 0: (q, b, n) = (b // n, n, b % n) (x0, x1) = (x1, x0 - q * x1) (y0, y1) = (y1, y0 - q * y1) return (b, x0, y0)
[ "def", "egcd", "(", "b", ",", "n", ")", ":", "(", "x0", ",", "x1", ",", "y0", ",", "y1", ")", "=", "(", "1", ",", "0", ",", "0", ",", "1", ")", "while", "n", "!=", "0", ":", "(", "q", ",", "b", ",", "n", ")", "=", "(", "b", "//", "n", ",", "n", ",", "b", "%", "n", ")", "(", "x0", ",", "x1", ")", "=", "(", "x1", ",", "x0", "-", "q", "*", "x1", ")", "(", "y0", ",", "y1", ")", "=", "(", "y1", ",", "y0", "-", "q", "*", "y1", ")", "return", "(", "b", ",", "x0", ",", "y0", ")" ]
Given two integers (b, n), returns (gcd(b, n), a, m) such that a*b + n*m = gcd(b, n). Adapted from several sources: https://brilliant.org/wiki/extended-euclidean-algorithm/ https://rosettacode.org/wiki/Modular_inverse https://en.wikibooks.org/wiki/Algorithm_Implementation/Mathematics/Extended_Euclidean_algorithm https://en.wikipedia.org/wiki/Euclidean_algorithm >>> egcd(1, 1) (1, 0, 1) >>> egcd(12, 8) (4, 1, -1) >>> egcd(23894798501898, 23948178468116) (2, 2437250447493, -2431817869532) >>> egcd(pow(2, 50), pow(3, 50)) (1, -260414429242905345185687, 408415383037561)
[ "Given", "two", "integers", "(", "b", "n", ")", "returns", "(", "gcd", "(", "b", "n", ")", "a", "m", ")", "such", "that", "a", "*", "b", "+", "n", "*", "m", "=", "gcd", "(", "b", "n", ")", ".", "Adapted", "from", "several", "sources", ":", "https", ":", "//", "brilliant", ".", "org", "/", "wiki", "/", "extended", "-", "euclidean", "-", "algorithm", "/", "https", ":", "//", "rosettacode", ".", "org", "/", "wiki", "/", "Modular_inverse", "https", ":", "//", "en", ".", "wikibooks", ".", "org", "/", "wiki", "/", "Algorithm_Implementation", "/", "Mathematics", "/", "Extended_Euclidean_algorithm", "https", ":", "//", "en", ".", "wikipedia", ".", "org", "/", "wiki", "/", "Euclidean_algorithm", ">>>", "egcd", "(", "1", "1", ")", "(", "1", "0", "1", ")", ">>>", "egcd", "(", "12", "8", ")", "(", "4", "1", "-", "1", ")", ">>>", "egcd", "(", "23894798501898", "23948178468116", ")", "(", "2", "2437250447493", "-", "2431817869532", ")", ">>>", "egcd", "(", "pow", "(", "2", "50", ")", "pow", "(", "3", "50", "))", "(", "1", "-", "260414429242905345185687", "408415383037561", ")" ]
44fbab4a8eb04ad198526b4f14ddce6dff2bae77
https://github.com/lapets/egcd/blob/44fbab4a8eb04ad198526b4f14ddce6dff2bae77/egcd/egcd.py#L16-L42
train
jucacrispim/pylint-mongoengine
pylint_mongoengine/__init__.py
register
def register(linter): """Add the needed transformations and supressions. """ linter.register_checker(MongoEngineChecker(linter)) add_transform('mongoengine') add_transform('mongomotor') suppress_qs_decorator_messages(linter) suppress_fields_attrs_messages(linter)
python
def register(linter): """Add the needed transformations and supressions. """ linter.register_checker(MongoEngineChecker(linter)) add_transform('mongoengine') add_transform('mongomotor') suppress_qs_decorator_messages(linter) suppress_fields_attrs_messages(linter)
[ "def", "register", "(", "linter", ")", ":", "linter", ".", "register_checker", "(", "MongoEngineChecker", "(", "linter", ")", ")", "add_transform", "(", "'mongoengine'", ")", "add_transform", "(", "'mongomotor'", ")", "suppress_qs_decorator_messages", "(", "linter", ")", "suppress_fields_attrs_messages", "(", "linter", ")" ]
Add the needed transformations and supressions.
[ "Add", "the", "needed", "transformations", "and", "supressions", "." ]
b873653d1224a5748f75dd507f492f8c60d95ce3
https://github.com/jucacrispim/pylint-mongoengine/blob/b873653d1224a5748f75dd507f492f8c60d95ce3/pylint_mongoengine/__init__.py#L25-L33
train
aguinane/nem-reader
nemreader/outputs.py
output_as_csv
def output_as_csv(file_name, nmi=None, output_file=None): """ Transpose all channels and output a csv that is easier to read and do charting on :param file_name: The NEM file to process :param nmi: Which NMI to output if more than one :param output_file: Specify different output location :returns: The file that was created """ m = read_nem_file(file_name) if nmi is None: nmi = list(m.readings.keys())[0] # Use first NMI channels = list(m.transactions[nmi].keys()) num_records = len(m.readings[nmi][channels[0]]) last_date = m.readings[nmi][channels[0]][-1].t_end if output_file is None: output_file = '{}_{}_transposed.csv'.format( nmi, last_date.strftime('%Y%m%d')) with open(output_file, 'w', newline='') as csvfile: cwriter = csv.writer( csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) heading_list = ['period_start', 'period_end'] for channel in channels: heading_list.append(channel) heading_list.append('quality_method') cwriter.writerow(heading_list) for i in range(0, num_records): t_start = m.readings[nmi][channels[0]][i].t_start t_end = m.readings[nmi][channels[0]][i].t_end quality_method = m.readings[nmi][channels[0]][i].quality_method row_list = [t_start, t_end] for ch in channels: val = m.readings[nmi][ch][i].read_value row_list.append(val) row_list.append(quality_method) cwriter.writerow(row_list) return output_file
python
def output_as_csv(file_name, nmi=None, output_file=None): """ Transpose all channels and output a csv that is easier to read and do charting on :param file_name: The NEM file to process :param nmi: Which NMI to output if more than one :param output_file: Specify different output location :returns: The file that was created """ m = read_nem_file(file_name) if nmi is None: nmi = list(m.readings.keys())[0] # Use first NMI channels = list(m.transactions[nmi].keys()) num_records = len(m.readings[nmi][channels[0]]) last_date = m.readings[nmi][channels[0]][-1].t_end if output_file is None: output_file = '{}_{}_transposed.csv'.format( nmi, last_date.strftime('%Y%m%d')) with open(output_file, 'w', newline='') as csvfile: cwriter = csv.writer( csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) heading_list = ['period_start', 'period_end'] for channel in channels: heading_list.append(channel) heading_list.append('quality_method') cwriter.writerow(heading_list) for i in range(0, num_records): t_start = m.readings[nmi][channels[0]][i].t_start t_end = m.readings[nmi][channels[0]][i].t_end quality_method = m.readings[nmi][channels[0]][i].quality_method row_list = [t_start, t_end] for ch in channels: val = m.readings[nmi][ch][i].read_value row_list.append(val) row_list.append(quality_method) cwriter.writerow(row_list) return output_file
[ "def", "output_as_csv", "(", "file_name", ",", "nmi", "=", "None", ",", "output_file", "=", "None", ")", ":", "m", "=", "read_nem_file", "(", "file_name", ")", "if", "nmi", "is", "None", ":", "nmi", "=", "list", "(", "m", ".", "readings", ".", "keys", "(", ")", ")", "[", "0", "]", "# Use first NMI", "channels", "=", "list", "(", "m", ".", "transactions", "[", "nmi", "]", ".", "keys", "(", ")", ")", "num_records", "=", "len", "(", "m", ".", "readings", "[", "nmi", "]", "[", "channels", "[", "0", "]", "]", ")", "last_date", "=", "m", ".", "readings", "[", "nmi", "]", "[", "channels", "[", "0", "]", "]", "[", "-", "1", "]", ".", "t_end", "if", "output_file", "is", "None", ":", "output_file", "=", "'{}_{}_transposed.csv'", ".", "format", "(", "nmi", ",", "last_date", ".", "strftime", "(", "'%Y%m%d'", ")", ")", "with", "open", "(", "output_file", ",", "'w'", ",", "newline", "=", "''", ")", "as", "csvfile", ":", "cwriter", "=", "csv", ".", "writer", "(", "csvfile", ",", "delimiter", "=", "','", ",", "quotechar", "=", "'\"'", ",", "quoting", "=", "csv", ".", "QUOTE_MINIMAL", ")", "heading_list", "=", "[", "'period_start'", ",", "'period_end'", "]", "for", "channel", "in", "channels", ":", "heading_list", ".", "append", "(", "channel", ")", "heading_list", ".", "append", "(", "'quality_method'", ")", "cwriter", ".", "writerow", "(", "heading_list", ")", "for", "i", "in", "range", "(", "0", ",", "num_records", ")", ":", "t_start", "=", "m", ".", "readings", "[", "nmi", "]", "[", "channels", "[", "0", "]", "]", "[", "i", "]", ".", "t_start", "t_end", "=", "m", ".", "readings", "[", "nmi", "]", "[", "channels", "[", "0", "]", "]", "[", "i", "]", ".", "t_end", "quality_method", "=", "m", ".", "readings", "[", "nmi", "]", "[", "channels", "[", "0", "]", "]", "[", "i", "]", ".", "quality_method", "row_list", "=", "[", "t_start", ",", "t_end", "]", "for", "ch", "in", "channels", ":", "val", "=", "m", ".", "readings", "[", "nmi", "]", "[", "ch", "]", "[", "i", "]", ".", "read_value", "row_list", ".", "append", "(", "val", ")", "row_list", ".", "append", "(", "quality_method", ")", "cwriter", ".", "writerow", "(", "row_list", ")", "return", "output_file" ]
Transpose all channels and output a csv that is easier to read and do charting on :param file_name: The NEM file to process :param nmi: Which NMI to output if more than one :param output_file: Specify different output location :returns: The file that was created
[ "Transpose", "all", "channels", "and", "output", "a", "csv", "that", "is", "easier", "to", "read", "and", "do", "charting", "on" ]
5405a5cba4bb8ebdad05c28455d12bb34a6d3ce5
https://github.com/aguinane/nem-reader/blob/5405a5cba4bb8ebdad05c28455d12bb34a6d3ce5/nemreader/outputs.py#L11-L51
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/group.py
GroupService.get
def get(cls, group_id, db_session=None): """ Fetch row using primary key - will use existing object in session if already present :param group_id: :param db_session: :return: """ db_session = get_db_session(db_session) return db_session.query(cls.model).get(group_id)
python
def get(cls, group_id, db_session=None): """ Fetch row using primary key - will use existing object in session if already present :param group_id: :param db_session: :return: """ db_session = get_db_session(db_session) return db_session.query(cls.model).get(group_id)
[ "def", "get", "(", "cls", ",", "group_id", ",", "db_session", "=", "None", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "return", "db_session", ".", "query", "(", "cls", ".", "model", ")", ".", "get", "(", "group_id", ")" ]
Fetch row using primary key - will use existing object in session if already present :param group_id: :param db_session: :return:
[ "Fetch", "row", "using", "primary", "key", "-", "will", "use", "existing", "object", "in", "session", "if", "already", "present" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/group.py#L19-L29
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/group.py
GroupService.by_group_name
def by_group_name(cls, group_name, db_session=None): """ fetch group by name :param group_name: :param db_session: :return: """ db_session = get_db_session(db_session) query = db_session.query(cls.model).filter(cls.model.group_name == group_name) return query.first()
python
def by_group_name(cls, group_name, db_session=None): """ fetch group by name :param group_name: :param db_session: :return: """ db_session = get_db_session(db_session) query = db_session.query(cls.model).filter(cls.model.group_name == group_name) return query.first()
[ "def", "by_group_name", "(", "cls", ",", "group_name", ",", "db_session", "=", "None", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ")", "query", "=", "db_session", ".", "query", "(", "cls", ".", "model", ")", ".", "filter", "(", "cls", ".", "model", ".", "group_name", "==", "group_name", ")", "return", "query", ".", "first", "(", ")" ]
fetch group by name :param group_name: :param db_session: :return:
[ "fetch", "group", "by", "name" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/group.py#L32-L42
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/group.py
GroupService.get_user_paginator
def get_user_paginator( cls, instance, page=1, item_count=None, items_per_page=50, user_ids=None, GET_params=None, ): """ returns paginator over users belonging to the group :param instance: :param page: :param item_count: :param items_per_page: :param user_ids: :param GET_params: :return: """ if not GET_params: GET_params = {} GET_params.pop("page", None) query = instance.users_dynamic if user_ids: query = query.filter(cls.models_proxy.UserGroup.user_id.in_(user_ids)) return SqlalchemyOrmPage( query, page=page, item_count=item_count, items_per_page=items_per_page, **GET_params )
python
def get_user_paginator( cls, instance, page=1, item_count=None, items_per_page=50, user_ids=None, GET_params=None, ): """ returns paginator over users belonging to the group :param instance: :param page: :param item_count: :param items_per_page: :param user_ids: :param GET_params: :return: """ if not GET_params: GET_params = {} GET_params.pop("page", None) query = instance.users_dynamic if user_ids: query = query.filter(cls.models_proxy.UserGroup.user_id.in_(user_ids)) return SqlalchemyOrmPage( query, page=page, item_count=item_count, items_per_page=items_per_page, **GET_params )
[ "def", "get_user_paginator", "(", "cls", ",", "instance", ",", "page", "=", "1", ",", "item_count", "=", "None", ",", "items_per_page", "=", "50", ",", "user_ids", "=", "None", ",", "GET_params", "=", "None", ",", ")", ":", "if", "not", "GET_params", ":", "GET_params", "=", "{", "}", "GET_params", ".", "pop", "(", "\"page\"", ",", "None", ")", "query", "=", "instance", ".", "users_dynamic", "if", "user_ids", ":", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "UserGroup", ".", "user_id", ".", "in_", "(", "user_ids", ")", ")", "return", "SqlalchemyOrmPage", "(", "query", ",", "page", "=", "page", ",", "item_count", "=", "item_count", ",", "items_per_page", "=", "items_per_page", ",", "*", "*", "GET_params", ")" ]
returns paginator over users belonging to the group :param instance: :param page: :param item_count: :param items_per_page: :param user_ids: :param GET_params: :return:
[ "returns", "paginator", "over", "users", "belonging", "to", "the", "group" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/group.py#L45-L77
train
ergo/ziggurat_foundations
ziggurat_foundations/models/services/group.py
GroupService.resources_with_possible_perms
def resources_with_possible_perms( cls, instance, perm_names=None, resource_ids=None, resource_types=None, db_session=None, ): """ returns list of permissions and resources for this group, resource_ids restricts the search to specific resources :param instance: :param perm_names: :param resource_ids: :param resource_types: :param db_session: :return: """ db_session = get_db_session(db_session, instance) query = db_session.query( cls.models_proxy.GroupResourcePermission.perm_name, cls.models_proxy.Group, cls.models_proxy.Resource, ) query = query.filter( cls.models_proxy.Resource.resource_id == cls.models_proxy.GroupResourcePermission.resource_id ) query = query.filter( cls.models_proxy.Group.id == cls.models_proxy.GroupResourcePermission.group_id ) if resource_ids: query = query.filter( cls.models_proxy.GroupResourcePermission.resource_id.in_(resource_ids) ) if resource_types: query = query.filter( cls.models_proxy.Resource.resource_type.in_(resource_types) ) if perm_names not in ([ANY_PERMISSION], ANY_PERMISSION) and perm_names: query = query.filter( cls.models_proxy.GroupResourcePermission.perm_name.in_(perm_names) ) query = query.filter( cls.models_proxy.GroupResourcePermission.group_id == instance.id ) perms = [ PermissionTuple( None, row.perm_name, "group", instance, row.Resource, False, True ) for row in query ] for resource in instance.resources: perms.append( PermissionTuple( None, ALL_PERMISSIONS, "group", instance, resource, True, True ) ) return perms
python
def resources_with_possible_perms( cls, instance, perm_names=None, resource_ids=None, resource_types=None, db_session=None, ): """ returns list of permissions and resources for this group, resource_ids restricts the search to specific resources :param instance: :param perm_names: :param resource_ids: :param resource_types: :param db_session: :return: """ db_session = get_db_session(db_session, instance) query = db_session.query( cls.models_proxy.GroupResourcePermission.perm_name, cls.models_proxy.Group, cls.models_proxy.Resource, ) query = query.filter( cls.models_proxy.Resource.resource_id == cls.models_proxy.GroupResourcePermission.resource_id ) query = query.filter( cls.models_proxy.Group.id == cls.models_proxy.GroupResourcePermission.group_id ) if resource_ids: query = query.filter( cls.models_proxy.GroupResourcePermission.resource_id.in_(resource_ids) ) if resource_types: query = query.filter( cls.models_proxy.Resource.resource_type.in_(resource_types) ) if perm_names not in ([ANY_PERMISSION], ANY_PERMISSION) and perm_names: query = query.filter( cls.models_proxy.GroupResourcePermission.perm_name.in_(perm_names) ) query = query.filter( cls.models_proxy.GroupResourcePermission.group_id == instance.id ) perms = [ PermissionTuple( None, row.perm_name, "group", instance, row.Resource, False, True ) for row in query ] for resource in instance.resources: perms.append( PermissionTuple( None, ALL_PERMISSIONS, "group", instance, resource, True, True ) ) return perms
[ "def", "resources_with_possible_perms", "(", "cls", ",", "instance", ",", "perm_names", "=", "None", ",", "resource_ids", "=", "None", ",", "resource_types", "=", "None", ",", "db_session", "=", "None", ",", ")", ":", "db_session", "=", "get_db_session", "(", "db_session", ",", "instance", ")", "query", "=", "db_session", ".", "query", "(", "cls", ".", "models_proxy", ".", "GroupResourcePermission", ".", "perm_name", ",", "cls", ".", "models_proxy", ".", "Group", ",", "cls", ".", "models_proxy", ".", "Resource", ",", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "Resource", ".", "resource_id", "==", "cls", ".", "models_proxy", ".", "GroupResourcePermission", ".", "resource_id", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "Group", ".", "id", "==", "cls", ".", "models_proxy", ".", "GroupResourcePermission", ".", "group_id", ")", "if", "resource_ids", ":", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "GroupResourcePermission", ".", "resource_id", ".", "in_", "(", "resource_ids", ")", ")", "if", "resource_types", ":", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "Resource", ".", "resource_type", ".", "in_", "(", "resource_types", ")", ")", "if", "perm_names", "not", "in", "(", "[", "ANY_PERMISSION", "]", ",", "ANY_PERMISSION", ")", "and", "perm_names", ":", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "GroupResourcePermission", ".", "perm_name", ".", "in_", "(", "perm_names", ")", ")", "query", "=", "query", ".", "filter", "(", "cls", ".", "models_proxy", ".", "GroupResourcePermission", ".", "group_id", "==", "instance", ".", "id", ")", "perms", "=", "[", "PermissionTuple", "(", "None", ",", "row", ".", "perm_name", ",", "\"group\"", ",", "instance", ",", "row", ".", "Resource", ",", "False", ",", "True", ")", "for", "row", "in", "query", "]", "for", "resource", "in", "instance", ".", "resources", ":", "perms", ".", "append", "(", "PermissionTuple", "(", "None", ",", "ALL_PERMISSIONS", ",", "\"group\"", ",", "instance", ",", "resource", ",", "True", ",", "True", ")", ")", "return", "perms" ]
returns list of permissions and resources for this group, resource_ids restricts the search to specific resources :param instance: :param perm_names: :param resource_ids: :param resource_types: :param db_session: :return:
[ "returns", "list", "of", "permissions", "and", "resources", "for", "this", "group", "resource_ids", "restricts", "the", "search", "to", "specific", "resources" ]
9eeec894d08e8d7defa60ddc04b63f69cd4cbeba
https://github.com/ergo/ziggurat_foundations/blob/9eeec894d08e8d7defa60ddc04b63f69cd4cbeba/ziggurat_foundations/models/services/group.py#L80-L144
train
zeromake/aiosqlite3
aiosqlite3/pool.py
create_pool
def create_pool( database, minsize=1, maxsize=10, echo=False, loop=None, **kwargs ): """ 创建支持上下文管理的pool """ coro = _create_pool( database=database, minsize=minsize, maxsize=maxsize, echo=echo, loop=loop, **kwargs ) return _PoolContextManager(coro)
python
def create_pool( database, minsize=1, maxsize=10, echo=False, loop=None, **kwargs ): """ 创建支持上下文管理的pool """ coro = _create_pool( database=database, minsize=minsize, maxsize=maxsize, echo=echo, loop=loop, **kwargs ) return _PoolContextManager(coro)
[ "def", "create_pool", "(", "database", ",", "minsize", "=", "1", ",", "maxsize", "=", "10", ",", "echo", "=", "False", ",", "loop", "=", "None", ",", "*", "*", "kwargs", ")", ":", "coro", "=", "_create_pool", "(", "database", "=", "database", ",", "minsize", "=", "minsize", ",", "maxsize", "=", "maxsize", ",", "echo", "=", "echo", ",", "loop", "=", "loop", ",", "*", "*", "kwargs", ")", "return", "_PoolContextManager", "(", "coro", ")" ]
创建支持上下文管理的pool
[ "创建支持上下文管理的pool" ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/pool.py#L14-L33
train
zeromake/aiosqlite3
aiosqlite3/pool.py
Pool.wait_closed
def wait_closed(self): """ Wait for closing all pool's connections. """ if self._closed: return if not self._closing: raise RuntimeError( ".wait_closed() should be called " "after .close()" ) while self._free: conn = self._free.popleft() if not conn.closed: yield from conn.close() else: # pragma: no cover pass with (yield from self._cond): while self.size > self.freesize: yield from self._cond.wait() self._used.clear() self._closed = True
python
def wait_closed(self): """ Wait for closing all pool's connections. """ if self._closed: return if not self._closing: raise RuntimeError( ".wait_closed() should be called " "after .close()" ) while self._free: conn = self._free.popleft() if not conn.closed: yield from conn.close() else: # pragma: no cover pass with (yield from self._cond): while self.size > self.freesize: yield from self._cond.wait() self._used.clear() self._closed = True
[ "def", "wait_closed", "(", "self", ")", ":", "if", "self", ".", "_closed", ":", "return", "if", "not", "self", ".", "_closing", ":", "raise", "RuntimeError", "(", "\".wait_closed() should be called \"", "\"after .close()\"", ")", "while", "self", ".", "_free", ":", "conn", "=", "self", ".", "_free", ".", "popleft", "(", ")", "if", "not", "conn", ".", "closed", ":", "yield", "from", "conn", ".", "close", "(", ")", "else", ":", "# pragma: no cover", "pass", "with", "(", "yield", "from", "self", ".", "_cond", ")", ":", "while", "self", ".", "size", ">", "self", ".", "freesize", ":", "yield", "from", "self", ".", "_cond", ".", "wait", "(", ")", "self", ".", "_used", ".", "clear", "(", ")", "self", ".", "_closed", "=", "True" ]
Wait for closing all pool's connections.
[ "Wait", "for", "closing", "all", "pool", "s", "connections", "." ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/pool.py#L169-L192
train
zeromake/aiosqlite3
aiosqlite3/pool.py
Pool.sync_close
def sync_close(self): """ 同步关闭 """ if self._closed: return while self._free: conn = self._free.popleft() if not conn.closed: # pragma: no cover conn.sync_close() for conn in self._used: if not conn.closed: # pragma: no cover conn.sync_close() self._terminated.add(conn) self._used.clear() self._closed = True
python
def sync_close(self): """ 同步关闭 """ if self._closed: return while self._free: conn = self._free.popleft() if not conn.closed: # pragma: no cover conn.sync_close() for conn in self._used: if not conn.closed: # pragma: no cover conn.sync_close() self._terminated.add(conn) self._used.clear() self._closed = True
[ "def", "sync_close", "(", "self", ")", ":", "if", "self", ".", "_closed", ":", "return", "while", "self", ".", "_free", ":", "conn", "=", "self", ".", "_free", ".", "popleft", "(", ")", "if", "not", "conn", ".", "closed", ":", "# pragma: no cover", "conn", ".", "sync_close", "(", ")", "for", "conn", "in", "self", ".", "_used", ":", "if", "not", "conn", ".", "closed", ":", "# pragma: no cover", "conn", ".", "sync_close", "(", ")", "self", ".", "_terminated", ".", "add", "(", "conn", ")", "self", ".", "_used", ".", "clear", "(", ")", "self", ".", "_closed", "=", "True" ]
同步关闭
[ "同步关闭" ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/pool.py#L194-L211
train
zeromake/aiosqlite3
aiosqlite3/pool.py
Pool._fill_free_pool
def _fill_free_pool(self, override_min): """ iterate over free connections and remove timeouted ones """ while self.size < self.minsize: self._acquiring += 1 try: conn = yield from connect( database=self._database, echo=self._echo, loop=self._loop, **self._conn_kwargs ) self._free.append(conn) self._cond.notify() finally: self._acquiring -= 1 if self._free: return if override_min and self.size < self.maxsize: self._acquiring += 1 try: conn = yield from connect( database=self._database, echo=self._echo, loop=self._loop, **self._conn_kwargs ) self._free.append(conn) self._cond.notify() finally: self._acquiring -= 1
python
def _fill_free_pool(self, override_min): """ iterate over free connections and remove timeouted ones """ while self.size < self.minsize: self._acquiring += 1 try: conn = yield from connect( database=self._database, echo=self._echo, loop=self._loop, **self._conn_kwargs ) self._free.append(conn) self._cond.notify() finally: self._acquiring -= 1 if self._free: return if override_min and self.size < self.maxsize: self._acquiring += 1 try: conn = yield from connect( database=self._database, echo=self._echo, loop=self._loop, **self._conn_kwargs ) self._free.append(conn) self._cond.notify() finally: self._acquiring -= 1
[ "def", "_fill_free_pool", "(", "self", ",", "override_min", ")", ":", "while", "self", ".", "size", "<", "self", ".", "minsize", ":", "self", ".", "_acquiring", "+=", "1", "try", ":", "conn", "=", "yield", "from", "connect", "(", "database", "=", "self", ".", "_database", ",", "echo", "=", "self", ".", "_echo", ",", "loop", "=", "self", ".", "_loop", ",", "*", "*", "self", ".", "_conn_kwargs", ")", "self", ".", "_free", ".", "append", "(", "conn", ")", "self", ".", "_cond", ".", "notify", "(", ")", "finally", ":", "self", ".", "_acquiring", "-=", "1", "if", "self", ".", "_free", ":", "return", "if", "override_min", "and", "self", ".", "size", "<", "self", ".", "maxsize", ":", "self", ".", "_acquiring", "+=", "1", "try", ":", "conn", "=", "yield", "from", "connect", "(", "database", "=", "self", ".", "_database", ",", "echo", "=", "self", ".", "_echo", ",", "loop", "=", "self", ".", "_loop", ",", "*", "*", "self", ".", "_conn_kwargs", ")", "self", ".", "_free", ".", "append", "(", "conn", ")", "self", ".", "_cond", ".", "notify", "(", ")", "finally", ":", "self", ".", "_acquiring", "-=", "1" ]
iterate over free connections and remove timeouted ones
[ "iterate", "over", "free", "connections", "and", "remove", "timeouted", "ones" ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/pool.py#L242-L274
train
stepank/pyws
src/pyws/functions/managers.py
FixedFunctionManager.add_function
def add_function(self, function): """ Adds the function to the list of registered functions. """ function = self.build_function(function) if function.name in self.functions: raise FunctionAlreadyRegistered(function.name) self.functions[function.name] = function
python
def add_function(self, function): """ Adds the function to the list of registered functions. """ function = self.build_function(function) if function.name in self.functions: raise FunctionAlreadyRegistered(function.name) self.functions[function.name] = function
[ "def", "add_function", "(", "self", ",", "function", ")", ":", "function", "=", "self", ".", "build_function", "(", "function", ")", "if", "function", ".", "name", "in", "self", ".", "functions", ":", "raise", "FunctionAlreadyRegistered", "(", "function", ".", "name", ")", "self", ".", "functions", "[", "function", ".", "name", "]", "=", "function" ]
Adds the function to the list of registered functions.
[ "Adds", "the", "function", "to", "the", "list", "of", "registered", "functions", "." ]
ff39133aabeb56bbb08d66286ac0cc8731eda7dd
https://github.com/stepank/pyws/blob/ff39133aabeb56bbb08d66286ac0cc8731eda7dd/src/pyws/functions/managers.py#L44-L51
train
stepank/pyws
src/pyws/functions/managers.py
FixedFunctionManager.get_one
def get_one(self, context, name): """ Returns a function if it is registered, the context is ignored. """ try: return self.functions[name] except KeyError: raise FunctionNotFound(name)
python
def get_one(self, context, name): """ Returns a function if it is registered, the context is ignored. """ try: return self.functions[name] except KeyError: raise FunctionNotFound(name)
[ "def", "get_one", "(", "self", ",", "context", ",", "name", ")", ":", "try", ":", "return", "self", ".", "functions", "[", "name", "]", "except", "KeyError", ":", "raise", "FunctionNotFound", "(", "name", ")" ]
Returns a function if it is registered, the context is ignored.
[ "Returns", "a", "function", "if", "it", "is", "registered", "the", "context", "is", "ignored", "." ]
ff39133aabeb56bbb08d66286ac0cc8731eda7dd
https://github.com/stepank/pyws/blob/ff39133aabeb56bbb08d66286ac0cc8731eda7dd/src/pyws/functions/managers.py#L53-L60
train
opennode/waldur-core
waldur_core/core/monkeypatch.py
subfield_get
def subfield_get(self, obj, type=None): """ Verbatim copy from: https://github.com/django/django/blob/1.9.13/django/db/models/fields/subclassing.py#L38 """ if obj is None: return self return obj.__dict__[self.field.name]
python
def subfield_get(self, obj, type=None): """ Verbatim copy from: https://github.com/django/django/blob/1.9.13/django/db/models/fields/subclassing.py#L38 """ if obj is None: return self return obj.__dict__[self.field.name]
[ "def", "subfield_get", "(", "self", ",", "obj", ",", "type", "=", "None", ")", ":", "if", "obj", "is", "None", ":", "return", "self", "return", "obj", ".", "__dict__", "[", "self", ".", "field", ".", "name", "]" ]
Verbatim copy from: https://github.com/django/django/blob/1.9.13/django/db/models/fields/subclassing.py#L38
[ "Verbatim", "copy", "from", ":", "https", ":", "//", "github", ".", "com", "/", "django", "/", "django", "/", "blob", "/", "1", ".", "9", ".", "13", "/", "django", "/", "db", "/", "models", "/", "fields", "/", "subclassing", ".", "py#L38" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/monkeypatch.py#L14-L21
train
opennode/waldur-core
waldur_core/core/managers.py
GenericKeyMixin._preprocess_kwargs
def _preprocess_kwargs(self, initial_kwargs): """ Replace generic key related attribute with filters by object_id and content_type fields """ kwargs = initial_kwargs.copy() generic_key_related_kwargs = self._get_generic_key_related_kwargs(initial_kwargs) for key, value in generic_key_related_kwargs.items(): # delete old kwarg that was related to generic key del kwargs[key] try: suffix = key.split('__')[1] except IndexError: suffix = None # add new kwargs that related to object_id and content_type fields new_kwargs = self._get_filter_object_id_and_content_type_filter_kwargs(value, suffix) kwargs.update(new_kwargs) return kwargs
python
def _preprocess_kwargs(self, initial_kwargs): """ Replace generic key related attribute with filters by object_id and content_type fields """ kwargs = initial_kwargs.copy() generic_key_related_kwargs = self._get_generic_key_related_kwargs(initial_kwargs) for key, value in generic_key_related_kwargs.items(): # delete old kwarg that was related to generic key del kwargs[key] try: suffix = key.split('__')[1] except IndexError: suffix = None # add new kwargs that related to object_id and content_type fields new_kwargs = self._get_filter_object_id_and_content_type_filter_kwargs(value, suffix) kwargs.update(new_kwargs) return kwargs
[ "def", "_preprocess_kwargs", "(", "self", ",", "initial_kwargs", ")", ":", "kwargs", "=", "initial_kwargs", ".", "copy", "(", ")", "generic_key_related_kwargs", "=", "self", ".", "_get_generic_key_related_kwargs", "(", "initial_kwargs", ")", "for", "key", ",", "value", "in", "generic_key_related_kwargs", ".", "items", "(", ")", ":", "# delete old kwarg that was related to generic key", "del", "kwargs", "[", "key", "]", "try", ":", "suffix", "=", "key", ".", "split", "(", "'__'", ")", "[", "1", "]", "except", "IndexError", ":", "suffix", "=", "None", "# add new kwargs that related to object_id and content_type fields", "new_kwargs", "=", "self", ".", "_get_filter_object_id_and_content_type_filter_kwargs", "(", "value", ",", "suffix", ")", "kwargs", ".", "update", "(", "new_kwargs", ")", "return", "kwargs" ]
Replace generic key related attribute with filters by object_id and content_type fields
[ "Replace", "generic", "key", "related", "attribute", "with", "filters", "by", "object_id", "and", "content_type", "fields" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/managers.py#L29-L44
train
OpenDataScienceLab/skdata
skdata/cleaning.py
categorize
def categorize( data, col_name: str = None, new_col_name: str = None, categories: dict = None, max_categories: float = 0.15 ): """ :param data: :param col_name: :param new_col_name: :param categories: :param max_categories: max proportion threshold of categories :return: new categories :rtype dict: """ _categories = {} if col_name is None: if categories is not None: raise Exception( 'col_name is None when categories was defined.' ) # create a list of cols with all object columns cols = [ k for k in data.keys() if data[k].dtype == 'object' and (data[k].unique() / data[k].count()) <= max_categories ] else: # create a list with col_name if new_col_name is not None: data[new_col_name] = data[col_name] col_name = new_col_name cols = [col_name] for c in cols: if categories is not None: # assert all keys is a number assert all(type(k) in (int, float) for k in categories.keys()) # replace values using given categories dict data[c].replace(categories, inplace=True) # change column to categorical type data[c] = data[c].astype('category') # update categories information _categories.update({c: categories}) else: # change column to categorical type data[c] = data[c].astype('category') # change column to categorical type _categories.update({ c: dict(enumerate( data[c].cat.categories, )) }) return _categories
python
def categorize( data, col_name: str = None, new_col_name: str = None, categories: dict = None, max_categories: float = 0.15 ): """ :param data: :param col_name: :param new_col_name: :param categories: :param max_categories: max proportion threshold of categories :return: new categories :rtype dict: """ _categories = {} if col_name is None: if categories is not None: raise Exception( 'col_name is None when categories was defined.' ) # create a list of cols with all object columns cols = [ k for k in data.keys() if data[k].dtype == 'object' and (data[k].unique() / data[k].count()) <= max_categories ] else: # create a list with col_name if new_col_name is not None: data[new_col_name] = data[col_name] col_name = new_col_name cols = [col_name] for c in cols: if categories is not None: # assert all keys is a number assert all(type(k) in (int, float) for k in categories.keys()) # replace values using given categories dict data[c].replace(categories, inplace=True) # change column to categorical type data[c] = data[c].astype('category') # update categories information _categories.update({c: categories}) else: # change column to categorical type data[c] = data[c].astype('category') # change column to categorical type _categories.update({ c: dict(enumerate( data[c].cat.categories, )) }) return _categories
[ "def", "categorize", "(", "data", ",", "col_name", ":", "str", "=", "None", ",", "new_col_name", ":", "str", "=", "None", ",", "categories", ":", "dict", "=", "None", ",", "max_categories", ":", "float", "=", "0.15", ")", ":", "_categories", "=", "{", "}", "if", "col_name", "is", "None", ":", "if", "categories", "is", "not", "None", ":", "raise", "Exception", "(", "'col_name is None when categories was defined.'", ")", "# create a list of cols with all object columns", "cols", "=", "[", "k", "for", "k", "in", "data", ".", "keys", "(", ")", "if", "data", "[", "k", "]", ".", "dtype", "==", "'object'", "and", "(", "data", "[", "k", "]", ".", "unique", "(", ")", "/", "data", "[", "k", "]", ".", "count", "(", ")", ")", "<=", "max_categories", "]", "else", ":", "# create a list with col_name", "if", "new_col_name", "is", "not", "None", ":", "data", "[", "new_col_name", "]", "=", "data", "[", "col_name", "]", "col_name", "=", "new_col_name", "cols", "=", "[", "col_name", "]", "for", "c", "in", "cols", ":", "if", "categories", "is", "not", "None", ":", "# assert all keys is a number", "assert", "all", "(", "type", "(", "k", ")", "in", "(", "int", ",", "float", ")", "for", "k", "in", "categories", ".", "keys", "(", ")", ")", "# replace values using given categories dict", "data", "[", "c", "]", ".", "replace", "(", "categories", ",", "inplace", "=", "True", ")", "# change column to categorical type", "data", "[", "c", "]", "=", "data", "[", "c", "]", ".", "astype", "(", "'category'", ")", "# update categories information", "_categories", ".", "update", "(", "{", "c", ":", "categories", "}", ")", "else", ":", "# change column to categorical type", "data", "[", "c", "]", "=", "data", "[", "c", "]", ".", "astype", "(", "'category'", ")", "# change column to categorical type", "_categories", ".", "update", "(", "{", "c", ":", "dict", "(", "enumerate", "(", "data", "[", "c", "]", ".", "cat", ".", "categories", ",", ")", ")", "}", ")", "return", "_categories" ]
:param data: :param col_name: :param new_col_name: :param categories: :param max_categories: max proportion threshold of categories :return: new categories :rtype dict:
[ ":", "param", "data", ":", ":", "param", "col_name", ":", ":", "param", "new_col_name", ":", ":", "param", "categories", ":", ":", "param", "max_categories", ":", "max", "proportion", "threshold", "of", "categories", ":", "return", ":", "new", "categories", ":", "rtype", "dict", ":" ]
34f06845a944ff4f048b55c7babdd8420f71a6b9
https://github.com/OpenDataScienceLab/skdata/blob/34f06845a944ff4f048b55c7babdd8420f71a6b9/skdata/cleaning.py#L4-L57
train
OpenDataScienceLab/skdata
skdata/cleaning.py
dropna
def dropna(data: pd.DataFrame, axis: int, **params): """ Remove columns with more NA values than threshold level :param data: :param axis: Axes are defined for arrays with more than one dimension. A 2-dimensional array has two corresponding axes: the first running vertically downwards across rows (axis 0), and the second running horizontally across columns (axis 1). (https://docs.scipy.org/doc/numpy-1.10.0/glossary.html) :param params: :return: """ if axis == 0: dropna_rows(data=data, **params) else: dropna_columns(data=data, **params)
python
def dropna(data: pd.DataFrame, axis: int, **params): """ Remove columns with more NA values than threshold level :param data: :param axis: Axes are defined for arrays with more than one dimension. A 2-dimensional array has two corresponding axes: the first running vertically downwards across rows (axis 0), and the second running horizontally across columns (axis 1). (https://docs.scipy.org/doc/numpy-1.10.0/glossary.html) :param params: :return: """ if axis == 0: dropna_rows(data=data, **params) else: dropna_columns(data=data, **params)
[ "def", "dropna", "(", "data", ":", "pd", ".", "DataFrame", ",", "axis", ":", "int", ",", "*", "*", "params", ")", ":", "if", "axis", "==", "0", ":", "dropna_rows", "(", "data", "=", "data", ",", "*", "*", "params", ")", "else", ":", "dropna_columns", "(", "data", "=", "data", ",", "*", "*", "params", ")" ]
Remove columns with more NA values than threshold level :param data: :param axis: Axes are defined for arrays with more than one dimension. A 2-dimensional array has two corresponding axes: the first running vertically downwards across rows (axis 0), and the second running horizontally across columns (axis 1). (https://docs.scipy.org/doc/numpy-1.10.0/glossary.html) :param params: :return:
[ "Remove", "columns", "with", "more", "NA", "values", "than", "threshold", "level" ]
34f06845a944ff4f048b55c7babdd8420f71a6b9
https://github.com/OpenDataScienceLab/skdata/blob/34f06845a944ff4f048b55c7babdd8420f71a6b9/skdata/cleaning.py#L60-L78
train
OpenDataScienceLab/skdata
skdata/cleaning.py
dropna_columns
def dropna_columns(data: pd.DataFrame, max_na_values: int=0.15): """ Remove columns with more NA values than threshold level :param data: :param max_na_values: proportion threshold of max na values :return: """ size = data.shape[0] df_na = (data.isnull().sum()/size) >= max_na_values data.drop(df_na[df_na].index, axis=1, inplace=True)
python
def dropna_columns(data: pd.DataFrame, max_na_values: int=0.15): """ Remove columns with more NA values than threshold level :param data: :param max_na_values: proportion threshold of max na values :return: """ size = data.shape[0] df_na = (data.isnull().sum()/size) >= max_na_values data.drop(df_na[df_na].index, axis=1, inplace=True)
[ "def", "dropna_columns", "(", "data", ":", "pd", ".", "DataFrame", ",", "max_na_values", ":", "int", "=", "0.15", ")", ":", "size", "=", "data", ".", "shape", "[", "0", "]", "df_na", "=", "(", "data", ".", "isnull", "(", ")", ".", "sum", "(", ")", "/", "size", ")", ">=", "max_na_values", "data", ".", "drop", "(", "df_na", "[", "df_na", "]", ".", "index", ",", "axis", "=", "1", ",", "inplace", "=", "True", ")" ]
Remove columns with more NA values than threshold level :param data: :param max_na_values: proportion threshold of max na values :return:
[ "Remove", "columns", "with", "more", "NA", "values", "than", "threshold", "level" ]
34f06845a944ff4f048b55c7babdd8420f71a6b9
https://github.com/OpenDataScienceLab/skdata/blob/34f06845a944ff4f048b55c7babdd8420f71a6b9/skdata/cleaning.py#L81-L92
train
OpenDataScienceLab/skdata
skdata/cleaning.py
dropna_rows
def dropna_rows(data: pd.DataFrame, columns_name: str=None): """ Remove columns with more NA values than threshold level :param data: :param columns_name: :return: """ params = {} if columns_name is not None: params.update({'subset': columns_name.split(',')}) data.dropna(inplace=True, **params)
python
def dropna_rows(data: pd.DataFrame, columns_name: str=None): """ Remove columns with more NA values than threshold level :param data: :param columns_name: :return: """ params = {} if columns_name is not None: params.update({'subset': columns_name.split(',')}) data.dropna(inplace=True, **params)
[ "def", "dropna_rows", "(", "data", ":", "pd", ".", "DataFrame", ",", "columns_name", ":", "str", "=", "None", ")", ":", "params", "=", "{", "}", "if", "columns_name", "is", "not", "None", ":", "params", ".", "update", "(", "{", "'subset'", ":", "columns_name", ".", "split", "(", "','", ")", "}", ")", "data", ".", "dropna", "(", "inplace", "=", "True", ",", "*", "*", "params", ")" ]
Remove columns with more NA values than threshold level :param data: :param columns_name: :return:
[ "Remove", "columns", "with", "more", "NA", "values", "than", "threshold", "level" ]
34f06845a944ff4f048b55c7babdd8420f71a6b9
https://github.com/OpenDataScienceLab/skdata/blob/34f06845a944ff4f048b55c7babdd8420f71a6b9/skdata/cleaning.py#L95-L107
train
OpenDataScienceLab/skdata
skdata/cleaning.py
drop_columns_with_unique_values
def drop_columns_with_unique_values( data: pd.DataFrame, max_unique_values: int = 0.25 ): """ Remove columns when the proportion of the total of unique values is more than the max_unique_values threshold, just for columns with type as object or category :param data: :param max_unique_values: :return: """ size = data.shape[0] df_uv = data.apply( lambda se: ( (se.dropna().unique().shape[0]/size) > max_unique_values and se.dtype in ['object', 'category'] ) ) data.drop(df_uv[df_uv].index, axis=1, inplace=True)
python
def drop_columns_with_unique_values( data: pd.DataFrame, max_unique_values: int = 0.25 ): """ Remove columns when the proportion of the total of unique values is more than the max_unique_values threshold, just for columns with type as object or category :param data: :param max_unique_values: :return: """ size = data.shape[0] df_uv = data.apply( lambda se: ( (se.dropna().unique().shape[0]/size) > max_unique_values and se.dtype in ['object', 'category'] ) ) data.drop(df_uv[df_uv].index, axis=1, inplace=True)
[ "def", "drop_columns_with_unique_values", "(", "data", ":", "pd", ".", "DataFrame", ",", "max_unique_values", ":", "int", "=", "0.25", ")", ":", "size", "=", "data", ".", "shape", "[", "0", "]", "df_uv", "=", "data", ".", "apply", "(", "lambda", "se", ":", "(", "(", "se", ".", "dropna", "(", ")", ".", "unique", "(", ")", ".", "shape", "[", "0", "]", "/", "size", ")", ">", "max_unique_values", "and", "se", ".", "dtype", "in", "[", "'object'", ",", "'category'", "]", ")", ")", "data", ".", "drop", "(", "df_uv", "[", "df_uv", "]", ".", "index", ",", "axis", "=", "1", ",", "inplace", "=", "True", ")" ]
Remove columns when the proportion of the total of unique values is more than the max_unique_values threshold, just for columns with type as object or category :param data: :param max_unique_values: :return:
[ "Remove", "columns", "when", "the", "proportion", "of", "the", "total", "of", "unique", "values", "is", "more", "than", "the", "max_unique_values", "threshold", "just", "for", "columns", "with", "type", "as", "object", "or", "category" ]
34f06845a944ff4f048b55c7babdd8420f71a6b9
https://github.com/OpenDataScienceLab/skdata/blob/34f06845a944ff4f048b55c7babdd8420f71a6b9/skdata/cleaning.py#L110-L130
train
opennode/waldur-core
waldur_core/quotas/views.py
QuotaViewSet.list
def list(self, request, *args, **kwargs): """ To get an actual value for object quotas limit and usage issue a **GET** request against */api/<objects>/*. To get all quotas visible to the user issue a **GET** request against */api/quotas/* """ return super(QuotaViewSet, self).list(request, *args, **kwargs)
python
def list(self, request, *args, **kwargs): """ To get an actual value for object quotas limit and usage issue a **GET** request against */api/<objects>/*. To get all quotas visible to the user issue a **GET** request against */api/quotas/* """ return super(QuotaViewSet, self).list(request, *args, **kwargs)
[ "def", "list", "(", "self", ",", "request", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "super", "(", "QuotaViewSet", ",", "self", ")", ".", "list", "(", "request", ",", "*", "args", ",", "*", "*", "kwargs", ")" ]
To get an actual value for object quotas limit and usage issue a **GET** request against */api/<objects>/*. To get all quotas visible to the user issue a **GET** request against */api/quotas/*
[ "To", "get", "an", "actual", "value", "for", "object", "quotas", "limit", "and", "usage", "issue", "a", "**", "GET", "**", "request", "against", "*", "/", "api", "/", "<objects", ">", "/", "*", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/quotas/views.py#L27-L33
train
opennode/waldur-core
waldur_core/quotas/views.py
QuotaViewSet.retrieve
def retrieve(self, request, *args, **kwargs): """ To set quota limit issue a **PUT** request against */api/quotas/<quota uuid>** with limit values. Please note that if a quota is a cache of a backend quota (e.g. 'storage' size of an OpenStack tenant), it will be impossible to modify it through */api/quotas/<quota uuid>** endpoint. Example of changing quota limit: .. code-block:: http POST /api/quotas/6ad5f49d6d6c49648573b2b71f44a42b/ HTTP/1.1 Content-Type: application/json Accept: application/json Authorization: Token c84d653b9ec92c6cbac41c706593e66f567a7fa4 Host: example.com { "limit": 2000.0 } Example of changing quota threshold: .. code-block:: http PUT /api/quotas/6ad5f49d6d6c49648573b2b71f44a42b/ HTTP/1.1 Content-Type: application/json Accept: application/json Authorization: Token c84d653b9ec92c6cbac41c706593e66f567a7fa4 Host: example.com { "threshold": 100.0 } """ return super(QuotaViewSet, self).retrieve(request, *args, **kwargs)
python
def retrieve(self, request, *args, **kwargs): """ To set quota limit issue a **PUT** request against */api/quotas/<quota uuid>** with limit values. Please note that if a quota is a cache of a backend quota (e.g. 'storage' size of an OpenStack tenant), it will be impossible to modify it through */api/quotas/<quota uuid>** endpoint. Example of changing quota limit: .. code-block:: http POST /api/quotas/6ad5f49d6d6c49648573b2b71f44a42b/ HTTP/1.1 Content-Type: application/json Accept: application/json Authorization: Token c84d653b9ec92c6cbac41c706593e66f567a7fa4 Host: example.com { "limit": 2000.0 } Example of changing quota threshold: .. code-block:: http PUT /api/quotas/6ad5f49d6d6c49648573b2b71f44a42b/ HTTP/1.1 Content-Type: application/json Accept: application/json Authorization: Token c84d653b9ec92c6cbac41c706593e66f567a7fa4 Host: example.com { "threshold": 100.0 } """ return super(QuotaViewSet, self).retrieve(request, *args, **kwargs)
[ "def", "retrieve", "(", "self", ",", "request", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "super", "(", "QuotaViewSet", ",", "self", ")", ".", "retrieve", "(", "request", ",", "*", "args", ",", "*", "*", "kwargs", ")" ]
To set quota limit issue a **PUT** request against */api/quotas/<quota uuid>** with limit values. Please note that if a quota is a cache of a backend quota (e.g. 'storage' size of an OpenStack tenant), it will be impossible to modify it through */api/quotas/<quota uuid>** endpoint. Example of changing quota limit: .. code-block:: http POST /api/quotas/6ad5f49d6d6c49648573b2b71f44a42b/ HTTP/1.1 Content-Type: application/json Accept: application/json Authorization: Token c84d653b9ec92c6cbac41c706593e66f567a7fa4 Host: example.com { "limit": 2000.0 } Example of changing quota threshold: .. code-block:: http PUT /api/quotas/6ad5f49d6d6c49648573b2b71f44a42b/ HTTP/1.1 Content-Type: application/json Accept: application/json Authorization: Token c84d653b9ec92c6cbac41c706593e66f567a7fa4 Host: example.com { "threshold": 100.0 }
[ "To", "set", "quota", "limit", "issue", "a", "**", "PUT", "**", "request", "against", "*", "/", "api", "/", "quotas", "/", "<quota", "uuid", ">", "**", "with", "limit", "values", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/quotas/views.py#L35-L71
train
opennode/waldur-core
waldur_core/quotas/views.py
QuotaViewSet.history
def history(self, request, uuid=None): """ Historical data endpoints could be available for any objects (currently implemented for quotas and events count). The data is available at *<object_endpoint>/history/*, for example: */api/quotas/<uuid>/history/*. There are two ways to define datetime points for historical data. 1. Send *?point=<timestamp>* parameter that can list. Response will contain historical data for each given point in the same order. 2. Send *?start=<timestamp>*, *?end=<timestamp>*, *?points_count=<integer>* parameters. Result will contain <points_count> points from <start> to <end>. Response format: .. code-block:: javascript [ { "point": <timestamp>, "object": {<object_representation>} }, { "point": <timestamp> "object": {<object_representation>} }, ... ] NB! There will not be any "object" for corresponding point in response if there is no data about object for a given timestamp. """ mapped = { 'start': request.query_params.get('start'), 'end': request.query_params.get('end'), 'points_count': request.query_params.get('points_count'), 'point_list': request.query_params.getlist('point'), } history_serializer = HistorySerializer(data={k: v for k, v in mapped.items() if v}) history_serializer.is_valid(raise_exception=True) quota = self.get_object() serializer = self.get_serializer(quota) serialized_versions = [] for point_date in history_serializer.get_filter_data(): serialized = {'point': datetime_to_timestamp(point_date)} version = Version.objects.get_for_object(quota).filter(revision__date_created__lte=point_date) if version.exists(): # make copy of serialized data and update field that are stored in version version_object = version.first()._object_version.object serialized['object'] = serializer.data.copy() serialized['object'].update({ f: getattr(version_object, f) for f in quota.get_version_fields() }) serialized_versions.append(serialized) return response.Response(serialized_versions, status=status.HTTP_200_OK)
python
def history(self, request, uuid=None): """ Historical data endpoints could be available for any objects (currently implemented for quotas and events count). The data is available at *<object_endpoint>/history/*, for example: */api/quotas/<uuid>/history/*. There are two ways to define datetime points for historical data. 1. Send *?point=<timestamp>* parameter that can list. Response will contain historical data for each given point in the same order. 2. Send *?start=<timestamp>*, *?end=<timestamp>*, *?points_count=<integer>* parameters. Result will contain <points_count> points from <start> to <end>. Response format: .. code-block:: javascript [ { "point": <timestamp>, "object": {<object_representation>} }, { "point": <timestamp> "object": {<object_representation>} }, ... ] NB! There will not be any "object" for corresponding point in response if there is no data about object for a given timestamp. """ mapped = { 'start': request.query_params.get('start'), 'end': request.query_params.get('end'), 'points_count': request.query_params.get('points_count'), 'point_list': request.query_params.getlist('point'), } history_serializer = HistorySerializer(data={k: v for k, v in mapped.items() if v}) history_serializer.is_valid(raise_exception=True) quota = self.get_object() serializer = self.get_serializer(quota) serialized_versions = [] for point_date in history_serializer.get_filter_data(): serialized = {'point': datetime_to_timestamp(point_date)} version = Version.objects.get_for_object(quota).filter(revision__date_created__lte=point_date) if version.exists(): # make copy of serialized data and update field that are stored in version version_object = version.first()._object_version.object serialized['object'] = serializer.data.copy() serialized['object'].update({ f: getattr(version_object, f) for f in quota.get_version_fields() }) serialized_versions.append(serialized) return response.Response(serialized_versions, status=status.HTTP_200_OK)
[ "def", "history", "(", "self", ",", "request", ",", "uuid", "=", "None", ")", ":", "mapped", "=", "{", "'start'", ":", "request", ".", "query_params", ".", "get", "(", "'start'", ")", ",", "'end'", ":", "request", ".", "query_params", ".", "get", "(", "'end'", ")", ",", "'points_count'", ":", "request", ".", "query_params", ".", "get", "(", "'points_count'", ")", ",", "'point_list'", ":", "request", ".", "query_params", ".", "getlist", "(", "'point'", ")", ",", "}", "history_serializer", "=", "HistorySerializer", "(", "data", "=", "{", "k", ":", "v", "for", "k", ",", "v", "in", "mapped", ".", "items", "(", ")", "if", "v", "}", ")", "history_serializer", ".", "is_valid", "(", "raise_exception", "=", "True", ")", "quota", "=", "self", ".", "get_object", "(", ")", "serializer", "=", "self", ".", "get_serializer", "(", "quota", ")", "serialized_versions", "=", "[", "]", "for", "point_date", "in", "history_serializer", ".", "get_filter_data", "(", ")", ":", "serialized", "=", "{", "'point'", ":", "datetime_to_timestamp", "(", "point_date", ")", "}", "version", "=", "Version", ".", "objects", ".", "get_for_object", "(", "quota", ")", ".", "filter", "(", "revision__date_created__lte", "=", "point_date", ")", "if", "version", ".", "exists", "(", ")", ":", "# make copy of serialized data and update field that are stored in version", "version_object", "=", "version", ".", "first", "(", ")", ".", "_object_version", ".", "object", "serialized", "[", "'object'", "]", "=", "serializer", ".", "data", ".", "copy", "(", ")", "serialized", "[", "'object'", "]", ".", "update", "(", "{", "f", ":", "getattr", "(", "version_object", ",", "f", ")", "for", "f", "in", "quota", ".", "get_version_fields", "(", ")", "}", ")", "serialized_versions", ".", "append", "(", "serialized", ")", "return", "response", ".", "Response", "(", "serialized_versions", ",", "status", "=", "status", ".", "HTTP_200_OK", ")" ]
Historical data endpoints could be available for any objects (currently implemented for quotas and events count). The data is available at *<object_endpoint>/history/*, for example: */api/quotas/<uuid>/history/*. There are two ways to define datetime points for historical data. 1. Send *?point=<timestamp>* parameter that can list. Response will contain historical data for each given point in the same order. 2. Send *?start=<timestamp>*, *?end=<timestamp>*, *?points_count=<integer>* parameters. Result will contain <points_count> points from <start> to <end>. Response format: .. code-block:: javascript [ { "point": <timestamp>, "object": {<object_representation>} }, { "point": <timestamp> "object": {<object_representation>} }, ... ] NB! There will not be any "object" for corresponding point in response if there is no data about object for a given timestamp.
[ "Historical", "data", "endpoints", "could", "be", "available", "for", "any", "objects", "(", "currently", "implemented", "for", "quotas", "and", "events", "count", ")", ".", "The", "data", "is", "available", "at", "*", "<object_endpoint", ">", "/", "history", "/", "*", "for", "example", ":", "*", "/", "api", "/", "quotas", "/", "<uuid", ">", "/", "history", "/", "*", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/quotas/views.py#L96-L151
train
opennode/waldur-core
waldur_core/core/serializers.py
GenericRelatedField._get_url
def _get_url(self, obj): """ Gets object url """ format_kwargs = { 'app_label': obj._meta.app_label, } try: format_kwargs['model_name'] = getattr(obj.__class__, 'get_url_name')() except AttributeError: format_kwargs['model_name'] = obj._meta.object_name.lower() return self._default_view_name % format_kwargs
python
def _get_url(self, obj): """ Gets object url """ format_kwargs = { 'app_label': obj._meta.app_label, } try: format_kwargs['model_name'] = getattr(obj.__class__, 'get_url_name')() except AttributeError: format_kwargs['model_name'] = obj._meta.object_name.lower() return self._default_view_name % format_kwargs
[ "def", "_get_url", "(", "self", ",", "obj", ")", ":", "format_kwargs", "=", "{", "'app_label'", ":", "obj", ".", "_meta", ".", "app_label", ",", "}", "try", ":", "format_kwargs", "[", "'model_name'", "]", "=", "getattr", "(", "obj", ".", "__class__", ",", "'get_url_name'", ")", "(", ")", "except", "AttributeError", ":", "format_kwargs", "[", "'model_name'", "]", "=", "obj", ".", "_meta", ".", "object_name", ".", "lower", "(", ")", "return", "self", ".", "_default_view_name", "%", "format_kwargs" ]
Gets object url
[ "Gets", "object", "url" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/serializers.py#L82-L93
train
opennode/waldur-core
waldur_core/core/serializers.py
GenericRelatedField.to_representation
def to_representation(self, obj): """ Serializes any object to his url representation """ kwargs = None for field in self.lookup_fields: if hasattr(obj, field): kwargs = {field: getattr(obj, field)} break if kwargs is None: raise AttributeError('Related object does not have any of lookup_fields') request = self._get_request() return request.build_absolute_uri(reverse(self._get_url(obj), kwargs=kwargs))
python
def to_representation(self, obj): """ Serializes any object to his url representation """ kwargs = None for field in self.lookup_fields: if hasattr(obj, field): kwargs = {field: getattr(obj, field)} break if kwargs is None: raise AttributeError('Related object does not have any of lookup_fields') request = self._get_request() return request.build_absolute_uri(reverse(self._get_url(obj), kwargs=kwargs))
[ "def", "to_representation", "(", "self", ",", "obj", ")", ":", "kwargs", "=", "None", "for", "field", "in", "self", ".", "lookup_fields", ":", "if", "hasattr", "(", "obj", ",", "field", ")", ":", "kwargs", "=", "{", "field", ":", "getattr", "(", "obj", ",", "field", ")", "}", "break", "if", "kwargs", "is", "None", ":", "raise", "AttributeError", "(", "'Related object does not have any of lookup_fields'", ")", "request", "=", "self", ".", "_get_request", "(", ")", "return", "request", ".", "build_absolute_uri", "(", "reverse", "(", "self", ".", "_get_url", "(", "obj", ")", ",", "kwargs", "=", "kwargs", ")", ")" ]
Serializes any object to his url representation
[ "Serializes", "any", "object", "to", "his", "url", "representation" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/serializers.py#L101-L113
train
opennode/waldur-core
waldur_core/core/serializers.py
GenericRelatedField.to_internal_value
def to_internal_value(self, data): """ Restores model instance from its url """ if not data: return None request = self._get_request() user = request.user try: obj = core_utils.instance_from_url(data, user=user) model = obj.__class__ except ValueError: raise serializers.ValidationError(_('URL is invalid: %s.') % data) except (Resolver404, AttributeError, MultipleObjectsReturned, ObjectDoesNotExist): raise serializers.ValidationError(_("Can't restore object from url: %s") % data) if model not in self.related_models: raise serializers.ValidationError(_('%s object does not support such relationship.') % six.text_type(obj)) return obj
python
def to_internal_value(self, data): """ Restores model instance from its url """ if not data: return None request = self._get_request() user = request.user try: obj = core_utils.instance_from_url(data, user=user) model = obj.__class__ except ValueError: raise serializers.ValidationError(_('URL is invalid: %s.') % data) except (Resolver404, AttributeError, MultipleObjectsReturned, ObjectDoesNotExist): raise serializers.ValidationError(_("Can't restore object from url: %s") % data) if model not in self.related_models: raise serializers.ValidationError(_('%s object does not support such relationship.') % six.text_type(obj)) return obj
[ "def", "to_internal_value", "(", "self", ",", "data", ")", ":", "if", "not", "data", ":", "return", "None", "request", "=", "self", ".", "_get_request", "(", ")", "user", "=", "request", ".", "user", "try", ":", "obj", "=", "core_utils", ".", "instance_from_url", "(", "data", ",", "user", "=", "user", ")", "model", "=", "obj", ".", "__class__", "except", "ValueError", ":", "raise", "serializers", ".", "ValidationError", "(", "_", "(", "'URL is invalid: %s.'", ")", "%", "data", ")", "except", "(", "Resolver404", ",", "AttributeError", ",", "MultipleObjectsReturned", ",", "ObjectDoesNotExist", ")", ":", "raise", "serializers", ".", "ValidationError", "(", "_", "(", "\"Can't restore object from url: %s\"", ")", "%", "data", ")", "if", "model", "not", "in", "self", ".", "related_models", ":", "raise", "serializers", ".", "ValidationError", "(", "_", "(", "'%s object does not support such relationship.'", ")", "%", "six", ".", "text_type", "(", "obj", ")", ")", "return", "obj" ]
Restores model instance from its url
[ "Restores", "model", "instance", "from", "its", "url" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/serializers.py#L115-L132
train
opennode/waldur-core
waldur_core/core/serializers.py
TimestampIntervalSerializer.validate
def validate(self, data): """ Check that the start is before the end. """ if 'start' in data and 'end' in data and data['start'] >= data['end']: raise serializers.ValidationError(_('End must occur after start.')) return data
python
def validate(self, data): """ Check that the start is before the end. """ if 'start' in data and 'end' in data and data['start'] >= data['end']: raise serializers.ValidationError(_('End must occur after start.')) return data
[ "def", "validate", "(", "self", ",", "data", ")", ":", "if", "'start'", "in", "data", "and", "'end'", "in", "data", "and", "data", "[", "'start'", "]", ">=", "data", "[", "'end'", "]", ":", "raise", "serializers", ".", "ValidationError", "(", "_", "(", "'End must occur after start.'", ")", ")", "return", "data" ]
Check that the start is before the end.
[ "Check", "that", "the", "start", "is", "before", "the", "end", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/serializers.py#L373-L379
train
opennode/waldur-core
waldur_core/core/tasks.py
send_task
def send_task(app_label, task_name): """ A helper function to deal with waldur_core "high-level" tasks. Define high-level task with explicit name using a pattern: waldur_core.<app_label>.<task_name> .. code-block:: python @shared_task(name='waldur_core.openstack.provision_instance') def provision_instance_fn(instance_uuid, backend_flavor_id) pass Call it by name: .. code-block:: python send_task('openstack', 'provision_instance')(instance_uuid, backend_flavor_id) Which is identical to: .. code-block:: python provision_instance_fn.delay(instance_uuid, backend_flavor_id) """ def delay(*args, **kwargs): full_task_name = 'waldur_core.%s.%s' % (app_label, task_name) send_celery_task(full_task_name, args, kwargs, countdown=2) return delay
python
def send_task(app_label, task_name): """ A helper function to deal with waldur_core "high-level" tasks. Define high-level task with explicit name using a pattern: waldur_core.<app_label>.<task_name> .. code-block:: python @shared_task(name='waldur_core.openstack.provision_instance') def provision_instance_fn(instance_uuid, backend_flavor_id) pass Call it by name: .. code-block:: python send_task('openstack', 'provision_instance')(instance_uuid, backend_flavor_id) Which is identical to: .. code-block:: python provision_instance_fn.delay(instance_uuid, backend_flavor_id) """ def delay(*args, **kwargs): full_task_name = 'waldur_core.%s.%s' % (app_label, task_name) send_celery_task(full_task_name, args, kwargs, countdown=2) return delay
[ "def", "send_task", "(", "app_label", ",", "task_name", ")", ":", "def", "delay", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "full_task_name", "=", "'waldur_core.%s.%s'", "%", "(", "app_label", ",", "task_name", ")", "send_celery_task", "(", "full_task_name", ",", "args", ",", "kwargs", ",", "countdown", "=", "2", ")", "return", "delay" ]
A helper function to deal with waldur_core "high-level" tasks. Define high-level task with explicit name using a pattern: waldur_core.<app_label>.<task_name> .. code-block:: python @shared_task(name='waldur_core.openstack.provision_instance') def provision_instance_fn(instance_uuid, backend_flavor_id) pass Call it by name: .. code-block:: python send_task('openstack', 'provision_instance')(instance_uuid, backend_flavor_id) Which is identical to: .. code-block:: python provision_instance_fn.delay(instance_uuid, backend_flavor_id)
[ "A", "helper", "function", "to", "deal", "with", "waldur_core", "high", "-", "level", "tasks", ".", "Define", "high", "-", "level", "task", "with", "explicit", "name", "using", "a", "pattern", ":", "waldur_core", ".", "<app_label", ">", ".", "<task_name", ">" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L54-L80
train
opennode/waldur-core
waldur_core/core/tasks.py
log_celery_task
def log_celery_task(request): """ Add description to celery log output """ task = request.task description = None if isinstance(task, Task): try: description = task.get_description(*request.args, **request.kwargs) except NotImplementedError: pass except Exception as e: # Logging should never break workflow. logger.exception('Cannot get description for task %s. Error: %s' % (task.__class__.__name__, e)) return '{0.name}[{0.id}]{1}{2}{3}'.format( request, ' {0}'.format(description) if description else '', ' eta:[{0}]'.format(request.eta) if request.eta else '', ' expires:[{0}]'.format(request.expires) if request.expires else '', )
python
def log_celery_task(request): """ Add description to celery log output """ task = request.task description = None if isinstance(task, Task): try: description = task.get_description(*request.args, **request.kwargs) except NotImplementedError: pass except Exception as e: # Logging should never break workflow. logger.exception('Cannot get description for task %s. Error: %s' % (task.__class__.__name__, e)) return '{0.name}[{0.id}]{1}{2}{3}'.format( request, ' {0}'.format(description) if description else '', ' eta:[{0}]'.format(request.eta) if request.eta else '', ' expires:[{0}]'.format(request.expires) if request.expires else '', )
[ "def", "log_celery_task", "(", "request", ")", ":", "task", "=", "request", ".", "task", "description", "=", "None", "if", "isinstance", "(", "task", ",", "Task", ")", ":", "try", ":", "description", "=", "task", ".", "get_description", "(", "*", "request", ".", "args", ",", "*", "*", "request", ".", "kwargs", ")", "except", "NotImplementedError", ":", "pass", "except", "Exception", "as", "e", ":", "# Logging should never break workflow.", "logger", ".", "exception", "(", "'Cannot get description for task %s. Error: %s'", "%", "(", "task", ".", "__class__", ".", "__name__", ",", "e", ")", ")", "return", "'{0.name}[{0.id}]{1}{2}{3}'", ".", "format", "(", "request", ",", "' {0}'", ".", "format", "(", "description", ")", "if", "description", "else", "''", ",", "' eta:[{0}]'", ".", "format", "(", "request", ".", "eta", ")", "if", "request", ".", "eta", "else", "''", ",", "' expires:[{0}]'", ".", "format", "(", "request", ".", "expires", ")", "if", "request", ".", "expires", "else", "''", ",", ")" ]
Add description to celery log output
[ "Add", "description", "to", "celery", "log", "output" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L440-L458
train
opennode/waldur-core
waldur_core/core/tasks.py
Task.run
def run(self, serialized_instance, *args, **kwargs): """ Deserialize input data and start backend operation execution """ try: instance = utils.deserialize_instance(serialized_instance) except ObjectDoesNotExist: message = ('Cannot restore instance from serialized object %s. Probably it was deleted.' % serialized_instance) six.reraise(ObjectDoesNotExist, message) self.args = args self.kwargs = kwargs self.pre_execute(instance) result = self.execute(instance, *self.args, **self.kwargs) self.post_execute(instance) if result and isinstance(result, django_models.Model): result = utils.serialize_instance(result) return result
python
def run(self, serialized_instance, *args, **kwargs): """ Deserialize input data and start backend operation execution """ try: instance = utils.deserialize_instance(serialized_instance) except ObjectDoesNotExist: message = ('Cannot restore instance from serialized object %s. Probably it was deleted.' % serialized_instance) six.reraise(ObjectDoesNotExist, message) self.args = args self.kwargs = kwargs self.pre_execute(instance) result = self.execute(instance, *self.args, **self.kwargs) self.post_execute(instance) if result and isinstance(result, django_models.Model): result = utils.serialize_instance(result) return result
[ "def", "run", "(", "self", ",", "serialized_instance", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "try", ":", "instance", "=", "utils", ".", "deserialize_instance", "(", "serialized_instance", ")", "except", "ObjectDoesNotExist", ":", "message", "=", "(", "'Cannot restore instance from serialized object %s. Probably it was deleted.'", "%", "serialized_instance", ")", "six", ".", "reraise", "(", "ObjectDoesNotExist", ",", "message", ")", "self", ".", "args", "=", "args", "self", ".", "kwargs", "=", "kwargs", "self", ".", "pre_execute", "(", "instance", ")", "result", "=", "self", ".", "execute", "(", "instance", ",", "*", "self", ".", "args", ",", "*", "*", "self", ".", "kwargs", ")", "self", ".", "post_execute", "(", "instance", ")", "if", "result", "and", "isinstance", "(", "result", ",", "django_models", ".", "Model", ")", ":", "result", "=", "utils", ".", "serialize_instance", "(", "result", ")", "return", "result" ]
Deserialize input data and start backend operation execution
[ "Deserialize", "input", "data", "and", "start", "backend", "operation", "execution" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L97-L114
train
opennode/waldur-core
waldur_core/core/tasks.py
BackgroundTask.is_previous_task_processing
def is_previous_task_processing(self, *args, **kwargs): """ Return True if exist task that is equal to current and is uncompleted """ app = self._get_app() inspect = app.control.inspect() active = inspect.active() or {} scheduled = inspect.scheduled() or {} reserved = inspect.reserved() or {} uncompleted = sum(list(active.values()) + list(scheduled.values()) + reserved.values(), []) return any(self.is_equal(task, *args, **kwargs) for task in uncompleted)
python
def is_previous_task_processing(self, *args, **kwargs): """ Return True if exist task that is equal to current and is uncompleted """ app = self._get_app() inspect = app.control.inspect() active = inspect.active() or {} scheduled = inspect.scheduled() or {} reserved = inspect.reserved() or {} uncompleted = sum(list(active.values()) + list(scheduled.values()) + reserved.values(), []) return any(self.is_equal(task, *args, **kwargs) for task in uncompleted)
[ "def", "is_previous_task_processing", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "app", "=", "self", ".", "_get_app", "(", ")", "inspect", "=", "app", ".", "control", ".", "inspect", "(", ")", "active", "=", "inspect", ".", "active", "(", ")", "or", "{", "}", "scheduled", "=", "inspect", ".", "scheduled", "(", ")", "or", "{", "}", "reserved", "=", "inspect", ".", "reserved", "(", ")", "or", "{", "}", "uncompleted", "=", "sum", "(", "list", "(", "active", ".", "values", "(", ")", ")", "+", "list", "(", "scheduled", ".", "values", "(", ")", ")", "+", "reserved", ".", "values", "(", ")", ",", "[", "]", ")", "return", "any", "(", "self", ".", "is_equal", "(", "task", ",", "*", "args", ",", "*", "*", "kwargs", ")", "for", "task", "in", "uncompleted", ")" ]
Return True if exist task that is equal to current and is uncompleted
[ "Return", "True", "if", "exist", "task", "that", "is", "equal", "to", "current", "and", "is", "uncompleted" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L353-L361
train
opennode/waldur-core
waldur_core/core/tasks.py
BackgroundTask.apply_async
def apply_async(self, args=None, kwargs=None, **options): """ Do not run background task if previous task is uncompleted """ if self.is_previous_task_processing(*args, **kwargs): message = 'Background task %s was not scheduled, because its predecessor is not completed yet.' % self.name logger.info(message) # It is expected by Celery that apply_async return AsyncResult, otherwise celerybeat dies return self.AsyncResult(options.get('task_id') or str(uuid4())) return super(BackgroundTask, self).apply_async(args=args, kwargs=kwargs, **options)
python
def apply_async(self, args=None, kwargs=None, **options): """ Do not run background task if previous task is uncompleted """ if self.is_previous_task_processing(*args, **kwargs): message = 'Background task %s was not scheduled, because its predecessor is not completed yet.' % self.name logger.info(message) # It is expected by Celery that apply_async return AsyncResult, otherwise celerybeat dies return self.AsyncResult(options.get('task_id') or str(uuid4())) return super(BackgroundTask, self).apply_async(args=args, kwargs=kwargs, **options)
[ "def", "apply_async", "(", "self", ",", "args", "=", "None", ",", "kwargs", "=", "None", ",", "*", "*", "options", ")", ":", "if", "self", ".", "is_previous_task_processing", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "message", "=", "'Background task %s was not scheduled, because its predecessor is not completed yet.'", "%", "self", ".", "name", "logger", ".", "info", "(", "message", ")", "# It is expected by Celery that apply_async return AsyncResult, otherwise celerybeat dies", "return", "self", ".", "AsyncResult", "(", "options", ".", "get", "(", "'task_id'", ")", "or", "str", "(", "uuid4", "(", ")", ")", ")", "return", "super", "(", "BackgroundTask", ",", "self", ")", ".", "apply_async", "(", "args", "=", "args", ",", "kwargs", "=", "kwargs", ",", "*", "*", "options", ")" ]
Do not run background task if previous task is uncompleted
[ "Do", "not", "run", "background", "task", "if", "previous", "task", "is", "uncompleted" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L363-L370
train
opennode/waldur-core
waldur_core/core/tasks.py
PenalizedBackgroundTask._get_cache_key
def _get_cache_key(self, args, kwargs): """ Returns key to be used in cache """ hash_input = json.dumps({'name': self.name, 'args': args, 'kwargs': kwargs}, sort_keys=True) # md5 is used for internal caching, not need to care about security return hashlib.md5(hash_input).hexdigest()
python
def _get_cache_key(self, args, kwargs): """ Returns key to be used in cache """ hash_input = json.dumps({'name': self.name, 'args': args, 'kwargs': kwargs}, sort_keys=True) # md5 is used for internal caching, not need to care about security return hashlib.md5(hash_input).hexdigest()
[ "def", "_get_cache_key", "(", "self", ",", "args", ",", "kwargs", ")", ":", "hash_input", "=", "json", ".", "dumps", "(", "{", "'name'", ":", "self", ".", "name", ",", "'args'", ":", "args", ",", "'kwargs'", ":", "kwargs", "}", ",", "sort_keys", "=", "True", ")", "# md5 is used for internal caching, not need to care about security", "return", "hashlib", ".", "md5", "(", "hash_input", ")", ".", "hexdigest", "(", ")" ]
Returns key to be used in cache
[ "Returns", "key", "to", "be", "used", "in", "cache" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L396-L400
train
opennode/waldur-core
waldur_core/core/tasks.py
PenalizedBackgroundTask.apply_async
def apply_async(self, args=None, kwargs=None, **options): """ Checks whether task must be skipped and decreases the counter in that case. """ key = self._get_cache_key(args, kwargs) counter, penalty = cache.get(key, (0, 0)) if not counter: return super(PenalizedBackgroundTask, self).apply_async(args=args, kwargs=kwargs, **options) cache.set(key, (counter - 1, penalty), self.CACHE_LIFETIME) logger.info('The task %s will not be executed due to the penalty.' % self.name) return self.AsyncResult(options.get('task_id') or str(uuid4()))
python
def apply_async(self, args=None, kwargs=None, **options): """ Checks whether task must be skipped and decreases the counter in that case. """ key = self._get_cache_key(args, kwargs) counter, penalty = cache.get(key, (0, 0)) if not counter: return super(PenalizedBackgroundTask, self).apply_async(args=args, kwargs=kwargs, **options) cache.set(key, (counter - 1, penalty), self.CACHE_LIFETIME) logger.info('The task %s will not be executed due to the penalty.' % self.name) return self.AsyncResult(options.get('task_id') or str(uuid4()))
[ "def", "apply_async", "(", "self", ",", "args", "=", "None", ",", "kwargs", "=", "None", ",", "*", "*", "options", ")", ":", "key", "=", "self", ".", "_get_cache_key", "(", "args", ",", "kwargs", ")", "counter", ",", "penalty", "=", "cache", ".", "get", "(", "key", ",", "(", "0", ",", "0", ")", ")", "if", "not", "counter", ":", "return", "super", "(", "PenalizedBackgroundTask", ",", "self", ")", ".", "apply_async", "(", "args", "=", "args", ",", "kwargs", "=", "kwargs", ",", "*", "*", "options", ")", "cache", ".", "set", "(", "key", ",", "(", "counter", "-", "1", ",", "penalty", ")", ",", "self", ".", "CACHE_LIFETIME", ")", "logger", ".", "info", "(", "'The task %s will not be executed due to the penalty.'", "%", "self", ".", "name", ")", "return", "self", ".", "AsyncResult", "(", "options", ".", "get", "(", "'task_id'", ")", "or", "str", "(", "uuid4", "(", ")", ")", ")" ]
Checks whether task must be skipped and decreases the counter in that case.
[ "Checks", "whether", "task", "must", "be", "skipped", "and", "decreases", "the", "counter", "in", "that", "case", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L402-L413
train
opennode/waldur-core
waldur_core/core/tasks.py
PenalizedBackgroundTask.on_failure
def on_failure(self, exc, task_id, args, kwargs, einfo): """ Increases penalty for the task and resets the counter. """ key = self._get_cache_key(args, kwargs) _, penalty = cache.get(key, (0, 0)) if penalty < self.MAX_PENALTY: penalty += 1 logger.debug('The task %s is penalized and will be executed on %d run.' % (self.name, penalty)) cache.set(key, (penalty, penalty), self.CACHE_LIFETIME) return super(PenalizedBackgroundTask, self).on_failure(exc, task_id, args, kwargs, einfo)
python
def on_failure(self, exc, task_id, args, kwargs, einfo): """ Increases penalty for the task and resets the counter. """ key = self._get_cache_key(args, kwargs) _, penalty = cache.get(key, (0, 0)) if penalty < self.MAX_PENALTY: penalty += 1 logger.debug('The task %s is penalized and will be executed on %d run.' % (self.name, penalty)) cache.set(key, (penalty, penalty), self.CACHE_LIFETIME) return super(PenalizedBackgroundTask, self).on_failure(exc, task_id, args, kwargs, einfo)
[ "def", "on_failure", "(", "self", ",", "exc", ",", "task_id", ",", "args", ",", "kwargs", ",", "einfo", ")", ":", "key", "=", "self", ".", "_get_cache_key", "(", "args", ",", "kwargs", ")", "_", ",", "penalty", "=", "cache", ".", "get", "(", "key", ",", "(", "0", ",", "0", ")", ")", "if", "penalty", "<", "self", ".", "MAX_PENALTY", ":", "penalty", "+=", "1", "logger", ".", "debug", "(", "'The task %s is penalized and will be executed on %d run.'", "%", "(", "self", ".", "name", ",", "penalty", ")", ")", "cache", ".", "set", "(", "key", ",", "(", "penalty", ",", "penalty", ")", ",", "self", ".", "CACHE_LIFETIME", ")", "return", "super", "(", "PenalizedBackgroundTask", ",", "self", ")", ".", "on_failure", "(", "exc", ",", "task_id", ",", "args", ",", "kwargs", ",", "einfo", ")" ]
Increases penalty for the task and resets the counter.
[ "Increases", "penalty", "for", "the", "task", "and", "resets", "the", "counter", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L415-L426
train
opennode/waldur-core
waldur_core/core/tasks.py
PenalizedBackgroundTask.on_success
def on_success(self, retval, task_id, args, kwargs): """ Clears cache for the task. """ key = self._get_cache_key(args, kwargs) if cache.get(key) is not None: cache.delete(key) logger.debug('Penalty for the task %s has been removed.' % self.name) return super(PenalizedBackgroundTask, self).on_success(retval, task_id, args, kwargs)
python
def on_success(self, retval, task_id, args, kwargs): """ Clears cache for the task. """ key = self._get_cache_key(args, kwargs) if cache.get(key) is not None: cache.delete(key) logger.debug('Penalty for the task %s has been removed.' % self.name) return super(PenalizedBackgroundTask, self).on_success(retval, task_id, args, kwargs)
[ "def", "on_success", "(", "self", ",", "retval", ",", "task_id", ",", "args", ",", "kwargs", ")", ":", "key", "=", "self", ".", "_get_cache_key", "(", "args", ",", "kwargs", ")", "if", "cache", ".", "get", "(", "key", ")", "is", "not", "None", ":", "cache", ".", "delete", "(", "key", ")", "logger", ".", "debug", "(", "'Penalty for the task %s has been removed.'", "%", "self", ".", "name", ")", "return", "super", "(", "PenalizedBackgroundTask", ",", "self", ")", ".", "on_success", "(", "retval", ",", "task_id", ",", "args", ",", "kwargs", ")" ]
Clears cache for the task.
[ "Clears", "cache", "for", "the", "task", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/tasks.py#L428-L437
train
opennode/waldur-core
waldur_core/structure/__init__.py
log_backend_action
def log_backend_action(action=None): """ Logging for backend method. Expects django model instance as first argument. """ def decorator(func): @functools.wraps(func) def wrapped(self, instance, *args, **kwargs): action_name = func.func_name.replace('_', ' ') if action is None else action logger.debug('About to %s `%s` (PK: %s).', action_name, instance, instance.pk) result = func(self, instance, *args, **kwargs) logger.debug('Action `%s` was executed successfully for `%s` (PK: %s).', action_name, instance, instance.pk) return result return wrapped return decorator
python
def log_backend_action(action=None): """ Logging for backend method. Expects django model instance as first argument. """ def decorator(func): @functools.wraps(func) def wrapped(self, instance, *args, **kwargs): action_name = func.func_name.replace('_', ' ') if action is None else action logger.debug('About to %s `%s` (PK: %s).', action_name, instance, instance.pk) result = func(self, instance, *args, **kwargs) logger.debug('Action `%s` was executed successfully for `%s` (PK: %s).', action_name, instance, instance.pk) return result return wrapped return decorator
[ "def", "log_backend_action", "(", "action", "=", "None", ")", ":", "def", "decorator", "(", "func", ")", ":", "@", "functools", ".", "wraps", "(", "func", ")", "def", "wrapped", "(", "self", ",", "instance", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "action_name", "=", "func", ".", "func_name", ".", "replace", "(", "'_'", ",", "' '", ")", "if", "action", "is", "None", "else", "action", "logger", ".", "debug", "(", "'About to %s `%s` (PK: %s).'", ",", "action_name", ",", "instance", ",", "instance", ".", "pk", ")", "result", "=", "func", "(", "self", ",", "instance", ",", "*", "args", ",", "*", "*", "kwargs", ")", "logger", ".", "debug", "(", "'Action `%s` was executed successfully for `%s` (PK: %s).'", ",", "action_name", ",", "instance", ",", "instance", ".", "pk", ")", "return", "result", "return", "wrapped", "return", "decorator" ]
Logging for backend method. Expects django model instance as first argument.
[ "Logging", "for", "backend", "method", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L417-L433
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_services
def get_services(cls, request=None): """ Get a list of services endpoints. { "Oracle": "/api/oracle/", "OpenStack": "/api/openstack/", "GitLab": "/api/gitlab/", "DigitalOcean": "/api/digitalocean/" } """ return {service['name']: reverse(service['list_view'], request=request) for service in cls._registry.values()}
python
def get_services(cls, request=None): """ Get a list of services endpoints. { "Oracle": "/api/oracle/", "OpenStack": "/api/openstack/", "GitLab": "/api/gitlab/", "DigitalOcean": "/api/digitalocean/" } """ return {service['name']: reverse(service['list_view'], request=request) for service in cls._registry.values()}
[ "def", "get_services", "(", "cls", ",", "request", "=", "None", ")", ":", "return", "{", "service", "[", "'name'", "]", ":", "reverse", "(", "service", "[", "'list_view'", "]", ",", "request", "=", "request", ")", "for", "service", "in", "cls", ".", "_registry", ".", "values", "(", ")", "}" ]
Get a list of services endpoints. { "Oracle": "/api/oracle/", "OpenStack": "/api/openstack/", "GitLab": "/api/gitlab/", "DigitalOcean": "/api/digitalocean/" }
[ "Get", "a", "list", "of", "services", "endpoints", ".", "{", "Oracle", ":", "/", "api", "/", "oracle", "/", "OpenStack", ":", "/", "api", "/", "openstack", "/", "GitLab", ":", "/", "api", "/", "gitlab", "/", "DigitalOcean", ":", "/", "api", "/", "digitalocean", "/", "}" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L161-L171
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_resources
def get_resources(cls, request=None): """ Get a list of resources endpoints. { "DigitalOcean.Droplet": "/api/digitalocean-droplets/", "Oracle.Database": "/api/oracle-databases/", "GitLab.Group": "/api/gitlab-groups/", "GitLab.Project": "/api/gitlab-projects/" } """ return {'.'.join([service['name'], resource['name']]): reverse(resource['list_view'], request=request) for service in cls._registry.values() for resource in service['resources'].values()}
python
def get_resources(cls, request=None): """ Get a list of resources endpoints. { "DigitalOcean.Droplet": "/api/digitalocean-droplets/", "Oracle.Database": "/api/oracle-databases/", "GitLab.Group": "/api/gitlab-groups/", "GitLab.Project": "/api/gitlab-projects/" } """ return {'.'.join([service['name'], resource['name']]): reverse(resource['list_view'], request=request) for service in cls._registry.values() for resource in service['resources'].values()}
[ "def", "get_resources", "(", "cls", ",", "request", "=", "None", ")", ":", "return", "{", "'.'", ".", "join", "(", "[", "service", "[", "'name'", "]", ",", "resource", "[", "'name'", "]", "]", ")", ":", "reverse", "(", "resource", "[", "'list_view'", "]", ",", "request", "=", "request", ")", "for", "service", "in", "cls", ".", "_registry", ".", "values", "(", ")", "for", "resource", "in", "service", "[", "'resources'", "]", ".", "values", "(", ")", "}" ]
Get a list of resources endpoints. { "DigitalOcean.Droplet": "/api/digitalocean-droplets/", "Oracle.Database": "/api/oracle-databases/", "GitLab.Group": "/api/gitlab-groups/", "GitLab.Project": "/api/gitlab-projects/" }
[ "Get", "a", "list", "of", "resources", "endpoints", ".", "{", "DigitalOcean", ".", "Droplet", ":", "/", "api", "/", "digitalocean", "-", "droplets", "/", "Oracle", ".", "Database", ":", "/", "api", "/", "oracle", "-", "databases", "/", "GitLab", ".", "Group", ":", "/", "api", "/", "gitlab", "-", "groups", "/", "GitLab", ".", "Project", ":", "/", "api", "/", "gitlab", "-", "projects", "/", "}" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L184-L195
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_services_with_resources
def get_services_with_resources(cls, request=None): """ Get a list of services and resources endpoints. { ... "GitLab": { "url": "/api/gitlab/", "service_project_link_url": "/api/gitlab-service-project-link/", "resources": { "Project": "/api/gitlab-projects/", "Group": "/api/gitlab-groups/" } }, ... } """ from django.apps import apps data = {} for service in cls._registry.values(): service_model = apps.get_model(service['model_name']) service_project_link = service_model.projects.through service_project_link_url = reverse(cls.get_list_view_for_model(service_project_link), request=request) data[service['name']] = { 'url': reverse(service['list_view'], request=request), 'service_project_link_url': service_project_link_url, 'resources': {resource['name']: reverse(resource['list_view'], request=request) for resource in service['resources'].values()}, 'properties': {resource['name']: reverse(resource['list_view'], request=request) for resource in service.get('properties', {}).values()}, 'is_public_service': cls.is_public_service(service_model) } return data
python
def get_services_with_resources(cls, request=None): """ Get a list of services and resources endpoints. { ... "GitLab": { "url": "/api/gitlab/", "service_project_link_url": "/api/gitlab-service-project-link/", "resources": { "Project": "/api/gitlab-projects/", "Group": "/api/gitlab-groups/" } }, ... } """ from django.apps import apps data = {} for service in cls._registry.values(): service_model = apps.get_model(service['model_name']) service_project_link = service_model.projects.through service_project_link_url = reverse(cls.get_list_view_for_model(service_project_link), request=request) data[service['name']] = { 'url': reverse(service['list_view'], request=request), 'service_project_link_url': service_project_link_url, 'resources': {resource['name']: reverse(resource['list_view'], request=request) for resource in service['resources'].values()}, 'properties': {resource['name']: reverse(resource['list_view'], request=request) for resource in service.get('properties', {}).values()}, 'is_public_service': cls.is_public_service(service_model) } return data
[ "def", "get_services_with_resources", "(", "cls", ",", "request", "=", "None", ")", ":", "from", "django", ".", "apps", "import", "apps", "data", "=", "{", "}", "for", "service", "in", "cls", ".", "_registry", ".", "values", "(", ")", ":", "service_model", "=", "apps", ".", "get_model", "(", "service", "[", "'model_name'", "]", ")", "service_project_link", "=", "service_model", ".", "projects", ".", "through", "service_project_link_url", "=", "reverse", "(", "cls", ".", "get_list_view_for_model", "(", "service_project_link", ")", ",", "request", "=", "request", ")", "data", "[", "service", "[", "'name'", "]", "]", "=", "{", "'url'", ":", "reverse", "(", "service", "[", "'list_view'", "]", ",", "request", "=", "request", ")", ",", "'service_project_link_url'", ":", "service_project_link_url", ",", "'resources'", ":", "{", "resource", "[", "'name'", "]", ":", "reverse", "(", "resource", "[", "'list_view'", "]", ",", "request", "=", "request", ")", "for", "resource", "in", "service", "[", "'resources'", "]", ".", "values", "(", ")", "}", ",", "'properties'", ":", "{", "resource", "[", "'name'", "]", ":", "reverse", "(", "resource", "[", "'list_view'", "]", ",", "request", "=", "request", ")", "for", "resource", "in", "service", ".", "get", "(", "'properties'", ",", "{", "}", ")", ".", "values", "(", ")", "}", ",", "'is_public_service'", ":", "cls", ".", "is_public_service", "(", "service_model", ")", "}", "return", "data" ]
Get a list of services and resources endpoints. { ... "GitLab": { "url": "/api/gitlab/", "service_project_link_url": "/api/gitlab-service-project-link/", "resources": { "Project": "/api/gitlab-projects/", "Group": "/api/gitlab-groups/" } }, ... }
[ "Get", "a", "list", "of", "services", "and", "resources", "endpoints", ".", "{", "...", "GitLab", ":", "{", "url", ":", "/", "api", "/", "gitlab", "/", "service_project_link_url", ":", "/", "api", "/", "gitlab", "-", "service", "-", "project", "-", "link", "/", "resources", ":", "{", "Project", ":", "/", "api", "/", "gitlab", "-", "projects", "/", "Group", ":", "/", "api", "/", "gitlab", "-", "groups", "/", "}", "}", "...", "}" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L216-L248
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_service_models
def get_service_models(cls): """ Get a list of service models. { ... 'gitlab': { "service": nodeconductor_gitlab.models.GitLabService, "service_project_link": nodeconductor_gitlab.models.GitLabServiceProjectLink, "resources": [ nodeconductor_gitlab.models.Group, nodeconductor_gitlab.models.Project ], }, ... } """ from django.apps import apps data = {} for key, service in cls._registry.items(): service_model = apps.get_model(service['model_name']) service_project_link = service_model.projects.through data[key] = { 'service': service_model, 'service_project_link': service_project_link, 'resources': [apps.get_model(r) for r in service['resources'].keys()], 'properties': [apps.get_model(r) for r in service['properties'].keys() if '.' in r], } return data
python
def get_service_models(cls): """ Get a list of service models. { ... 'gitlab': { "service": nodeconductor_gitlab.models.GitLabService, "service_project_link": nodeconductor_gitlab.models.GitLabServiceProjectLink, "resources": [ nodeconductor_gitlab.models.Group, nodeconductor_gitlab.models.Project ], }, ... } """ from django.apps import apps data = {} for key, service in cls._registry.items(): service_model = apps.get_model(service['model_name']) service_project_link = service_model.projects.through data[key] = { 'service': service_model, 'service_project_link': service_project_link, 'resources': [apps.get_model(r) for r in service['resources'].keys()], 'properties': [apps.get_model(r) for r in service['properties'].keys() if '.' in r], } return data
[ "def", "get_service_models", "(", "cls", ")", ":", "from", "django", ".", "apps", "import", "apps", "data", "=", "{", "}", "for", "key", ",", "service", "in", "cls", ".", "_registry", ".", "items", "(", ")", ":", "service_model", "=", "apps", ".", "get_model", "(", "service", "[", "'model_name'", "]", ")", "service_project_link", "=", "service_model", ".", "projects", ".", "through", "data", "[", "key", "]", "=", "{", "'service'", ":", "service_model", ",", "'service_project_link'", ":", "service_project_link", ",", "'resources'", ":", "[", "apps", ".", "get_model", "(", "r", ")", "for", "r", "in", "service", "[", "'resources'", "]", ".", "keys", "(", ")", "]", ",", "'properties'", ":", "[", "apps", ".", "get_model", "(", "r", ")", "for", "r", "in", "service", "[", "'properties'", "]", ".", "keys", "(", ")", "if", "'.'", "in", "r", "]", ",", "}", "return", "data" ]
Get a list of service models. { ... 'gitlab': { "service": nodeconductor_gitlab.models.GitLabService, "service_project_link": nodeconductor_gitlab.models.GitLabServiceProjectLink, "resources": [ nodeconductor_gitlab.models.Group, nodeconductor_gitlab.models.Project ], }, ... }
[ "Get", "a", "list", "of", "service", "models", ".", "{", "...", "gitlab", ":", "{", "service", ":", "nodeconductor_gitlab", ".", "models", ".", "GitLabService", "service_project_link", ":", "nodeconductor_gitlab", ".", "models", ".", "GitLabServiceProjectLink", "resources", ":", "[", "nodeconductor_gitlab", ".", "models", ".", "Group", "nodeconductor_gitlab", ".", "models", ".", "Project", "]", "}", "...", "}" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L252-L281
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_resource_models
def get_resource_models(cls): """ Get a list of resource models. { 'DigitalOcean.Droplet': waldur_digitalocean.models.Droplet, 'JIRA.Project': waldur_jira.models.Project, 'OpenStack.Tenant': waldur_openstack.models.Tenant } """ from django.apps import apps return {'.'.join([service['name'], attrs['name']]): apps.get_model(resource) for service in cls._registry.values() for resource, attrs in service['resources'].items()}
python
def get_resource_models(cls): """ Get a list of resource models. { 'DigitalOcean.Droplet': waldur_digitalocean.models.Droplet, 'JIRA.Project': waldur_jira.models.Project, 'OpenStack.Tenant': waldur_openstack.models.Tenant } """ from django.apps import apps return {'.'.join([service['name'], attrs['name']]): apps.get_model(resource) for service in cls._registry.values() for resource, attrs in service['resources'].items()}
[ "def", "get_resource_models", "(", "cls", ")", ":", "from", "django", ".", "apps", "import", "apps", "return", "{", "'.'", ".", "join", "(", "[", "service", "[", "'name'", "]", ",", "attrs", "[", "'name'", "]", "]", ")", ":", "apps", ".", "get_model", "(", "resource", ")", "for", "service", "in", "cls", ".", "_registry", ".", "values", "(", ")", "for", "resource", ",", "attrs", "in", "service", "[", "'resources'", "]", ".", "items", "(", ")", "}" ]
Get a list of resource models. { 'DigitalOcean.Droplet': waldur_digitalocean.models.Droplet, 'JIRA.Project': waldur_jira.models.Project, 'OpenStack.Tenant': waldur_openstack.models.Tenant }
[ "Get", "a", "list", "of", "resource", "models", ".", "{", "DigitalOcean", ".", "Droplet", ":", "waldur_digitalocean", ".", "models", ".", "Droplet", "JIRA", ".", "Project", ":", "waldur_jira", ".", "models", ".", "Project", "OpenStack", ".", "Tenant", ":", "waldur_openstack", ".", "models", ".", "Tenant", "}" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L285-L298
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_service_resources
def get_service_resources(cls, model): """ Get resource models by service model """ key = cls.get_model_key(model) return cls.get_service_name_resources(key)
python
def get_service_resources(cls, model): """ Get resource models by service model """ key = cls.get_model_key(model) return cls.get_service_name_resources(key)
[ "def", "get_service_resources", "(", "cls", ",", "model", ")", ":", "key", "=", "cls", ".", "get_model_key", "(", "model", ")", "return", "cls", ".", "get_service_name_resources", "(", "key", ")" ]
Get resource models by service model
[ "Get", "resource", "models", "by", "service", "model" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L302-L305
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_service_name_resources
def get_service_name_resources(cls, service_name): """ Get resource models by service name """ from django.apps import apps resources = cls._registry[service_name]['resources'].keys() return [apps.get_model(resource) for resource in resources]
python
def get_service_name_resources(cls, service_name): """ Get resource models by service name """ from django.apps import apps resources = cls._registry[service_name]['resources'].keys() return [apps.get_model(resource) for resource in resources]
[ "def", "get_service_name_resources", "(", "cls", ",", "service_name", ")", ":", "from", "django", ".", "apps", "import", "apps", "resources", "=", "cls", ".", "_registry", "[", "service_name", "]", "[", "'resources'", "]", ".", "keys", "(", ")", "return", "[", "apps", ".", "get_model", "(", "resource", ")", "for", "resource", "in", "resources", "]" ]
Get resource models by service name
[ "Get", "resource", "models", "by", "service", "name" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L309-L314
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_name_for_model
def get_name_for_model(cls, model): """ Get a name for given class or model: -- it's a service type for a service -- it's a <service_type>.<resource_model_name> for a resource """ key = cls.get_model_key(model) model_str = cls._get_model_str(model) service = cls._registry[key] if model_str in service['resources']: return '{}.{}'.format(service['name'], service['resources'][model_str]['name']) else: return service['name']
python
def get_name_for_model(cls, model): """ Get a name for given class or model: -- it's a service type for a service -- it's a <service_type>.<resource_model_name> for a resource """ key = cls.get_model_key(model) model_str = cls._get_model_str(model) service = cls._registry[key] if model_str in service['resources']: return '{}.{}'.format(service['name'], service['resources'][model_str]['name']) else: return service['name']
[ "def", "get_name_for_model", "(", "cls", ",", "model", ")", ":", "key", "=", "cls", ".", "get_model_key", "(", "model", ")", "model_str", "=", "cls", ".", "_get_model_str", "(", "model", ")", "service", "=", "cls", ".", "_registry", "[", "key", "]", "if", "model_str", "in", "service", "[", "'resources'", "]", ":", "return", "'{}.{}'", ".", "format", "(", "service", "[", "'name'", "]", ",", "service", "[", "'resources'", "]", "[", "model_str", "]", "[", "'name'", "]", ")", "else", ":", "return", "service", "[", "'name'", "]" ]
Get a name for given class or model: -- it's a service type for a service -- it's a <service_type>.<resource_model_name> for a resource
[ "Get", "a", "name", "for", "given", "class", "or", "model", ":", "--", "it", "s", "a", "service", "type", "for", "a", "service", "--", "it", "s", "a", "<service_type", ">", ".", "<resource_model_name", ">", "for", "a", "resource" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L317-L328
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices.get_related_models
def get_related_models(cls, model): """ Get a dictionary with related structure models for given class or model: >> SupportedServices.get_related_models(gitlab_models.Project) { 'service': nodeconductor_gitlab.models.GitLabService, 'service_project_link': nodeconductor_gitlab.models.GitLabServiceProjectLink, 'resources': [ nodeconductor_gitlab.models.Group, nodeconductor_gitlab.models.Project, ] } """ from waldur_core.structure.models import ServiceSettings if isinstance(model, ServiceSettings): model_str = cls._registry.get(model.type, {}).get('model_name', '') else: model_str = cls._get_model_str(model) for models in cls.get_service_models().values(): if model_str == cls._get_model_str(models['service']) or \ model_str == cls._get_model_str(models['service_project_link']): return models for resource_model in models['resources']: if model_str == cls._get_model_str(resource_model): return models
python
def get_related_models(cls, model): """ Get a dictionary with related structure models for given class or model: >> SupportedServices.get_related_models(gitlab_models.Project) { 'service': nodeconductor_gitlab.models.GitLabService, 'service_project_link': nodeconductor_gitlab.models.GitLabServiceProjectLink, 'resources': [ nodeconductor_gitlab.models.Group, nodeconductor_gitlab.models.Project, ] } """ from waldur_core.structure.models import ServiceSettings if isinstance(model, ServiceSettings): model_str = cls._registry.get(model.type, {}).get('model_name', '') else: model_str = cls._get_model_str(model) for models in cls.get_service_models().values(): if model_str == cls._get_model_str(models['service']) or \ model_str == cls._get_model_str(models['service_project_link']): return models for resource_model in models['resources']: if model_str == cls._get_model_str(resource_model): return models
[ "def", "get_related_models", "(", "cls", ",", "model", ")", ":", "from", "waldur_core", ".", "structure", ".", "models", "import", "ServiceSettings", "if", "isinstance", "(", "model", ",", "ServiceSettings", ")", ":", "model_str", "=", "cls", ".", "_registry", ".", "get", "(", "model", ".", "type", ",", "{", "}", ")", ".", "get", "(", "'model_name'", ",", "''", ")", "else", ":", "model_str", "=", "cls", ".", "_get_model_str", "(", "model", ")", "for", "models", "in", "cls", ".", "get_service_models", "(", ")", ".", "values", "(", ")", ":", "if", "model_str", "==", "cls", ".", "_get_model_str", "(", "models", "[", "'service'", "]", ")", "or", "model_str", "==", "cls", ".", "_get_model_str", "(", "models", "[", "'service_project_link'", "]", ")", ":", "return", "models", "for", "resource_model", "in", "models", "[", "'resources'", "]", ":", "if", "model_str", "==", "cls", ".", "_get_model_str", "(", "resource_model", ")", ":", "return", "models" ]
Get a dictionary with related structure models for given class or model: >> SupportedServices.get_related_models(gitlab_models.Project) { 'service': nodeconductor_gitlab.models.GitLabService, 'service_project_link': nodeconductor_gitlab.models.GitLabServiceProjectLink, 'resources': [ nodeconductor_gitlab.models.Group, nodeconductor_gitlab.models.Project, ] }
[ "Get", "a", "dictionary", "with", "related", "structure", "models", "for", "given", "class", "or", "model", ":" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L331-L358
train
opennode/waldur-core
waldur_core/structure/__init__.py
SupportedServices._is_active_model
def _is_active_model(cls, model): """ Check is model app name is in list of INSTALLED_APPS """ # We need to use such tricky way to check because of inconsistent apps names: # some apps are included in format "<module_name>.<app_name>" like "waldur_core.openstack" # other apps are included in format "<app_name>" like "nodecondcutor_sugarcrm" return ('.'.join(model.__module__.split('.')[:2]) in settings.INSTALLED_APPS or '.'.join(model.__module__.split('.')[:1]) in settings.INSTALLED_APPS)
python
def _is_active_model(cls, model): """ Check is model app name is in list of INSTALLED_APPS """ # We need to use such tricky way to check because of inconsistent apps names: # some apps are included in format "<module_name>.<app_name>" like "waldur_core.openstack" # other apps are included in format "<app_name>" like "nodecondcutor_sugarcrm" return ('.'.join(model.__module__.split('.')[:2]) in settings.INSTALLED_APPS or '.'.join(model.__module__.split('.')[:1]) in settings.INSTALLED_APPS)
[ "def", "_is_active_model", "(", "cls", ",", "model", ")", ":", "# We need to use such tricky way to check because of inconsistent apps names:", "# some apps are included in format \"<module_name>.<app_name>\" like \"waldur_core.openstack\"", "# other apps are included in format \"<app_name>\" like \"nodecondcutor_sugarcrm\"", "return", "(", "'.'", ".", "join", "(", "model", ".", "__module__", ".", "split", "(", "'.'", ")", "[", ":", "2", "]", ")", "in", "settings", ".", "INSTALLED_APPS", "or", "'.'", ".", "join", "(", "model", ".", "__module__", ".", "split", "(", "'.'", ")", "[", ":", "1", "]", ")", "in", "settings", ".", "INSTALLED_APPS", ")" ]
Check is model app name is in list of INSTALLED_APPS
[ "Check", "is", "model", "app", "name", "is", "in", "list", "of", "INSTALLED_APPS" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/__init__.py#L361-L367
train
opennode/waldur-core
waldur_core/logging/models.py
PushHook.process
def process(self, event): """ Send events as push notification via Google Cloud Messaging. Expected settings as follows: # https://developers.google.com/mobile/add WALDUR_CORE['GOOGLE_API'] = { 'NOTIFICATION_TITLE': "Waldur notification", 'Android': { 'server_key': 'AIzaSyA2_7UaVIxXfKeFvxTjQNZbrzkXG9OTCkg', }, 'iOS': { 'server_key': 'AIzaSyA34zlG_y5uHOe2FmcJKwfk2vG-3RW05vk', } } """ conf = settings.WALDUR_CORE.get('GOOGLE_API') or {} keys = conf.get(dict(self.Type.CHOICES)[self.type]) if not keys or not self.token: return endpoint = 'https://gcm-http.googleapis.com/gcm/send' headers = { 'Content-Type': 'application/json', 'Authorization': 'key=%s' % keys['server_key'], } payload = { 'to': self.token, 'notification': { 'body': event.get('message', 'New event'), 'title': conf.get('NOTIFICATION_TITLE', 'Waldur notification'), 'image': 'icon', }, 'data': { 'event': event }, } if self.type == self.Type.IOS: payload['content-available'] = '1' logger.debug('Submitting GCM push notification with headers %s, payload: %s' % (headers, payload)) requests.post(endpoint, json=payload, headers=headers)
python
def process(self, event): """ Send events as push notification via Google Cloud Messaging. Expected settings as follows: # https://developers.google.com/mobile/add WALDUR_CORE['GOOGLE_API'] = { 'NOTIFICATION_TITLE': "Waldur notification", 'Android': { 'server_key': 'AIzaSyA2_7UaVIxXfKeFvxTjQNZbrzkXG9OTCkg', }, 'iOS': { 'server_key': 'AIzaSyA34zlG_y5uHOe2FmcJKwfk2vG-3RW05vk', } } """ conf = settings.WALDUR_CORE.get('GOOGLE_API') or {} keys = conf.get(dict(self.Type.CHOICES)[self.type]) if not keys or not self.token: return endpoint = 'https://gcm-http.googleapis.com/gcm/send' headers = { 'Content-Type': 'application/json', 'Authorization': 'key=%s' % keys['server_key'], } payload = { 'to': self.token, 'notification': { 'body': event.get('message', 'New event'), 'title': conf.get('NOTIFICATION_TITLE', 'Waldur notification'), 'image': 'icon', }, 'data': { 'event': event }, } if self.type == self.Type.IOS: payload['content-available'] = '1' logger.debug('Submitting GCM push notification with headers %s, payload: %s' % (headers, payload)) requests.post(endpoint, json=payload, headers=headers)
[ "def", "process", "(", "self", ",", "event", ")", ":", "conf", "=", "settings", ".", "WALDUR_CORE", ".", "get", "(", "'GOOGLE_API'", ")", "or", "{", "}", "keys", "=", "conf", ".", "get", "(", "dict", "(", "self", ".", "Type", ".", "CHOICES", ")", "[", "self", ".", "type", "]", ")", "if", "not", "keys", "or", "not", "self", ".", "token", ":", "return", "endpoint", "=", "'https://gcm-http.googleapis.com/gcm/send'", "headers", "=", "{", "'Content-Type'", ":", "'application/json'", ",", "'Authorization'", ":", "'key=%s'", "%", "keys", "[", "'server_key'", "]", ",", "}", "payload", "=", "{", "'to'", ":", "self", ".", "token", ",", "'notification'", ":", "{", "'body'", ":", "event", ".", "get", "(", "'message'", ",", "'New event'", ")", ",", "'title'", ":", "conf", ".", "get", "(", "'NOTIFICATION_TITLE'", ",", "'Waldur notification'", ")", ",", "'image'", ":", "'icon'", ",", "}", ",", "'data'", ":", "{", "'event'", ":", "event", "}", ",", "}", "if", "self", ".", "type", "==", "self", ".", "Type", ".", "IOS", ":", "payload", "[", "'content-available'", "]", "=", "'1'", "logger", ".", "debug", "(", "'Submitting GCM push notification with headers %s, payload: %s'", "%", "(", "headers", ",", "payload", ")", ")", "requests", ".", "post", "(", "endpoint", ",", "json", "=", "payload", ",", "headers", "=", "headers", ")" ]
Send events as push notification via Google Cloud Messaging. Expected settings as follows: # https://developers.google.com/mobile/add WALDUR_CORE['GOOGLE_API'] = { 'NOTIFICATION_TITLE': "Waldur notification", 'Android': { 'server_key': 'AIzaSyA2_7UaVIxXfKeFvxTjQNZbrzkXG9OTCkg', }, 'iOS': { 'server_key': 'AIzaSyA34zlG_y5uHOe2FmcJKwfk2vG-3RW05vk', } }
[ "Send", "events", "as", "push", "notification", "via", "Google", "Cloud", "Messaging", ".", "Expected", "settings", "as", "follows", ":" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/logging/models.py#L200-L241
train
stepank/pyws
src/pyws/protocols/soap/__init__.py
get_context_data_from_headers
def get_context_data_from_headers(request, headers_schema): """ Extracts context data from request headers according to specified schema. >>> from lxml import etree as et >>> from datetime import date >>> from pyws.functions.args import TypeFactory >>> Fake = type('Fake', (object, ), {}) >>> request = Fake() >>> request.parsed_data = Fake() >>> request.parsed_data.xml = et.fromstring( ... '<s:Envelope xmlns:s="http://schemas.xmlsoap.org/soap/envelope/">' ... '<s:Header>' ... '<headers>' ... '<string>hello</string>' ... '<number>100</number>' ... '<date>2011-08-12</date>' ... '</headers>' ... '</s:Header>' ... '</s:Envelope>') >>> data = get_context_data_from_headers(request, TypeFactory( ... {0: 'Headers', 'string': str, 'number': int, 'date': date})) >>> data == {'string': 'hello', 'number': 100, 'date': date(2011, 8, 12)} True """ if not headers_schema: return None env = request.parsed_data.xml.xpath( '/soap:Envelope', namespaces=SoapProtocol.namespaces)[0] header = env.xpath( './soap:Header/*', namespaces=SoapProtocol.namespaces) if len(header) < 1: return None return headers_schema.validate(xml2obj(header[0], headers_schema))
python
def get_context_data_from_headers(request, headers_schema): """ Extracts context data from request headers according to specified schema. >>> from lxml import etree as et >>> from datetime import date >>> from pyws.functions.args import TypeFactory >>> Fake = type('Fake', (object, ), {}) >>> request = Fake() >>> request.parsed_data = Fake() >>> request.parsed_data.xml = et.fromstring( ... '<s:Envelope xmlns:s="http://schemas.xmlsoap.org/soap/envelope/">' ... '<s:Header>' ... '<headers>' ... '<string>hello</string>' ... '<number>100</number>' ... '<date>2011-08-12</date>' ... '</headers>' ... '</s:Header>' ... '</s:Envelope>') >>> data = get_context_data_from_headers(request, TypeFactory( ... {0: 'Headers', 'string': str, 'number': int, 'date': date})) >>> data == {'string': 'hello', 'number': 100, 'date': date(2011, 8, 12)} True """ if not headers_schema: return None env = request.parsed_data.xml.xpath( '/soap:Envelope', namespaces=SoapProtocol.namespaces)[0] header = env.xpath( './soap:Header/*', namespaces=SoapProtocol.namespaces) if len(header) < 1: return None return headers_schema.validate(xml2obj(header[0], headers_schema))
[ "def", "get_context_data_from_headers", "(", "request", ",", "headers_schema", ")", ":", "if", "not", "headers_schema", ":", "return", "None", "env", "=", "request", ".", "parsed_data", ".", "xml", ".", "xpath", "(", "'/soap:Envelope'", ",", "namespaces", "=", "SoapProtocol", ".", "namespaces", ")", "[", "0", "]", "header", "=", "env", ".", "xpath", "(", "'./soap:Header/*'", ",", "namespaces", "=", "SoapProtocol", ".", "namespaces", ")", "if", "len", "(", "header", ")", "<", "1", ":", "return", "None", "return", "headers_schema", ".", "validate", "(", "xml2obj", "(", "header", "[", "0", "]", ",", "headers_schema", ")", ")" ]
Extracts context data from request headers according to specified schema. >>> from lxml import etree as et >>> from datetime import date >>> from pyws.functions.args import TypeFactory >>> Fake = type('Fake', (object, ), {}) >>> request = Fake() >>> request.parsed_data = Fake() >>> request.parsed_data.xml = et.fromstring( ... '<s:Envelope xmlns:s="http://schemas.xmlsoap.org/soap/envelope/">' ... '<s:Header>' ... '<headers>' ... '<string>hello</string>' ... '<number>100</number>' ... '<date>2011-08-12</date>' ... '</headers>' ... '</s:Header>' ... '</s:Envelope>') >>> data = get_context_data_from_headers(request, TypeFactory( ... {0: 'Headers', 'string': str, 'number': int, 'date': date})) >>> data == {'string': 'hello', 'number': 100, 'date': date(2011, 8, 12)} True
[ "Extracts", "context", "data", "from", "request", "headers", "according", "to", "specified", "schema", "." ]
ff39133aabeb56bbb08d66286ac0cc8731eda7dd
https://github.com/stepank/pyws/blob/ff39133aabeb56bbb08d66286ac0cc8731eda7dd/src/pyws/protocols/soap/__init__.py#L94-L131
train
quora/qcore
qcore/caching.py
lazy_constant
def lazy_constant(fn): """Decorator to make a function that takes no arguments use the LazyConstant class.""" class NewLazyConstant(LazyConstant): @functools.wraps(fn) def __call__(self): return self.get_value() return NewLazyConstant(fn)
python
def lazy_constant(fn): """Decorator to make a function that takes no arguments use the LazyConstant class.""" class NewLazyConstant(LazyConstant): @functools.wraps(fn) def __call__(self): return self.get_value() return NewLazyConstant(fn)
[ "def", "lazy_constant", "(", "fn", ")", ":", "class", "NewLazyConstant", "(", "LazyConstant", ")", ":", "@", "functools", ".", "wraps", "(", "fn", ")", "def", "__call__", "(", "self", ")", ":", "return", "self", ".", "get_value", "(", ")", "return", "NewLazyConstant", "(", "fn", ")" ]
Decorator to make a function that takes no arguments use the LazyConstant class.
[ "Decorator", "to", "make", "a", "function", "that", "takes", "no", "arguments", "use", "the", "LazyConstant", "class", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L84-L92
train
quora/qcore
qcore/caching.py
lru_cache
def lru_cache(maxsize=128, key_fn=None): """Decorator that adds an LRU cache of size maxsize to the decorated function. maxsize is the number of different keys cache can accomodate. key_fn is the function that builds key from args. The default key function creates a tuple out of args and kwargs. If you use the default, there is no reason not to use functools.lru_cache directly. Possible use cases: - Your cache key is very large, so you don't want to keep the whole key in memory. - The function takes some arguments that don't affect the result. """ def decorator(fn): cache = LRUCache(maxsize) argspec = inspect2.getfullargspec(fn) arg_names = argspec.args[1:] + argspec.kwonlyargs # remove self kwargs_defaults = get_kwargs_defaults(argspec) cache_key = key_fn if cache_key is None: def cache_key(args, kwargs): return get_args_tuple(args, kwargs, arg_names, kwargs_defaults) @functools.wraps(fn) def wrapper(*args, **kwargs): key = cache_key(args, kwargs) try: return cache[key] except KeyError: value = fn(*args, **kwargs) cache[key] = value return value wrapper.clear = cache.clear return wrapper return decorator
python
def lru_cache(maxsize=128, key_fn=None): """Decorator that adds an LRU cache of size maxsize to the decorated function. maxsize is the number of different keys cache can accomodate. key_fn is the function that builds key from args. The default key function creates a tuple out of args and kwargs. If you use the default, there is no reason not to use functools.lru_cache directly. Possible use cases: - Your cache key is very large, so you don't want to keep the whole key in memory. - The function takes some arguments that don't affect the result. """ def decorator(fn): cache = LRUCache(maxsize) argspec = inspect2.getfullargspec(fn) arg_names = argspec.args[1:] + argspec.kwonlyargs # remove self kwargs_defaults = get_kwargs_defaults(argspec) cache_key = key_fn if cache_key is None: def cache_key(args, kwargs): return get_args_tuple(args, kwargs, arg_names, kwargs_defaults) @functools.wraps(fn) def wrapper(*args, **kwargs): key = cache_key(args, kwargs) try: return cache[key] except KeyError: value = fn(*args, **kwargs) cache[key] = value return value wrapper.clear = cache.clear return wrapper return decorator
[ "def", "lru_cache", "(", "maxsize", "=", "128", ",", "key_fn", "=", "None", ")", ":", "def", "decorator", "(", "fn", ")", ":", "cache", "=", "LRUCache", "(", "maxsize", ")", "argspec", "=", "inspect2", ".", "getfullargspec", "(", "fn", ")", "arg_names", "=", "argspec", ".", "args", "[", "1", ":", "]", "+", "argspec", ".", "kwonlyargs", "# remove self", "kwargs_defaults", "=", "get_kwargs_defaults", "(", "argspec", ")", "cache_key", "=", "key_fn", "if", "cache_key", "is", "None", ":", "def", "cache_key", "(", "args", ",", "kwargs", ")", ":", "return", "get_args_tuple", "(", "args", ",", "kwargs", ",", "arg_names", ",", "kwargs_defaults", ")", "@", "functools", ".", "wraps", "(", "fn", ")", "def", "wrapper", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "key", "=", "cache_key", "(", "args", ",", "kwargs", ")", "try", ":", "return", "cache", "[", "key", "]", "except", "KeyError", ":", "value", "=", "fn", "(", "*", "args", ",", "*", "*", "kwargs", ")", "cache", "[", "key", "]", "=", "value", "return", "value", "wrapper", ".", "clear", "=", "cache", ".", "clear", "return", "wrapper", "return", "decorator" ]
Decorator that adds an LRU cache of size maxsize to the decorated function. maxsize is the number of different keys cache can accomodate. key_fn is the function that builds key from args. The default key function creates a tuple out of args and kwargs. If you use the default, there is no reason not to use functools.lru_cache directly. Possible use cases: - Your cache key is very large, so you don't want to keep the whole key in memory. - The function takes some arguments that don't affect the result.
[ "Decorator", "that", "adds", "an", "LRU", "cache", "of", "size", "maxsize", "to", "the", "decorated", "function", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L238-L278
train
quora/qcore
qcore/caching.py
cached_per_instance
def cached_per_instance(): """Decorator that adds caching to an instance method. The cached value is stored so that it gets garbage collected together with the instance. The cached values are not stored when the object is pickled. """ def cache_fun(fun): argspec = inspect2.getfullargspec(fun) arg_names = argspec.args[1:] + argspec.kwonlyargs # remove self kwargs_defaults = get_kwargs_defaults(argspec) cache = {} def cache_key(args, kwargs): return get_args_tuple(args, kwargs, arg_names, kwargs_defaults) def clear_cache(instance_key, ref): del cache[instance_key] @functools.wraps(fun) def new_fun(self, *args, **kwargs): instance_key = id(self) if instance_key not in cache: ref = weakref.ref(self, functools.partial(clear_cache, instance_key)) cache[instance_key] = (ref, {}) instance_cache = cache[instance_key][1] k = cache_key(args, kwargs) if k not in instance_cache: instance_cache[k] = fun(self, *args, **kwargs) return instance_cache[k] # just so unit tests can check that this is cleaned up correctly new_fun.__cached_per_instance_cache__ = cache return new_fun return cache_fun
python
def cached_per_instance(): """Decorator that adds caching to an instance method. The cached value is stored so that it gets garbage collected together with the instance. The cached values are not stored when the object is pickled. """ def cache_fun(fun): argspec = inspect2.getfullargspec(fun) arg_names = argspec.args[1:] + argspec.kwonlyargs # remove self kwargs_defaults = get_kwargs_defaults(argspec) cache = {} def cache_key(args, kwargs): return get_args_tuple(args, kwargs, arg_names, kwargs_defaults) def clear_cache(instance_key, ref): del cache[instance_key] @functools.wraps(fun) def new_fun(self, *args, **kwargs): instance_key = id(self) if instance_key not in cache: ref = weakref.ref(self, functools.partial(clear_cache, instance_key)) cache[instance_key] = (ref, {}) instance_cache = cache[instance_key][1] k = cache_key(args, kwargs) if k not in instance_cache: instance_cache[k] = fun(self, *args, **kwargs) return instance_cache[k] # just so unit tests can check that this is cleaned up correctly new_fun.__cached_per_instance_cache__ = cache return new_fun return cache_fun
[ "def", "cached_per_instance", "(", ")", ":", "def", "cache_fun", "(", "fun", ")", ":", "argspec", "=", "inspect2", ".", "getfullargspec", "(", "fun", ")", "arg_names", "=", "argspec", ".", "args", "[", "1", ":", "]", "+", "argspec", ".", "kwonlyargs", "# remove self", "kwargs_defaults", "=", "get_kwargs_defaults", "(", "argspec", ")", "cache", "=", "{", "}", "def", "cache_key", "(", "args", ",", "kwargs", ")", ":", "return", "get_args_tuple", "(", "args", ",", "kwargs", ",", "arg_names", ",", "kwargs_defaults", ")", "def", "clear_cache", "(", "instance_key", ",", "ref", ")", ":", "del", "cache", "[", "instance_key", "]", "@", "functools", ".", "wraps", "(", "fun", ")", "def", "new_fun", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "instance_key", "=", "id", "(", "self", ")", "if", "instance_key", "not", "in", "cache", ":", "ref", "=", "weakref", ".", "ref", "(", "self", ",", "functools", ".", "partial", "(", "clear_cache", ",", "instance_key", ")", ")", "cache", "[", "instance_key", "]", "=", "(", "ref", ",", "{", "}", ")", "instance_cache", "=", "cache", "[", "instance_key", "]", "[", "1", "]", "k", "=", "cache_key", "(", "args", ",", "kwargs", ")", "if", "k", "not", "in", "instance_cache", ":", "instance_cache", "[", "k", "]", "=", "fun", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", "return", "instance_cache", "[", "k", "]", "# just so unit tests can check that this is cleaned up correctly", "new_fun", ".", "__cached_per_instance_cache__", "=", "cache", "return", "new_fun", "return", "cache_fun" ]
Decorator that adds caching to an instance method. The cached value is stored so that it gets garbage collected together with the instance. The cached values are not stored when the object is pickled.
[ "Decorator", "that", "adds", "caching", "to", "an", "instance", "method", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L281-L319
train
quora/qcore
qcore/caching.py
get_args_tuple
def get_args_tuple(args, kwargs, arg_names, kwargs_defaults): """Generates a cache key from the passed in arguments.""" args_list = list(args) args_len = len(args) all_args_len = len(arg_names) try: while args_len < all_args_len: arg_name = arg_names[args_len] if arg_name in kwargs_defaults: args_list.append(kwargs.get(arg_name, kwargs_defaults[arg_name])) else: args_list.append(kwargs[arg_name]) args_len += 1 except KeyError as e: raise TypeError("Missing argument %r" % (e.args[0],)) return tuple(args_list)
python
def get_args_tuple(args, kwargs, arg_names, kwargs_defaults): """Generates a cache key from the passed in arguments.""" args_list = list(args) args_len = len(args) all_args_len = len(arg_names) try: while args_len < all_args_len: arg_name = arg_names[args_len] if arg_name in kwargs_defaults: args_list.append(kwargs.get(arg_name, kwargs_defaults[arg_name])) else: args_list.append(kwargs[arg_name]) args_len += 1 except KeyError as e: raise TypeError("Missing argument %r" % (e.args[0],)) return tuple(args_list)
[ "def", "get_args_tuple", "(", "args", ",", "kwargs", ",", "arg_names", ",", "kwargs_defaults", ")", ":", "args_list", "=", "list", "(", "args", ")", "args_len", "=", "len", "(", "args", ")", "all_args_len", "=", "len", "(", "arg_names", ")", "try", ":", "while", "args_len", "<", "all_args_len", ":", "arg_name", "=", "arg_names", "[", "args_len", "]", "if", "arg_name", "in", "kwargs_defaults", ":", "args_list", ".", "append", "(", "kwargs", ".", "get", "(", "arg_name", ",", "kwargs_defaults", "[", "arg_name", "]", ")", ")", "else", ":", "args_list", ".", "append", "(", "kwargs", "[", "arg_name", "]", ")", "args_len", "+=", "1", "except", "KeyError", "as", "e", ":", "raise", "TypeError", "(", "\"Missing argument %r\"", "%", "(", "e", ".", "args", "[", "0", "]", ",", ")", ")", "return", "tuple", "(", "args_list", ")" ]
Generates a cache key from the passed in arguments.
[ "Generates", "a", "cache", "key", "from", "the", "passed", "in", "arguments", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L322-L337
train
quora/qcore
qcore/caching.py
get_kwargs_defaults
def get_kwargs_defaults(argspec): """Computes a kwargs_defaults dictionary for use by get_args_tuple given an argspec.""" arg_names = tuple(argspec.args) defaults = argspec.defaults or () num_args = len(argspec.args) - len(defaults) kwargs_defaults = {} for i, default_value in enumerate(defaults): kwargs_defaults[arg_names[num_args + i]] = default_value if getattr(argspec, "kwonlydefaults", None): kwargs_defaults.update(argspec.kwonlydefaults) return kwargs_defaults
python
def get_kwargs_defaults(argspec): """Computes a kwargs_defaults dictionary for use by get_args_tuple given an argspec.""" arg_names = tuple(argspec.args) defaults = argspec.defaults or () num_args = len(argspec.args) - len(defaults) kwargs_defaults = {} for i, default_value in enumerate(defaults): kwargs_defaults[arg_names[num_args + i]] = default_value if getattr(argspec, "kwonlydefaults", None): kwargs_defaults.update(argspec.kwonlydefaults) return kwargs_defaults
[ "def", "get_kwargs_defaults", "(", "argspec", ")", ":", "arg_names", "=", "tuple", "(", "argspec", ".", "args", ")", "defaults", "=", "argspec", ".", "defaults", "or", "(", ")", "num_args", "=", "len", "(", "argspec", ".", "args", ")", "-", "len", "(", "defaults", ")", "kwargs_defaults", "=", "{", "}", "for", "i", ",", "default_value", "in", "enumerate", "(", "defaults", ")", ":", "kwargs_defaults", "[", "arg_names", "[", "num_args", "+", "i", "]", "]", "=", "default_value", "if", "getattr", "(", "argspec", ",", "\"kwonlydefaults\"", ",", "None", ")", ":", "kwargs_defaults", ".", "update", "(", "argspec", ".", "kwonlydefaults", ")", "return", "kwargs_defaults" ]
Computes a kwargs_defaults dictionary for use by get_args_tuple given an argspec.
[ "Computes", "a", "kwargs_defaults", "dictionary", "for", "use", "by", "get_args_tuple", "given", "an", "argspec", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L340-L350
train
quora/qcore
qcore/caching.py
memoize
def memoize(fun): """Memoizes return values of the decorated function. Similar to l0cache, but the cache persists for the duration of the process, unless clear_cache() is called on the function. """ argspec = inspect2.getfullargspec(fun) arg_names = argspec.args + argspec.kwonlyargs kwargs_defaults = get_kwargs_defaults(argspec) def cache_key(args, kwargs): return get_args_tuple(args, kwargs, arg_names, kwargs_defaults) @functools.wraps(fun) def new_fun(*args, **kwargs): k = cache_key(args, kwargs) if k not in new_fun.__cache: new_fun.__cache[k] = fun(*args, **kwargs) return new_fun.__cache[k] def clear_cache(): """Removes all cached values for this function.""" new_fun.__cache.clear() new_fun.__cache = {} new_fun.clear_cache = clear_cache return new_fun
python
def memoize(fun): """Memoizes return values of the decorated function. Similar to l0cache, but the cache persists for the duration of the process, unless clear_cache() is called on the function. """ argspec = inspect2.getfullargspec(fun) arg_names = argspec.args + argspec.kwonlyargs kwargs_defaults = get_kwargs_defaults(argspec) def cache_key(args, kwargs): return get_args_tuple(args, kwargs, arg_names, kwargs_defaults) @functools.wraps(fun) def new_fun(*args, **kwargs): k = cache_key(args, kwargs) if k not in new_fun.__cache: new_fun.__cache[k] = fun(*args, **kwargs) return new_fun.__cache[k] def clear_cache(): """Removes all cached values for this function.""" new_fun.__cache.clear() new_fun.__cache = {} new_fun.clear_cache = clear_cache return new_fun
[ "def", "memoize", "(", "fun", ")", ":", "argspec", "=", "inspect2", ".", "getfullargspec", "(", "fun", ")", "arg_names", "=", "argspec", ".", "args", "+", "argspec", ".", "kwonlyargs", "kwargs_defaults", "=", "get_kwargs_defaults", "(", "argspec", ")", "def", "cache_key", "(", "args", ",", "kwargs", ")", ":", "return", "get_args_tuple", "(", "args", ",", "kwargs", ",", "arg_names", ",", "kwargs_defaults", ")", "@", "functools", ".", "wraps", "(", "fun", ")", "def", "new_fun", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "k", "=", "cache_key", "(", "args", ",", "kwargs", ")", "if", "k", "not", "in", "new_fun", ".", "__cache", ":", "new_fun", ".", "__cache", "[", "k", "]", "=", "fun", "(", "*", "args", ",", "*", "*", "kwargs", ")", "return", "new_fun", ".", "__cache", "[", "k", "]", "def", "clear_cache", "(", ")", ":", "\"\"\"Removes all cached values for this function.\"\"\"", "new_fun", ".", "__cache", ".", "clear", "(", ")", "new_fun", ".", "__cache", "=", "{", "}", "new_fun", ".", "clear_cache", "=", "clear_cache", "return", "new_fun" ]
Memoizes return values of the decorated function. Similar to l0cache, but the cache persists for the duration of the process, unless clear_cache() is called on the function.
[ "Memoizes", "return", "values", "of", "the", "decorated", "function", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L353-L380
train
quora/qcore
qcore/caching.py
memoize_with_ttl
def memoize_with_ttl(ttl_secs=60 * 60 * 24): """Memoizes return values of the decorated function for a given time-to-live. Similar to l0cache, but the cache persists for the duration of the process, unless clear_cache() is called on the function or the time-to-live expires. By default, the time-to-live is set to 24 hours. """ error_msg = ( "Incorrect usage of qcore.caching.memoize_with_ttl: " "ttl_secs must be a positive integer." ) assert_is_instance(ttl_secs, six.integer_types, error_msg) assert_gt(ttl_secs, 0, error_msg) def cache_fun(fun): argspec = inspect2.getfullargspec(fun) arg_names = argspec.args + argspec.kwonlyargs kwargs_defaults = get_kwargs_defaults(argspec) def cache_key(args, kwargs): return repr(get_args_tuple(args, kwargs, arg_names, kwargs_defaults)) @functools.wraps(fun) def new_fun(*args, **kwargs): k = cache_key(args, kwargs) current_time = int(time.time()) # k is not in the cache; perform the function and cache the result. if k not in new_fun.__cache or k not in new_fun.__cache_times: new_fun.__cache[k] = fun(*args, **kwargs) new_fun.__cache_times[k] = current_time return new_fun.__cache[k] # k is in the cache at this point. Check if the ttl has expired; # if so, recompute the value and cache it. cache_time = new_fun.__cache_times[k] if current_time - cache_time > ttl_secs: new_fun.__cache[k] = fun(*args, **kwargs) new_fun.__cache_times[k] = current_time # finally, return the cached result. return new_fun.__cache[k] def clear_cache(): """Removes all cached values for this function.""" new_fun.__cache.clear() new_fun.__cache_times.clear() def dirty(*args, **kwargs): """Dirties the function for a given set of arguments.""" k = cache_key(args, kwargs) new_fun.__cache.pop(k, None) new_fun.__cache_times.pop(k, None) new_fun.__cache = {} new_fun.__cache_times = {} new_fun.clear_cache = clear_cache new_fun.dirty = dirty return new_fun return cache_fun
python
def memoize_with_ttl(ttl_secs=60 * 60 * 24): """Memoizes return values of the decorated function for a given time-to-live. Similar to l0cache, but the cache persists for the duration of the process, unless clear_cache() is called on the function or the time-to-live expires. By default, the time-to-live is set to 24 hours. """ error_msg = ( "Incorrect usage of qcore.caching.memoize_with_ttl: " "ttl_secs must be a positive integer." ) assert_is_instance(ttl_secs, six.integer_types, error_msg) assert_gt(ttl_secs, 0, error_msg) def cache_fun(fun): argspec = inspect2.getfullargspec(fun) arg_names = argspec.args + argspec.kwonlyargs kwargs_defaults = get_kwargs_defaults(argspec) def cache_key(args, kwargs): return repr(get_args_tuple(args, kwargs, arg_names, kwargs_defaults)) @functools.wraps(fun) def new_fun(*args, **kwargs): k = cache_key(args, kwargs) current_time = int(time.time()) # k is not in the cache; perform the function and cache the result. if k not in new_fun.__cache or k not in new_fun.__cache_times: new_fun.__cache[k] = fun(*args, **kwargs) new_fun.__cache_times[k] = current_time return new_fun.__cache[k] # k is in the cache at this point. Check if the ttl has expired; # if so, recompute the value and cache it. cache_time = new_fun.__cache_times[k] if current_time - cache_time > ttl_secs: new_fun.__cache[k] = fun(*args, **kwargs) new_fun.__cache_times[k] = current_time # finally, return the cached result. return new_fun.__cache[k] def clear_cache(): """Removes all cached values for this function.""" new_fun.__cache.clear() new_fun.__cache_times.clear() def dirty(*args, **kwargs): """Dirties the function for a given set of arguments.""" k = cache_key(args, kwargs) new_fun.__cache.pop(k, None) new_fun.__cache_times.pop(k, None) new_fun.__cache = {} new_fun.__cache_times = {} new_fun.clear_cache = clear_cache new_fun.dirty = dirty return new_fun return cache_fun
[ "def", "memoize_with_ttl", "(", "ttl_secs", "=", "60", "*", "60", "*", "24", ")", ":", "error_msg", "=", "(", "\"Incorrect usage of qcore.caching.memoize_with_ttl: \"", "\"ttl_secs must be a positive integer.\"", ")", "assert_is_instance", "(", "ttl_secs", ",", "six", ".", "integer_types", ",", "error_msg", ")", "assert_gt", "(", "ttl_secs", ",", "0", ",", "error_msg", ")", "def", "cache_fun", "(", "fun", ")", ":", "argspec", "=", "inspect2", ".", "getfullargspec", "(", "fun", ")", "arg_names", "=", "argspec", ".", "args", "+", "argspec", ".", "kwonlyargs", "kwargs_defaults", "=", "get_kwargs_defaults", "(", "argspec", ")", "def", "cache_key", "(", "args", ",", "kwargs", ")", ":", "return", "repr", "(", "get_args_tuple", "(", "args", ",", "kwargs", ",", "arg_names", ",", "kwargs_defaults", ")", ")", "@", "functools", ".", "wraps", "(", "fun", ")", "def", "new_fun", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "k", "=", "cache_key", "(", "args", ",", "kwargs", ")", "current_time", "=", "int", "(", "time", ".", "time", "(", ")", ")", "# k is not in the cache; perform the function and cache the result.", "if", "k", "not", "in", "new_fun", ".", "__cache", "or", "k", "not", "in", "new_fun", ".", "__cache_times", ":", "new_fun", ".", "__cache", "[", "k", "]", "=", "fun", "(", "*", "args", ",", "*", "*", "kwargs", ")", "new_fun", ".", "__cache_times", "[", "k", "]", "=", "current_time", "return", "new_fun", ".", "__cache", "[", "k", "]", "# k is in the cache at this point. Check if the ttl has expired;", "# if so, recompute the value and cache it.", "cache_time", "=", "new_fun", ".", "__cache_times", "[", "k", "]", "if", "current_time", "-", "cache_time", ">", "ttl_secs", ":", "new_fun", ".", "__cache", "[", "k", "]", "=", "fun", "(", "*", "args", ",", "*", "*", "kwargs", ")", "new_fun", ".", "__cache_times", "[", "k", "]", "=", "current_time", "# finally, return the cached result.", "return", "new_fun", ".", "__cache", "[", "k", "]", "def", "clear_cache", "(", ")", ":", "\"\"\"Removes all cached values for this function.\"\"\"", "new_fun", ".", "__cache", ".", "clear", "(", ")", "new_fun", ".", "__cache_times", ".", "clear", "(", ")", "def", "dirty", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "\"\"\"Dirties the function for a given set of arguments.\"\"\"", "k", "=", "cache_key", "(", "args", ",", "kwargs", ")", "new_fun", ".", "__cache", ".", "pop", "(", "k", ",", "None", ")", "new_fun", ".", "__cache_times", ".", "pop", "(", "k", ",", "None", ")", "new_fun", ".", "__cache", "=", "{", "}", "new_fun", ".", "__cache_times", "=", "{", "}", "new_fun", ".", "clear_cache", "=", "clear_cache", "new_fun", ".", "dirty", "=", "dirty", "return", "new_fun", "return", "cache_fun" ]
Memoizes return values of the decorated function for a given time-to-live. Similar to l0cache, but the cache persists for the duration of the process, unless clear_cache() is called on the function or the time-to-live expires. By default, the time-to-live is set to 24 hours.
[ "Memoizes", "return", "values", "of", "the", "decorated", "function", "for", "a", "given", "time", "-", "to", "-", "live", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L383-L445
train
quora/qcore
qcore/caching.py
LazyConstant.get_value
def get_value(self): """Returns the value of the constant.""" if self.value is not_computed: self.value = self.value_provider() if self.value is not_computed: return None return self.value
python
def get_value(self): """Returns the value of the constant.""" if self.value is not_computed: self.value = self.value_provider() if self.value is not_computed: return None return self.value
[ "def", "get_value", "(", "self", ")", ":", "if", "self", ".", "value", "is", "not_computed", ":", "self", ".", "value", "=", "self", ".", "value_provider", "(", ")", "if", "self", ".", "value", "is", "not_computed", ":", "return", "None", "return", "self", ".", "value" ]
Returns the value of the constant.
[ "Returns", "the", "value", "of", "the", "constant", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L64-L70
train
quora/qcore
qcore/caching.py
LazyConstant.compute
def compute(self): """Computes the value. Does not look at the cache.""" self.value = self.value_provider() if self.value is not_computed: return None else: return self.value
python
def compute(self): """Computes the value. Does not look at the cache.""" self.value = self.value_provider() if self.value is not_computed: return None else: return self.value
[ "def", "compute", "(", "self", ")", ":", "self", ".", "value", "=", "self", ".", "value_provider", "(", ")", "if", "self", ".", "value", "is", "not_computed", ":", "return", "None", "else", ":", "return", "self", ".", "value" ]
Computes the value. Does not look at the cache.
[ "Computes", "the", "value", ".", "Does", "not", "look", "at", "the", "cache", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L72-L78
train
quora/qcore
qcore/caching.py
LRUCache.get
def get(self, key, default=miss): """Return the value for given key if it exists.""" if key not in self._dict: return default # invokes __getitem__, which updates the item return self[key]
python
def get(self, key, default=miss): """Return the value for given key if it exists.""" if key not in self._dict: return default # invokes __getitem__, which updates the item return self[key]
[ "def", "get", "(", "self", ",", "key", ",", "default", "=", "miss", ")", ":", "if", "key", "not", "in", "self", ".", "_dict", ":", "return", "default", "# invokes __getitem__, which updates the item", "return", "self", "[", "key", "]" ]
Return the value for given key if it exists.
[ "Return", "the", "value", "for", "given", "key", "if", "it", "exists", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L187-L193
train
quora/qcore
qcore/caching.py
LRUCache.clear
def clear(self, omit_item_evicted=False): """Empty the cache and optionally invoke item_evicted callback.""" if not omit_item_evicted: items = self._dict.items() for key, value in items: self._evict_item(key, value) self._dict.clear()
python
def clear(self, omit_item_evicted=False): """Empty the cache and optionally invoke item_evicted callback.""" if not omit_item_evicted: items = self._dict.items() for key, value in items: self._evict_item(key, value) self._dict.clear()
[ "def", "clear", "(", "self", ",", "omit_item_evicted", "=", "False", ")", ":", "if", "not", "omit_item_evicted", ":", "items", "=", "self", ".", "_dict", ".", "items", "(", ")", "for", "key", ",", "value", "in", "items", ":", "self", ".", "_evict_item", "(", "key", ",", "value", ")", "self", ".", "_dict", ".", "clear", "(", ")" ]
Empty the cache and optionally invoke item_evicted callback.
[ "Empty", "the", "cache", "and", "optionally", "invoke", "item_evicted", "callback", "." ]
fa5cd438eea554db35fd29cbc8dfbde69f09961c
https://github.com/quora/qcore/blob/fa5cd438eea554db35fd29cbc8dfbde69f09961c/qcore/caching.py#L209-L215
train
zeromake/aiosqlite3
aiosqlite3/utils.py
create_future
def create_future(loop): # pragma: no cover """Compatibility wrapper for the loop.create_future() call introduced in 3.5.2.""" if hasattr(loop, 'create_future'): return loop.create_future() return asyncio.Future(loop=loop)
python
def create_future(loop): # pragma: no cover """Compatibility wrapper for the loop.create_future() call introduced in 3.5.2.""" if hasattr(loop, 'create_future'): return loop.create_future() return asyncio.Future(loop=loop)
[ "def", "create_future", "(", "loop", ")", ":", "# pragma: no cover", "if", "hasattr", "(", "loop", ",", "'create_future'", ")", ":", "return", "loop", ".", "create_future", "(", ")", "return", "asyncio", ".", "Future", "(", "loop", "=", "loop", ")" ]
Compatibility wrapper for the loop.create_future() call introduced in 3.5.2.
[ "Compatibility", "wrapper", "for", "the", "loop", ".", "create_future", "()", "call", "introduced", "in", "3", ".", "5", ".", "2", "." ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/utils.py#L17-L23
train
zeromake/aiosqlite3
aiosqlite3/utils.py
create_task
def create_task(coro, loop): # pragma: no cover """Compatibility wrapper for the loop.create_task() call introduced in 3.4.2.""" if hasattr(loop, 'create_task'): return loop.create_task(coro) return asyncio.Task(coro, loop=loop)
python
def create_task(coro, loop): # pragma: no cover """Compatibility wrapper for the loop.create_task() call introduced in 3.4.2.""" if hasattr(loop, 'create_task'): return loop.create_task(coro) return asyncio.Task(coro, loop=loop)
[ "def", "create_task", "(", "coro", ",", "loop", ")", ":", "# pragma: no cover", "if", "hasattr", "(", "loop", ",", "'create_task'", ")", ":", "return", "loop", ".", "create_task", "(", "coro", ")", "return", "asyncio", ".", "Task", "(", "coro", ",", "loop", "=", "loop", ")" ]
Compatibility wrapper for the loop.create_task() call introduced in 3.4.2.
[ "Compatibility", "wrapper", "for", "the", "loop", ".", "create_task", "()", "call", "introduced", "in", "3", ".", "4", ".", "2", "." ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/utils.py#L26-L32
train
zeromake/aiosqlite3
aiosqlite3/utils.py
proxy_property_directly
def proxy_property_directly(bind_attr, attrs): """ 为类添加代理属性 """ def cls_builder(cls): """ 添加到类 """ for attr_name in attrs: setattr(cls, attr_name, _make_proxy_property(bind_attr, attr_name)) return cls return cls_builder
python
def proxy_property_directly(bind_attr, attrs): """ 为类添加代理属性 """ def cls_builder(cls): """ 添加到类 """ for attr_name in attrs: setattr(cls, attr_name, _make_proxy_property(bind_attr, attr_name)) return cls return cls_builder
[ "def", "proxy_property_directly", "(", "bind_attr", ",", "attrs", ")", ":", "def", "cls_builder", "(", "cls", ")", ":", "\"\"\"\n 添加到类\n \"\"\"", "for", "attr_name", "in", "attrs", ":", "setattr", "(", "cls", ",", "attr_name", ",", "_make_proxy_property", "(", "bind_attr", ",", "attr_name", ")", ")", "return", "cls", "return", "cls_builder" ]
为类添加代理属性
[ "为类添加代理属性" ]
1a74a062507e2df8f833a70885e69dca0ab3e7e7
https://github.com/zeromake/aiosqlite3/blob/1a74a062507e2df8f833a70885e69dca0ab3e7e7/aiosqlite3/utils.py#L265-L276
train
opennode/waldur-core
waldur_core/structure/models.py
StructureLoggableMixin.get_permitted_objects_uuids
def get_permitted_objects_uuids(cls, user): """ Return query dictionary to search objects available to user. """ uuids = filter_queryset_for_user(cls.objects.all(), user).values_list('uuid', flat=True) key = core_utils.camel_case_to_underscore(cls.__name__) + '_uuid' return {key: uuids}
python
def get_permitted_objects_uuids(cls, user): """ Return query dictionary to search objects available to user. """ uuids = filter_queryset_for_user(cls.objects.all(), user).values_list('uuid', flat=True) key = core_utils.camel_case_to_underscore(cls.__name__) + '_uuid' return {key: uuids}
[ "def", "get_permitted_objects_uuids", "(", "cls", ",", "user", ")", ":", "uuids", "=", "filter_queryset_for_user", "(", "cls", ".", "objects", ".", "all", "(", ")", ",", "user", ")", ".", "values_list", "(", "'uuid'", ",", "flat", "=", "True", ")", "key", "=", "core_utils", ".", "camel_case_to_underscore", "(", "cls", ".", "__name__", ")", "+", "'_uuid'", "return", "{", "key", ":", "uuids", "}" ]
Return query dictionary to search objects available to user.
[ "Return", "query", "dictionary", "to", "search", "objects", "available", "to", "user", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/models.py#L79-L85
train
opennode/waldur-core
waldur_core/structure/models.py
PermissionMixin.has_user
def has_user(self, user, role=None, timestamp=False): """ Checks whether user has role in entity. `timestamp` can have following values: - False - check whether user has role in entity at the moment. - None - check whether user has permanent role in entity. - Datetime object - check whether user will have role in entity at specific timestamp. """ permissions = self.permissions.filter(user=user, is_active=True) if role is not None: permissions = permissions.filter(role=role) if timestamp is None: permissions = permissions.filter(expiration_time=None) elif timestamp: permissions = permissions.filter(Q(expiration_time=None) | Q(expiration_time__gte=timestamp)) return permissions.exists()
python
def has_user(self, user, role=None, timestamp=False): """ Checks whether user has role in entity. `timestamp` can have following values: - False - check whether user has role in entity at the moment. - None - check whether user has permanent role in entity. - Datetime object - check whether user will have role in entity at specific timestamp. """ permissions = self.permissions.filter(user=user, is_active=True) if role is not None: permissions = permissions.filter(role=role) if timestamp is None: permissions = permissions.filter(expiration_time=None) elif timestamp: permissions = permissions.filter(Q(expiration_time=None) | Q(expiration_time__gte=timestamp)) return permissions.exists()
[ "def", "has_user", "(", "self", ",", "user", ",", "role", "=", "None", ",", "timestamp", "=", "False", ")", ":", "permissions", "=", "self", ".", "permissions", ".", "filter", "(", "user", "=", "user", ",", "is_active", "=", "True", ")", "if", "role", "is", "not", "None", ":", "permissions", "=", "permissions", ".", "filter", "(", "role", "=", "role", ")", "if", "timestamp", "is", "None", ":", "permissions", "=", "permissions", ".", "filter", "(", "expiration_time", "=", "None", ")", "elif", "timestamp", ":", "permissions", "=", "permissions", ".", "filter", "(", "Q", "(", "expiration_time", "=", "None", ")", "|", "Q", "(", "expiration_time__gte", "=", "timestamp", ")", ")", "return", "permissions", ".", "exists", "(", ")" ]
Checks whether user has role in entity. `timestamp` can have following values: - False - check whether user has role in entity at the moment. - None - check whether user has permanent role in entity. - Datetime object - check whether user will have role in entity at specific timestamp.
[ "Checks", "whether", "user", "has", "role", "in", "entity", ".", "timestamp", "can", "have", "following", "values", ":", "-", "False", "-", "check", "whether", "user", "has", "role", "in", "entity", "at", "the", "moment", ".", "-", "None", "-", "check", "whether", "user", "has", "permanent", "role", "in", "entity", ".", "-", "Datetime", "object", "-", "check", "whether", "user", "will", "have", "role", "in", "entity", "at", "specific", "timestamp", "." ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/structure/models.py#L191-L209
train
opennode/waldur-core
waldur_core/core/magic.py
Magic.from_buffer
def from_buffer(self, buf): """ Identify the contents of `buf` """ with self.lock: try: # if we're on python3, convert buf to bytes # otherwise this string is passed as wchar* # which is not what libmagic expects if isinstance(buf, str) and str != bytes: buf = buf.encode('utf-8', errors='replace') return maybe_decode(magic_buffer(self.cookie, buf)) except MagicException as e: return self._handle509Bug(e)
python
def from_buffer(self, buf): """ Identify the contents of `buf` """ with self.lock: try: # if we're on python3, convert buf to bytes # otherwise this string is passed as wchar* # which is not what libmagic expects if isinstance(buf, str) and str != bytes: buf = buf.encode('utf-8', errors='replace') return maybe_decode(magic_buffer(self.cookie, buf)) except MagicException as e: return self._handle509Bug(e)
[ "def", "from_buffer", "(", "self", ",", "buf", ")", ":", "with", "self", ".", "lock", ":", "try", ":", "# if we're on python3, convert buf to bytes", "# otherwise this string is passed as wchar*", "# which is not what libmagic expects", "if", "isinstance", "(", "buf", ",", "str", ")", "and", "str", "!=", "bytes", ":", "buf", "=", "buf", ".", "encode", "(", "'utf-8'", ",", "errors", "=", "'replace'", ")", "return", "maybe_decode", "(", "magic_buffer", "(", "self", ".", "cookie", ",", "buf", ")", ")", "except", "MagicException", "as", "e", ":", "return", "self", ".", "_handle509Bug", "(", "e", ")" ]
Identify the contents of `buf`
[ "Identify", "the", "contents", "of", "buf" ]
d6c17a9592bb6c49c33567542eef8d099605a46a
https://github.com/opennode/waldur-core/blob/d6c17a9592bb6c49c33567542eef8d099605a46a/waldur_core/core/magic.py#L71-L84
train
nuagenetworks/bambou
bambou/nurest_push_center.py
NURESTPushCenter.start
def start(self, timeout=None, root_object=None): """ Starts listening to events. Args: timeout (int): number of seconds before timeout. Used for testing purpose only. root_object (bambou.NURESTRootObject): NURESTRootObject object that is listening. Used for testing purpose only. """ if self._is_running: return if timeout: self._timeout = timeout self._start_time = int(time()) pushcenter_logger.debug("[NURESTPushCenter] Starting push center on url %s ..." % self.url) self._is_running = True self.__root_object = root_object from .nurest_session import NURESTSession current_session = NURESTSession.get_current_session() args_session = {'session': current_session} self._thread = StoppableThread(target=self._listen, name='push-center', kwargs=args_session) self._thread.daemon = True self._thread.start()
python
def start(self, timeout=None, root_object=None): """ Starts listening to events. Args: timeout (int): number of seconds before timeout. Used for testing purpose only. root_object (bambou.NURESTRootObject): NURESTRootObject object that is listening. Used for testing purpose only. """ if self._is_running: return if timeout: self._timeout = timeout self._start_time = int(time()) pushcenter_logger.debug("[NURESTPushCenter] Starting push center on url %s ..." % self.url) self._is_running = True self.__root_object = root_object from .nurest_session import NURESTSession current_session = NURESTSession.get_current_session() args_session = {'session': current_session} self._thread = StoppableThread(target=self._listen, name='push-center', kwargs=args_session) self._thread.daemon = True self._thread.start()
[ "def", "start", "(", "self", ",", "timeout", "=", "None", ",", "root_object", "=", "None", ")", ":", "if", "self", ".", "_is_running", ":", "return", "if", "timeout", ":", "self", ".", "_timeout", "=", "timeout", "self", ".", "_start_time", "=", "int", "(", "time", "(", ")", ")", "pushcenter_logger", ".", "debug", "(", "\"[NURESTPushCenter] Starting push center on url %s ...\"", "%", "self", ".", "url", ")", "self", ".", "_is_running", "=", "True", "self", ".", "__root_object", "=", "root_object", "from", ".", "nurest_session", "import", "NURESTSession", "current_session", "=", "NURESTSession", ".", "get_current_session", "(", ")", "args_session", "=", "{", "'session'", ":", "current_session", "}", "self", ".", "_thread", "=", "StoppableThread", "(", "target", "=", "self", ".", "_listen", ",", "name", "=", "'push-center'", ",", "kwargs", "=", "args_session", ")", "self", ".", "_thread", ".", "daemon", "=", "True", "self", ".", "_thread", ".", "start", "(", ")" ]
Starts listening to events. Args: timeout (int): number of seconds before timeout. Used for testing purpose only. root_object (bambou.NURESTRootObject): NURESTRootObject object that is listening. Used for testing purpose only.
[ "Starts", "listening", "to", "events", "." ]
d334fea23e384d3df8e552fe1849ad707941c666
https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_push_center.py#L103-L128
train
nuagenetworks/bambou
bambou/nurest_push_center.py
NURESTPushCenter.stop
def stop(self): """ Stops listening for events. """ if not self._is_running: return pushcenter_logger.debug("[NURESTPushCenter] Stopping...") self._thread.stop() self._thread.join() self._is_running = False self._current_connection = None self._start_time = None self._timeout = None
python
def stop(self): """ Stops listening for events. """ if not self._is_running: return pushcenter_logger.debug("[NURESTPushCenter] Stopping...") self._thread.stop() self._thread.join() self._is_running = False self._current_connection = None self._start_time = None self._timeout = None
[ "def", "stop", "(", "self", ")", ":", "if", "not", "self", ".", "_is_running", ":", "return", "pushcenter_logger", ".", "debug", "(", "\"[NURESTPushCenter] Stopping...\"", ")", "self", ".", "_thread", ".", "stop", "(", ")", "self", ".", "_thread", ".", "join", "(", ")", "self", ".", "_is_running", "=", "False", "self", ".", "_current_connection", "=", "None", "self", ".", "_start_time", "=", "None", "self", ".", "_timeout", "=", "None" ]
Stops listening for events.
[ "Stops", "listening", "for", "events", "." ]
d334fea23e384d3df8e552fe1849ad707941c666
https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_push_center.py#L130-L144
train
nuagenetworks/bambou
bambou/nurest_push_center.py
NURESTPushCenter.wait_until_exit
def wait_until_exit(self): """ Wait until thread exit Used for testing purpose only """ if self._timeout is None: raise Exception("Thread will never exit. Use stop or specify timeout when starting it!") self._thread.join() self.stop()
python
def wait_until_exit(self): """ Wait until thread exit Used for testing purpose only """ if self._timeout is None: raise Exception("Thread will never exit. Use stop or specify timeout when starting it!") self._thread.join() self.stop()
[ "def", "wait_until_exit", "(", "self", ")", ":", "if", "self", ".", "_timeout", "is", "None", ":", "raise", "Exception", "(", "\"Thread will never exit. Use stop or specify timeout when starting it!\"", ")", "self", ".", "_thread", ".", "join", "(", ")", "self", ".", "stop", "(", ")" ]
Wait until thread exit Used for testing purpose only
[ "Wait", "until", "thread", "exit" ]
d334fea23e384d3df8e552fe1849ad707941c666
https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_push_center.py#L146-L156
train
nuagenetworks/bambou
bambou/nurest_push_center.py
NURESTPushCenter._did_receive_event
def _did_receive_event(self, connection): """ Receive an event from connection """ if not self._is_running: return if connection.has_timeouted: return response = connection.response data = None if response.status_code != 200: pushcenter_logger.error("[NURESTPushCenter]: Connection failure [%s] %s" % (response.status_code, response.errors)) else: data = response.data if len(self._delegate_methods) > 0: for m in self._delegate_methods: try: m(data) except Exception as exc: pushcenter_logger.error("[NURESTPushCenter] Delegate method %s failed:\n%s" % (m, exc)) elif data: events = data['events'] self.nb_events_received += len(events) self.nb_push_received += 1 pushcenter_logger.info("[NURESTPushCenter] Received Push #%s (total=%s, latest=%s)\n%s" % (self.nb_push_received, self.nb_events_received, len(events), json.dumps(events, indent=4))) self._last_events.extend(events) if self._is_running: uuid = None if data and 'uuid' in data: uuid = data['uuid'] self._listen(uuid)
python
def _did_receive_event(self, connection): """ Receive an event from connection """ if not self._is_running: return if connection.has_timeouted: return response = connection.response data = None if response.status_code != 200: pushcenter_logger.error("[NURESTPushCenter]: Connection failure [%s] %s" % (response.status_code, response.errors)) else: data = response.data if len(self._delegate_methods) > 0: for m in self._delegate_methods: try: m(data) except Exception as exc: pushcenter_logger.error("[NURESTPushCenter] Delegate method %s failed:\n%s" % (m, exc)) elif data: events = data['events'] self.nb_events_received += len(events) self.nb_push_received += 1 pushcenter_logger.info("[NURESTPushCenter] Received Push #%s (total=%s, latest=%s)\n%s" % (self.nb_push_received, self.nb_events_received, len(events), json.dumps(events, indent=4))) self._last_events.extend(events) if self._is_running: uuid = None if data and 'uuid' in data: uuid = data['uuid'] self._listen(uuid)
[ "def", "_did_receive_event", "(", "self", ",", "connection", ")", ":", "if", "not", "self", ".", "_is_running", ":", "return", "if", "connection", ".", "has_timeouted", ":", "return", "response", "=", "connection", ".", "response", "data", "=", "None", "if", "response", ".", "status_code", "!=", "200", ":", "pushcenter_logger", ".", "error", "(", "\"[NURESTPushCenter]: Connection failure [%s] %s\"", "%", "(", "response", ".", "status_code", ",", "response", ".", "errors", ")", ")", "else", ":", "data", "=", "response", ".", "data", "if", "len", "(", "self", ".", "_delegate_methods", ")", ">", "0", ":", "for", "m", "in", "self", ".", "_delegate_methods", ":", "try", ":", "m", "(", "data", ")", "except", "Exception", "as", "exc", ":", "pushcenter_logger", ".", "error", "(", "\"[NURESTPushCenter] Delegate method %s failed:\\n%s\"", "%", "(", "m", ",", "exc", ")", ")", "elif", "data", ":", "events", "=", "data", "[", "'events'", "]", "self", ".", "nb_events_received", "+=", "len", "(", "events", ")", "self", ".", "nb_push_received", "+=", "1", "pushcenter_logger", ".", "info", "(", "\"[NURESTPushCenter] Received Push #%s (total=%s, latest=%s)\\n%s\"", "%", "(", "self", ".", "nb_push_received", ",", "self", ".", "nb_events_received", ",", "len", "(", "events", ")", ",", "json", ".", "dumps", "(", "events", ",", "indent", "=", "4", ")", ")", ")", "self", ".", "_last_events", ".", "extend", "(", "events", ")", "if", "self", ".", "_is_running", ":", "uuid", "=", "None", "if", "data", "and", "'uuid'", "in", "data", ":", "uuid", "=", "data", "[", "'uuid'", "]", "self", ".", "_listen", "(", "uuid", ")" ]
Receive an event from connection
[ "Receive", "an", "event", "from", "connection" ]
d334fea23e384d3df8e552fe1849ad707941c666
https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_push_center.py#L173-L210
train
nuagenetworks/bambou
bambou/nurest_push_center.py
NURESTPushCenter._listen
def _listen(self, uuid=None, session=None): """ Listen a connection uuid """ if self.url is None: raise Exception("NURESTPushCenter needs to have a valid URL. please use setURL: before starting it.") events_url = "%s/events" % self.url if uuid: events_url = "%s?uuid=%s" % (events_url, uuid) request = NURESTRequest(method='GET', url=events_url) # Force async to False so the push center will have only 1 thread running connection = NURESTConnection(request=request, async=True, callback=self._did_receive_event, root_object=self._root_object) if self._timeout: if int(time()) - self._start_time >= self._timeout: pushcenter_logger.debug("[NURESTPushCenter] Timeout (timeout=%ss)." % self._timeout) return else: connection.timeout = self._timeout pushcenter_logger.info('Bambou Sending >>>>>>\n%s %s' % (request.method, request.url)) # connection.ignore_request_idle = True connection.start()
python
def _listen(self, uuid=None, session=None): """ Listen a connection uuid """ if self.url is None: raise Exception("NURESTPushCenter needs to have a valid URL. please use setURL: before starting it.") events_url = "%s/events" % self.url if uuid: events_url = "%s?uuid=%s" % (events_url, uuid) request = NURESTRequest(method='GET', url=events_url) # Force async to False so the push center will have only 1 thread running connection = NURESTConnection(request=request, async=True, callback=self._did_receive_event, root_object=self._root_object) if self._timeout: if int(time()) - self._start_time >= self._timeout: pushcenter_logger.debug("[NURESTPushCenter] Timeout (timeout=%ss)." % self._timeout) return else: connection.timeout = self._timeout pushcenter_logger.info('Bambou Sending >>>>>>\n%s %s' % (request.method, request.url)) # connection.ignore_request_idle = True connection.start()
[ "def", "_listen", "(", "self", ",", "uuid", "=", "None", ",", "session", "=", "None", ")", ":", "if", "self", ".", "url", "is", "None", ":", "raise", "Exception", "(", "\"NURESTPushCenter needs to have a valid URL. please use setURL: before starting it.\"", ")", "events_url", "=", "\"%s/events\"", "%", "self", ".", "url", "if", "uuid", ":", "events_url", "=", "\"%s?uuid=%s\"", "%", "(", "events_url", ",", "uuid", ")", "request", "=", "NURESTRequest", "(", "method", "=", "'GET'", ",", "url", "=", "events_url", ")", "# Force async to False so the push center will have only 1 thread running", "connection", "=", "NURESTConnection", "(", "request", "=", "request", ",", "async", "=", "True", ",", "callback", "=", "self", ".", "_did_receive_event", ",", "root_object", "=", "self", ".", "_root_object", ")", "if", "self", ".", "_timeout", ":", "if", "int", "(", "time", "(", ")", ")", "-", "self", ".", "_start_time", ">=", "self", ".", "_timeout", ":", "pushcenter_logger", ".", "debug", "(", "\"[NURESTPushCenter] Timeout (timeout=%ss).\"", "%", "self", ".", "_timeout", ")", "return", "else", ":", "connection", ".", "timeout", "=", "self", ".", "_timeout", "pushcenter_logger", ".", "info", "(", "'Bambou Sending >>>>>>\\n%s %s'", "%", "(", "request", ".", "method", ",", "request", ".", "url", ")", ")", "# connection.ignore_request_idle = True", "connection", ".", "start", "(", ")" ]
Listen a connection uuid
[ "Listen", "a", "connection", "uuid" ]
d334fea23e384d3df8e552fe1849ad707941c666
https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_push_center.py#L212-L238
train
nuagenetworks/bambou
bambou/nurest_push_center.py
NURESTPushCenter.add_delegate
def add_delegate(self, callback): """ Registers a new delegate callback The prototype should be function(data), where data will be the decoded json push Args: callback (function): method to trigger when push center receives events """ if callback in self._delegate_methods: return self._delegate_methods.append(callback)
python
def add_delegate(self, callback): """ Registers a new delegate callback The prototype should be function(data), where data will be the decoded json push Args: callback (function): method to trigger when push center receives events """ if callback in self._delegate_methods: return self._delegate_methods.append(callback)
[ "def", "add_delegate", "(", "self", ",", "callback", ")", ":", "if", "callback", "in", "self", ".", "_delegate_methods", ":", "return", "self", ".", "_delegate_methods", ".", "append", "(", "callback", ")" ]
Registers a new delegate callback The prototype should be function(data), where data will be the decoded json push Args: callback (function): method to trigger when push center receives events
[ "Registers", "a", "new", "delegate", "callback" ]
d334fea23e384d3df8e552fe1849ad707941c666
https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_push_center.py#L240-L252
train