hexsha
stringlengths
40
40
size
int64
2
1.02M
ext
stringclasses
10 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
4
245
max_stars_repo_name
stringlengths
6
130
max_stars_repo_head_hexsha
stringlengths
40
40
max_stars_repo_licenses
listlengths
1
10
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
4
245
max_issues_repo_name
stringlengths
6
130
max_issues_repo_head_hexsha
stringlengths
40
40
max_issues_repo_licenses
listlengths
1
10
max_issues_count
int64
1
67k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
4
245
max_forks_repo_name
stringlengths
6
130
max_forks_repo_head_hexsha
stringlengths
40
40
max_forks_repo_licenses
listlengths
1
10
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
2
1.02M
avg_line_length
float64
1
958k
max_line_length
int64
1
987k
alphanum_fraction
float64
0
1
content_no_comment
stringlengths
0
1.01M
is_comment_constant_removed
bool
2 classes
is_sharp_comment_removed
bool
1 class
1c48308b9835d4ad17cc2c255db05b765a7dd3a3
2,479
py
Python
conpaas-services/contrib/libcloud/common/hostvirtual.py
bopopescu/conpaas-1
cea3c02f499a729464697de7cf98c2041febc0ab
[ "BSD-3-Clause" ]
5
2016-02-24T14:44:03.000Z
2020-11-29T19:18:40.000Z
conpaas-services/contrib/libcloud/common/hostvirtual.py
bopopescu/conpaas-1
cea3c02f499a729464697de7cf98c2041febc0ab
[ "BSD-3-Clause" ]
25
2015-03-23T16:05:19.000Z
2018-02-13T17:22:22.000Z
conpaas-services/contrib/libcloud/common/hostvirtual.py
bopopescu/conpaas-1
cea3c02f499a729464697de7cf98c2041febc0ab
[ "BSD-3-Clause" ]
3
2018-09-14T16:54:14.000Z
2020-07-26T03:14:56.000Z
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License.You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. try: import simplejson as json except ImportError: import json from libcloud.utils.py3 import httplib from libcloud.common.base import ConnectionKey, JsonResponse from libcloud.compute.types import InvalidCredsError from libcloud.common.types import LibcloudError API_HOST = 'www.vr.org' class HostVirtualException(LibcloudError): def __init__(self, code, message): self.code = code self.message = message self.args = (code, message) def __str__(self): return self.__repr__() def __repr__(self): return '<HostVirtualException in %d: %s>' % (self.code, self.message) class HostVirtualConnection(ConnectionKey): host = API_HOST def add_default_params(self, params): params['key'] = self.key return params class HostVirtualResponse(JsonResponse): valid_response_codes = [httplib.OK, httplib.ACCEPTED, httplib.CREATED, httplib.NO_CONTENT] def parse_body(self): if not self.body: return None data = json.loads(self.body) return data def parse_error(self): data = self.parse_body() if self.status == httplib.UNAUTHORIZED: raise InvalidCredsError('%(code)s:%(message)s' % (data['error'])) elif self.status == httplib.PRECONDITION_FAILED: raise HostVirtualException( data['error']['code'], data['error']['message']) elif self.status == httplib.NOT_FOUND: raise HostVirtualException( data['error']['code'], data['error']['message']) return self.body def success(self): return self.status in self.valid_response_codes
32.618421
77
0.689794
try: import simplejson as json except ImportError: import json from libcloud.utils.py3 import httplib from libcloud.common.base import ConnectionKey, JsonResponse from libcloud.compute.types import InvalidCredsError from libcloud.common.types import LibcloudError API_HOST = 'www.vr.org' class HostVirtualException(LibcloudError): def __init__(self, code, message): self.code = code self.message = message self.args = (code, message) def __str__(self): return self.__repr__() def __repr__(self): return '<HostVirtualException in %d: %s>' % (self.code, self.message) class HostVirtualConnection(ConnectionKey): host = API_HOST def add_default_params(self, params): params['key'] = self.key return params class HostVirtualResponse(JsonResponse): valid_response_codes = [httplib.OK, httplib.ACCEPTED, httplib.CREATED, httplib.NO_CONTENT] def parse_body(self): if not self.body: return None data = json.loads(self.body) return data def parse_error(self): data = self.parse_body() if self.status == httplib.UNAUTHORIZED: raise InvalidCredsError('%(code)s:%(message)s' % (data['error'])) elif self.status == httplib.PRECONDITION_FAILED: raise HostVirtualException( data['error']['code'], data['error']['message']) elif self.status == httplib.NOT_FOUND: raise HostVirtualException( data['error']['code'], data['error']['message']) return self.body def success(self): return self.status in self.valid_response_codes
true
true
1c483094595d09d08bba4956265bb0fbca8a59fc
538
py
Python
src/myres/migrations/0003_auto_20170404_0924.py
tsotetsi/myres-api
9ca8f6762168d07a767c30a490520dfad54079d9
[ "MIT" ]
null
null
null
src/myres/migrations/0003_auto_20170404_0924.py
tsotetsi/myres-api
9ca8f6762168d07a767c30a490520dfad54079d9
[ "MIT" ]
null
null
null
src/myres/migrations/0003_auto_20170404_0924.py
tsotetsi/myres-api
9ca8f6762168d07a767c30a490520dfad54079d9
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # Generated by Django 1.10.5 on 2017-04-04 07:24 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('myres', '0002_student_residence'), ] operations = [ migrations.AlterField( model_name='student', name='residence', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='myres.Residence'), ), ]
24.454545
103
0.650558
from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('myres', '0002_student_residence'), ] operations = [ migrations.AlterField( model_name='student', name='residence', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='myres.Residence'), ), ]
true
true
1c48317e7445a689b165340a69a42ccf53b75ba5
15,085
py
Python
src/models/data_loader.py
tikhonovpavel/LdaSummarization
fbfb229e83548d9dd8f921626fd3fbf423b0305a
[ "MIT" ]
null
null
null
src/models/data_loader.py
tikhonovpavel/LdaSummarization
fbfb229e83548d9dd8f921626fd3fbf423b0305a
[ "MIT" ]
null
null
null
src/models/data_loader.py
tikhonovpavel/LdaSummarization
fbfb229e83548d9dd8f921626fd3fbf423b0305a
[ "MIT" ]
null
null
null
import bisect import gc import glob import pickle import random import torch from others.logging import logger import gensim from gensim.utils import simple_preprocess from gensim.parsing.preprocessing import STOPWORDS from nltk.stem import WordNetLemmatizer, SnowballStemmer from nltk.stem.porter import * import numpy as np np.random.seed(2018) # import nltk # nltk.download('wordnet') # # with open('../topic_modelling_data/dictionary.pkl', 'rb') as f: # tm_dictionary = pickle.load(f) # # with open('../topic_modelling_data/lda_model.pkl', 'rb') as f: # lda_model = pickle.load(f) # # stemmer = SnowballStemmer('english') # # def lemmatize_stemming(text): # return stemmer.stem(WordNetLemmatizer().lemmatize(text, pos='v')) # # def preprocess(text): # result = [] # for token in gensim.utils.simple_preprocess(text): # if token not in gensim.parsing.preprocessing.STOPWORDS and len(token) > 3: # result.append(lemmatize_stemming(token)) # return result class Batch(object): def _pad(self, data, pad_id, width=-1): if (width == -1): width = max(len(d) for d in data) rtn_data = [d + [pad_id] * (width - len(d)) for d in data] return rtn_data def __init__(self, data=None, device=None, is_test=False): """Create a Batch from a list of examples.""" if data is not None: self.batch_size = len(data) pre_src = [x[0] for x in data] pre_tgt = [x[1] for x in data] pre_segs = [x[2] for x in data] pre_clss = [x[3] for x in data] pre_src_sent_labels = [x[4] for x in data] src = torch.tensor(self._pad(pre_src, 0)) tgt = torch.tensor(self._pad(pre_tgt, 0)) segs = torch.tensor(self._pad(pre_segs, 0)) try: mask_src = 1 - (src == 0) mask_tgt = 1 - (tgt == 0) except RuntimeError as err: if 'Subtraction, the `-` operator, with a bool tensor is not supported' not in str(err): raise err mask_src = ~(src == 0) mask_tgt = ~(tgt == 0) clss = torch.tensor(self._pad(pre_clss, -1)) src_sent_labels = torch.tensor(self._pad(pre_src_sent_labels, 0)) try: mask_cls = 1 - (clss == -1) except RuntimeError as err: if 'Subtraction, the `-` operator, with a bool tensor is not supported' not in str(err): raise err mask_cls = ~(clss == -1) clss[clss == -1] = 0 setattr(self, 'clss', clss.to(device)) setattr(self, 'mask_cls', mask_cls.to(device)) setattr(self, 'src_sent_labels', src_sent_labels.to(device)) setattr(self, 'src', src.to(device)) setattr(self, 'tgt', tgt.to(device)) setattr(self, 'segs', segs.to(device)) setattr(self, 'mask_src', mask_src.to(device)) setattr(self, 'mask_tgt', mask_tgt.to(device)) # setattr(self, 'topics', topics.to(device)) if (is_test) or True: src_str = [x[-3] for x in data] setattr(self, 'src_str', src_str) tgt_str = [x[-2] for x in data] setattr(self, 'tgt_str', tgt_str) topics = [x[-1] for x in data] setattr(self, 'topics', topics) def __len__(self): return self.batch_size def load_dataset(args, corpus_type, shuffle): """ Dataset generator. Don't do extra stuff here, like printing, because they will be postponed to the first loading time. Args: corpus_type: 'train' or 'valid' Returns: A list of dataset, the dataset(s) are lazily loaded. """ assert corpus_type in ["train", "valid", "test"] def _lazy_dataset_loader(pt_file, corpus_type, use_topic_modelling): dataset = torch.load(pt_file) # if use_topic_modelling: # for article in dataset: # # unseen_document = 'How a Pentagon deal became an identity crisis for Google' # bow_vector = tm_dictionary.doc2bow(preprocess(' '.join(article['src_txt']))) # # article_topic = sorted(lda_model[bow_vector], key=lambda tup: -1 * tup[1])[0] # article_topic = article_topic[0] # DICTIONARY_SIZE = 30_000 # article_topic = DICTIONARY_SIZE + article_topic # # article['src'] = [article_topic] + article['src'] # # # for index, score in sorted(lda_model[bow_vector], key=lambda tup: -1 * tup[1]): # # print("Score: {}\t Topic: {}".format(score, lda_model.print_topic(index, 5))) logger.info('Loading %s dataset from %s, number of examples: %d' % (corpus_type, pt_file, len(dataset))) return dataset # Sort the glob output by file name (by increasing indexes). pts = sorted(glob.glob(args.bert_data_path + '.' + corpus_type + '.[0-9]*.pt')) if pts: if (shuffle): random.shuffle(pts) for pt in pts: yield _lazy_dataset_loader(pt, corpus_type, args.use_topic_modelling) else: # Only one inputters.*Dataset, simple! pt = args.bert_data_path + '.' + corpus_type + '.pt' yield _lazy_dataset_loader(pt, corpus_type, args.use_topic_modelling) def abs_batch_size_fn(new, count): src, tgt = new[0], new[1] global max_n_sents, max_n_tokens, max_size if count == 1: max_size = 0 max_n_sents=0 max_n_tokens=0 max_n_sents = max(max_n_sents, len(tgt)) max_size = max(max_size, max_n_sents) src_elements = count * max_size if (count > 6): return src_elements + 1e3 return src_elements def ext_batch_size_fn(new, count): if (len(new) == 4): pass src, labels = new[0], new[4] global max_n_sents, max_n_tokens, max_size if count == 1: max_size = 0 max_n_sents = 0 max_n_tokens = 0 max_n_sents = max(max_n_sents, len(src)) max_size = max(max_size, max_n_sents) src_elements = count * max_size return src_elements class Dataloader(object): def __init__(self, args, datasets, batch_size, device, shuffle, is_test): self.args = args self.datasets = datasets self.batch_size = batch_size self.device = device self.shuffle = shuffle self.is_test = is_test self.use_topic_modelling = args.use_topic_modelling self.cur_iter = self._next_dataset_iterator(datasets) assert self.cur_iter is not None def __iter__(self): dataset_iter = (d for d in self.datasets) while self.cur_iter is not None: for batch in self.cur_iter: yield batch self.cur_iter = self._next_dataset_iterator(dataset_iter) def _next_dataset_iterator(self, dataset_iter): try: # Drop the current dataset for decreasing memory if hasattr(self, "cur_dataset"): self.cur_dataset = None gc.collect() del self.cur_dataset gc.collect() self.cur_dataset = next(dataset_iter) except StopIteration: return None return DataIterator(args = self.args, dataset=self.cur_dataset, batch_size=self.batch_size, device=self.device, shuffle=self.shuffle, is_test=self.is_test) class DataIterator(object): def __init__(self, args, dataset, batch_size, device=None, is_test=False, shuffle=True): self.args = args self.batch_size, self.is_test, self.dataset = batch_size, is_test, dataset self.iterations = 0 self.device = device self.shuffle = shuffle self.sort_key = lambda x: len(x[1]) self._iterations_this_epoch = 0 if (self.args.task == 'abs'): self.batch_size_fn = abs_batch_size_fn else: self.batch_size_fn = ext_batch_size_fn def data(self): if self.shuffle: random.shuffle(self.dataset) xs = self.dataset return xs def preprocess(self, ex, is_test): src = ex['src'] tgt = ex['tgt'][:self.args.max_tgt_len][:-1]+[2] src_sent_labels = ex['src_sent_labels'] segs = ex['segs'] if(not self.args.use_interval): segs=[0]*len(segs) clss = ex['clss'] src_txt = ex['src_txt'] tgt_txt = ex['tgt_txt'] try: topics = ex['topics'] except KeyError: print('Warning: topics are not presented!') topics = None end_id = [src[-1]] src = src[:-1][:self.args.max_pos - 1] + end_id segs = segs[:self.args.max_pos] max_sent_id = bisect.bisect_left(clss, self.args.max_pos) src_sent_labels = src_sent_labels[:max_sent_id] clss = clss[:max_sent_id] # src_txt = src_txt[:max_sent_id] if(is_test): return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics else: return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics def batch_buffer(self, data, batch_size): minibatch, size_so_far = [], 0 for ex in data: if(len(ex['src'])==0): continue ex = self.preprocess(ex, self.is_test) if(ex is None): continue minibatch.append(ex) size_so_far = self.batch_size_fn(ex, len(minibatch)) if size_so_far == batch_size: yield minibatch minibatch, size_so_far = [], 0 elif size_so_far > batch_size: yield minibatch[:-1] minibatch, size_so_far = minibatch[-1:], self.batch_size_fn(ex, 1) if minibatch: yield minibatch def batch(self, data, batch_size): """Yield elements from data in chunks of batch_size.""" minibatch, size_so_far = [], 0 for ex in data: minibatch.append(ex) size_so_far = self.batch_size_fn(ex, len(minibatch)) if size_so_far == batch_size: yield minibatch minibatch, size_so_far = [], 0 elif size_so_far > batch_size: yield minibatch[:-1] minibatch, size_so_far = minibatch[-1:], self.batch_size_fn(ex, 1) if minibatch: yield minibatch def create_batches(self): """ Create batches """ data = self.data() for buffer in self.batch_buffer(data, self.batch_size * 300): if (self.args.task == 'abs'): p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = sorted(p_batch, key=lambda x: len(x[1])) else: p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = self.batch(p_batch, self.batch_size) p_batch = list(p_batch) if (self.shuffle): random.shuffle(p_batch) for b in p_batch: if(len(b)==0): continue yield b def __iter__(self): while True: self.batches = self.create_batches() for idx, minibatch in enumerate(self.batches): # fast-forward if loaded from state if self._iterations_this_epoch > idx: continue self.iterations += 1 self._iterations_this_epoch += 1 batch = Batch(minibatch, self.device, self.is_test) yield batch return class TextDataloader(object): def __init__(self, args, datasets, batch_size, device, shuffle, is_test): self.args = args self.batch_size = batch_size self.device = device def data(self): if self.shuffle: random.shuffle(self.dataset) xs = self.dataset return xs def preprocess(self, ex, is_test): src = ex['src'] tgt = ex['tgt'][:self.args.max_tgt_len][:-1] + [2] src_sent_labels = ex['src_sent_labels'] segs = ex['segs'] if (not self.args.use_interval): segs = [0] * len(segs) clss = ex['clss'] src_txt = ex['src_txt'] tgt_txt = ex['tgt_txt'] topics = ex['topics'] end_id = [src[-1]] src = src[:-1][:self.args.max_pos - 1] + end_id segs = segs[:self.args.max_pos] max_sent_id = bisect.bisect_left(clss, self.args.max_pos) src_sent_labels = src_sent_labels[:max_sent_id] clss = clss[:max_sent_id] # src_txt = src_txt[:max_sent_id] if (is_test): return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics else: return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics def batch_buffer(self, data, batch_size): minibatch, size_so_far = [], 0 for ex in data: if (len(ex['src']) == 0): continue ex = self.preprocess(ex, self.is_test) if (ex is None): continue minibatch.append(ex) size_so_far = simple_batch_size_fn(ex, len(minibatch)) if size_so_far == batch_size: yield minibatch minibatch, size_so_far = [], 0 elif size_so_far > batch_size: yield minibatch[:-1] minibatch, size_so_far = minibatch[-1:], simple_batch_size_fn(ex, 1) if minibatch: yield minibatch def create_batches(self): """ Create batches """ data = self.data() for buffer in self.batch_buffer(data, self.batch_size * 300): if (self.args.task == 'abs'): p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = sorted(p_batch, key=lambda x: len(x[1])) else: p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = batch(p_batch, self.batch_size) p_batch = batch(p_batch, self.batch_size) p_batch = list(p_batch) if (self.shuffle): random.shuffle(p_batch) for b in p_batch: if (len(b) == 0): continue yield b def __iter__(self): while True: self.batches = self.create_batches() for idx, minibatch in enumerate(self.batches): # fast-forward if loaded from state if self._iterations_this_epoch > idx: continue self.iterations += 1 self._iterations_this_epoch += 1 batch = Batch(minibatch, self.device, self.is_test) yield batch return
33.522222
104
0.562678
import bisect import gc import glob import pickle import random import torch from others.logging import logger import gensim from gensim.utils import simple_preprocess from gensim.parsing.preprocessing import STOPWORDS from nltk.stem import WordNetLemmatizer, SnowballStemmer from nltk.stem.porter import * import numpy as np np.random.seed(2018) class Batch(object): def _pad(self, data, pad_id, width=-1): if (width == -1): width = max(len(d) for d in data) rtn_data = [d + [pad_id] * (width - len(d)) for d in data] return rtn_data def __init__(self, data=None, device=None, is_test=False): if data is not None: self.batch_size = len(data) pre_src = [x[0] for x in data] pre_tgt = [x[1] for x in data] pre_segs = [x[2] for x in data] pre_clss = [x[3] for x in data] pre_src_sent_labels = [x[4] for x in data] src = torch.tensor(self._pad(pre_src, 0)) tgt = torch.tensor(self._pad(pre_tgt, 0)) segs = torch.tensor(self._pad(pre_segs, 0)) try: mask_src = 1 - (src == 0) mask_tgt = 1 - (tgt == 0) except RuntimeError as err: if 'Subtraction, the `-` operator, with a bool tensor is not supported' not in str(err): raise err mask_src = ~(src == 0) mask_tgt = ~(tgt == 0) clss = torch.tensor(self._pad(pre_clss, -1)) src_sent_labels = torch.tensor(self._pad(pre_src_sent_labels, 0)) try: mask_cls = 1 - (clss == -1) except RuntimeError as err: if 'Subtraction, the `-` operator, with a bool tensor is not supported' not in str(err): raise err mask_cls = ~(clss == -1) clss[clss == -1] = 0 setattr(self, 'clss', clss.to(device)) setattr(self, 'mask_cls', mask_cls.to(device)) setattr(self, 'src_sent_labels', src_sent_labels.to(device)) setattr(self, 'src', src.to(device)) setattr(self, 'tgt', tgt.to(device)) setattr(self, 'segs', segs.to(device)) setattr(self, 'mask_src', mask_src.to(device)) setattr(self, 'mask_tgt', mask_tgt.to(device)) if (is_test) or True: src_str = [x[-3] for x in data] setattr(self, 'src_str', src_str) tgt_str = [x[-2] for x in data] setattr(self, 'tgt_str', tgt_str) topics = [x[-1] for x in data] setattr(self, 'topics', topics) def __len__(self): return self.batch_size def load_dataset(args, corpus_type, shuffle): assert corpus_type in ["train", "valid", "test"] def _lazy_dataset_loader(pt_file, corpus_type, use_topic_modelling): dataset = torch.load(pt_file) et pts = sorted(glob.glob(args.bert_data_path + '.' + corpus_type + '.[0-9]*.pt')) if pts: if (shuffle): random.shuffle(pts) for pt in pts: yield _lazy_dataset_loader(pt, corpus_type, args.use_topic_modelling) else: pt = args.bert_data_path + '.' + corpus_type + '.pt' yield _lazy_dataset_loader(pt, corpus_type, args.use_topic_modelling) def abs_batch_size_fn(new, count): src, tgt = new[0], new[1] global max_n_sents, max_n_tokens, max_size if count == 1: max_size = 0 max_n_sents=0 max_n_tokens=0 max_n_sents = max(max_n_sents, len(tgt)) max_size = max(max_size, max_n_sents) src_elements = count * max_size if (count > 6): return src_elements + 1e3 return src_elements def ext_batch_size_fn(new, count): if (len(new) == 4): pass src, labels = new[0], new[4] global max_n_sents, max_n_tokens, max_size if count == 1: max_size = 0 max_n_sents = 0 max_n_tokens = 0 max_n_sents = max(max_n_sents, len(src)) max_size = max(max_size, max_n_sents) src_elements = count * max_size return src_elements class Dataloader(object): def __init__(self, args, datasets, batch_size, device, shuffle, is_test): self.args = args self.datasets = datasets self.batch_size = batch_size self.device = device self.shuffle = shuffle self.is_test = is_test self.use_topic_modelling = args.use_topic_modelling self.cur_iter = self._next_dataset_iterator(datasets) assert self.cur_iter is not None def __iter__(self): dataset_iter = (d for d in self.datasets) while self.cur_iter is not None: for batch in self.cur_iter: yield batch self.cur_iter = self._next_dataset_iterator(dataset_iter) def _next_dataset_iterator(self, dataset_iter): try: if hasattr(self, "cur_dataset"): self.cur_dataset = None gc.collect() del self.cur_dataset gc.collect() self.cur_dataset = next(dataset_iter) except StopIteration: return None return DataIterator(args = self.args, dataset=self.cur_dataset, batch_size=self.batch_size, device=self.device, shuffle=self.shuffle, is_test=self.is_test) class DataIterator(object): def __init__(self, args, dataset, batch_size, device=None, is_test=False, shuffle=True): self.args = args self.batch_size, self.is_test, self.dataset = batch_size, is_test, dataset self.iterations = 0 self.device = device self.shuffle = shuffle self.sort_key = lambda x: len(x[1]) self._iterations_this_epoch = 0 if (self.args.task == 'abs'): self.batch_size_fn = abs_batch_size_fn else: self.batch_size_fn = ext_batch_size_fn def data(self): if self.shuffle: random.shuffle(self.dataset) xs = self.dataset return xs def preprocess(self, ex, is_test): src = ex['src'] tgt = ex['tgt'][:self.args.max_tgt_len][:-1]+[2] src_sent_labels = ex['src_sent_labels'] segs = ex['segs'] if(not self.args.use_interval): segs=[0]*len(segs) clss = ex['clss'] src_txt = ex['src_txt'] tgt_txt = ex['tgt_txt'] try: topics = ex['topics'] except KeyError: print('Warning: topics are not presented!') topics = None end_id = [src[-1]] src = src[:-1][:self.args.max_pos - 1] + end_id segs = segs[:self.args.max_pos] max_sent_id = bisect.bisect_left(clss, self.args.max_pos) src_sent_labels = src_sent_labels[:max_sent_id] clss = clss[:max_sent_id] if(is_test): return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics else: return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics def batch_buffer(self, data, batch_size): minibatch, size_so_far = [], 0 for ex in data: if(len(ex['src'])==0): continue ex = self.preprocess(ex, self.is_test) if(ex is None): continue minibatch.append(ex) size_so_far = self.batch_size_fn(ex, len(minibatch)) if size_so_far == batch_size: yield minibatch minibatch, size_so_far = [], 0 elif size_so_far > batch_size: yield minibatch[:-1] minibatch, size_so_far = minibatch[-1:], self.batch_size_fn(ex, 1) if minibatch: yield minibatch def batch(self, data, batch_size): minibatch, size_so_far = [], 0 for ex in data: minibatch.append(ex) size_so_far = self.batch_size_fn(ex, len(minibatch)) if size_so_far == batch_size: yield minibatch minibatch, size_so_far = [], 0 elif size_so_far > batch_size: yield minibatch[:-1] minibatch, size_so_far = minibatch[-1:], self.batch_size_fn(ex, 1) if minibatch: yield minibatch def create_batches(self): data = self.data() for buffer in self.batch_buffer(data, self.batch_size * 300): if (self.args.task == 'abs'): p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = sorted(p_batch, key=lambda x: len(x[1])) else: p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = self.batch(p_batch, self.batch_size) p_batch = list(p_batch) if (self.shuffle): random.shuffle(p_batch) for b in p_batch: if(len(b)==0): continue yield b def __iter__(self): while True: self.batches = self.create_batches() for idx, minibatch in enumerate(self.batches): if self._iterations_this_epoch > idx: continue self.iterations += 1 self._iterations_this_epoch += 1 batch = Batch(minibatch, self.device, self.is_test) yield batch return class TextDataloader(object): def __init__(self, args, datasets, batch_size, device, shuffle, is_test): self.args = args self.batch_size = batch_size self.device = device def data(self): if self.shuffle: random.shuffle(self.dataset) xs = self.dataset return xs def preprocess(self, ex, is_test): src = ex['src'] tgt = ex['tgt'][:self.args.max_tgt_len][:-1] + [2] src_sent_labels = ex['src_sent_labels'] segs = ex['segs'] if (not self.args.use_interval): segs = [0] * len(segs) clss = ex['clss'] src_txt = ex['src_txt'] tgt_txt = ex['tgt_txt'] topics = ex['topics'] end_id = [src[-1]] src = src[:-1][:self.args.max_pos - 1] + end_id segs = segs[:self.args.max_pos] max_sent_id = bisect.bisect_left(clss, self.args.max_pos) src_sent_labels = src_sent_labels[:max_sent_id] clss = clss[:max_sent_id] if (is_test): return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics else: return src, tgt, segs, clss, src_sent_labels, src_txt, tgt_txt, topics def batch_buffer(self, data, batch_size): minibatch, size_so_far = [], 0 for ex in data: if (len(ex['src']) == 0): continue ex = self.preprocess(ex, self.is_test) if (ex is None): continue minibatch.append(ex) size_so_far = simple_batch_size_fn(ex, len(minibatch)) if size_so_far == batch_size: yield minibatch minibatch, size_so_far = [], 0 elif size_so_far > batch_size: yield minibatch[:-1] minibatch, size_so_far = minibatch[-1:], simple_batch_size_fn(ex, 1) if minibatch: yield minibatch def create_batches(self): data = self.data() for buffer in self.batch_buffer(data, self.batch_size * 300): if (self.args.task == 'abs'): p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = sorted(p_batch, key=lambda x: len(x[1])) else: p_batch = sorted(buffer, key=lambda x: len(x[2])) p_batch = batch(p_batch, self.batch_size) p_batch = batch(p_batch, self.batch_size) p_batch = list(p_batch) if (self.shuffle): random.shuffle(p_batch) for b in p_batch: if (len(b) == 0): continue yield b def __iter__(self): while True: self.batches = self.create_batches() for idx, minibatch in enumerate(self.batches): if self._iterations_this_epoch > idx: continue self.iterations += 1 self._iterations_this_epoch += 1 batch = Batch(minibatch, self.device, self.is_test) yield batch return
true
true
1c4831c92bce75f0f171e01cc007b17f4ff0e01b
291
py
Python
tests/urls.py
movermeyer/django-umanage
9327772efbf1f13c05b22afcbccaebb2c8595850
[ "MIT" ]
4
2015-04-21T01:01:23.000Z
2016-01-15T08:41:56.000Z
tests/urls.py
movermeyer/django-umanage
9327772efbf1f13c05b22afcbccaebb2c8595850
[ "MIT" ]
1
2018-03-04T20:46:41.000Z
2018-03-04T20:46:41.000Z
tests/urls.py
movermeyer/django-umanage
9327772efbf1f13c05b22afcbccaebb2c8595850
[ "MIT" ]
3
2017-08-14T01:53:44.000Z
2019-06-06T17:47:49.000Z
from django.conf.urls import include from django.conf.urls import url urlpatterns = [ url(r'', include('umanage.auth.urls')), url(r'', include('umanage.forgot_username.urls')), url(r'', include('umanage.forgot_password.urls')), url(r'^account', include('umanage.urls')), ]
26.454545
54
0.680412
from django.conf.urls import include from django.conf.urls import url urlpatterns = [ url(r'', include('umanage.auth.urls')), url(r'', include('umanage.forgot_username.urls')), url(r'', include('umanage.forgot_password.urls')), url(r'^account', include('umanage.urls')), ]
true
true
1c4833f8843bf3183d8117ca678adcf0a5a840f2
4,689
py
Python
kubernetes_spawner/swagger_client/models/unversioned_list_meta.py
AdrianGPrado/k8s-jupyterhub-spawner
f3d28adf1d70102bc60ba57f5737a7ec864537d9
[ "Apache-2.0" ]
16
2016-09-18T21:20:49.000Z
2020-02-15T06:28:03.000Z
kubernetes_spawner/swagger_client/models/unversioned_list_meta.py
AdrianGPrado/k8s-jupyterhub-spawner
f3d28adf1d70102bc60ba57f5737a7ec864537d9
[ "Apache-2.0" ]
2
2016-11-10T17:51:55.000Z
2018-03-18T05:38:22.000Z
kubernetes_spawner/swagger_client/models/unversioned_list_meta.py
AdrianGPrado/k8s-jupyterhub-spawner
f3d28adf1d70102bc60ba57f5737a7ec864537d9
[ "Apache-2.0" ]
12
2016-09-28T20:48:56.000Z
2020-01-17T04:50:59.000Z
# coding: utf-8 """ Copyright 2015 SmartBear Software Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. Ref: https://github.com/swagger-api/swagger-codegen """ from pprint import pformat from six import iteritems class UnversionedListMeta(object): """ NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. """ def __init__(self): """ UnversionedListMeta - a model defined in Swagger :param dict swaggerTypes: The key is attribute name and the value is attribute type. :param dict attributeMap: The key is attribute name and the value is json key in definition. """ self.swagger_types = { 'self_link': 'str', 'resource_version': 'str' } self.attribute_map = { 'self_link': 'selfLink', 'resource_version': 'resourceVersion' } self._self_link = None self._resource_version = None @property def self_link(self): """ Gets the self_link of this UnversionedListMeta. SelfLink is a URL representing this object. Populated by the system. Read-only. :return: The self_link of this UnversionedListMeta. :rtype: str """ return self._self_link @self_link.setter def self_link(self, self_link): """ Sets the self_link of this UnversionedListMeta. SelfLink is a URL representing this object. Populated by the system. Read-only. :param self_link: The self_link of this UnversionedListMeta. :type: str """ self._self_link = self_link @property def resource_version(self): """ Gets the resource_version of this UnversionedListMeta. String that identifies the server's internal version of this object that can be used by clients to determine when objects have changed. Value must be treated as opaque by clients and passed unmodified back to the server. Populated by the system. Read-only. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#concurrency-control-and-consistency :return: The resource_version of this UnversionedListMeta. :rtype: str """ return self._resource_version @resource_version.setter def resource_version(self, resource_version): """ Sets the resource_version of this UnversionedListMeta. String that identifies the server's internal version of this object that can be used by clients to determine when objects have changed. Value must be treated as opaque by clients and passed unmodified back to the server. Populated by the system. Read-only. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#concurrency-control-and-consistency :param resource_version: The resource_version of this UnversionedListMeta. :type: str """ self._resource_version = resource_version def to_dict(self): """ Returns the model properties as a dict """ result = {} for attr, _ in iteritems(self.swagger_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() else: result[attr] = value return result def to_str(self): """ Returns the string representation of the model """ return pformat(self.to_dict()) def __repr__(self): """ For `print` and `pprint` """ return self.to_str() def __eq__(self, other): """ Returns true if both objects are equal """ return self.__dict__ == other.__dict__ def __ne__(self, other): """ Returns true if both objects are not equal """ return not self == other
33.492857
369
0.629132
from pprint import pformat from six import iteritems class UnversionedListMeta(object): def __init__(self): self.swagger_types = { 'self_link': 'str', 'resource_version': 'str' } self.attribute_map = { 'self_link': 'selfLink', 'resource_version': 'resourceVersion' } self._self_link = None self._resource_version = None @property def self_link(self): return self._self_link @self_link.setter def self_link(self, self_link): self._self_link = self_link @property def resource_version(self): return self._resource_version @resource_version.setter def resource_version(self, resource_version): self._resource_version = resource_version def to_dict(self): result = {} for attr, _ in iteritems(self.swagger_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() else: result[attr] = value return result def to_str(self): return pformat(self.to_dict()) def __repr__(self): return self.to_str() def __eq__(self, other): return self.__dict__ == other.__dict__ def __ne__(self, other): return not self == other
true
true
1c48346c9bf89ec65e95fe3f36582e8e59a98ca7
2,281
py
Python
account/forms.py
mijiFernandes/pa4
3850c5bf6af1f89cd3876b08c9a9fa319d583fae
[ "Unlicense" ]
null
null
null
account/forms.py
mijiFernandes/pa4
3850c5bf6af1f89cd3876b08c9a9fa319d583fae
[ "Unlicense" ]
null
null
null
account/forms.py
mijiFernandes/pa4
3850c5bf6af1f89cd3876b08c9a9fa319d583fae
[ "Unlicense" ]
null
null
null
from django import forms from django.contrib.auth.forms import ReadOnlyPasswordHashField from django.utils.translation import ugettext_lazy as _ from account.models import User, UserManager class UserCreationForm(forms.ModelForm): # 사용자 생성 폼 username = forms.CharField( label=_('Username'), required=True, widget=forms.TextInput( attrs={ 'class': 'form-control', 'placeholder': _('Username'), 'required': 'True', } ) ) password1 = forms.CharField( label=_('Password'), widget=forms.PasswordInput( attrs={ 'class': 'form-control', 'placeholder': _('Password'), 'required': 'True', } ) ) password2 = forms.CharField( label=_('Password confirmation'), widget=forms.PasswordInput( attrs={ 'class': 'form-control', 'placeholder': _('Password confirmation'), 'required': 'True', } ) ) class Meta: model = User fields = ('username',) def clean_password2(self): # 두 비밀번호 입력 일치 확인 password1 = self.cleaned_data.get("password1") password2 = self.cleaned_data.get("password2") if password1 and password2 and password1 != password2: raise forms.ValidationError("Passwords don't match") return password2 def save(self, commit=True): # Save the provided password in hashed format user = super(UserCreationForm, self).save(commit=False) user.set_password(self.cleaned_data["password1"]) if commit: user.save() return user class UserChangeForm(forms.ModelForm): # 비밀번호 변경 폼 password = ReadOnlyPasswordHashField( label=_('Password') ) class Meta: model = User fields = ('username', 'password', 'is_active', 'is_superuser') def clean_password(self): # Regardless of what the user provides, return the initial value. # This is done here, rather than on the field, because the # field does not have access to the initial value return self.initial["password"]
29.24359
73
0.577378
from django import forms from django.contrib.auth.forms import ReadOnlyPasswordHashField from django.utils.translation import ugettext_lazy as _ from account.models import User, UserManager class UserCreationForm(forms.ModelForm): username = forms.CharField( label=_('Username'), required=True, widget=forms.TextInput( attrs={ 'class': 'form-control', 'placeholder': _('Username'), 'required': 'True', } ) ) password1 = forms.CharField( label=_('Password'), widget=forms.PasswordInput( attrs={ 'class': 'form-control', 'placeholder': _('Password'), 'required': 'True', } ) ) password2 = forms.CharField( label=_('Password confirmation'), widget=forms.PasswordInput( attrs={ 'class': 'form-control', 'placeholder': _('Password confirmation'), 'required': 'True', } ) ) class Meta: model = User fields = ('username',) def clean_password2(self): password1 = self.cleaned_data.get("password1") password2 = self.cleaned_data.get("password2") if password1 and password2 and password1 != password2: raise forms.ValidationError("Passwords don't match") return password2 def save(self, commit=True): # Save the provided password in hashed format user = super(UserCreationForm, self).save(commit=False) user.set_password(self.cleaned_data["password1"]) if commit: user.save() return user class UserChangeForm(forms.ModelForm): # 비밀번호 변경 폼 password = ReadOnlyPasswordHashField( label=_('Password') ) class Meta: model = User fields = ('username', 'password', 'is_active', 'is_superuser') def clean_password(self): # Regardless of what the user provides, return the initial value. # This is done here, rather than on the field, because the # field does not have access to the initial value return self.initial["password"]
true
true
1c4834ad7231269805601500145db87c940a6876
2,080
py
Python
src/aiosdnotify/__init__.py
vivienm/python-aiosdnotify
b0fe62bccf55041b00f65d395bea96c0964de9a4
[ "MIT" ]
null
null
null
src/aiosdnotify/__init__.py
vivienm/python-aiosdnotify
b0fe62bccf55041b00f65d395bea96c0964de9a4
[ "MIT" ]
null
null
null
src/aiosdnotify/__init__.py
vivienm/python-aiosdnotify
b0fe62bccf55041b00f65d395bea96c0964de9a4
[ "MIT" ]
null
null
null
import asyncio import logging import os import socket from abc import ABC, abstractmethod from asyncio.events import AbstractEventLoop from typing import Optional, Union logger = logging.getLogger(__name__) __version__ = "0.1.0" class AbstractNotifier(ABC): @abstractmethod async def connect(self) -> None: pass @abstractmethod async def close(self) -> None: pass @abstractmethod async def notify(self, state: Union[str, bytes]) -> None: pass async def __aenter__(self): await self.connect() return self async def __aexit__(self, exc_type, exc_value, traceback): await self.close() class SystemdNotifier(AbstractNotifier): __slots__ = ( "addr", "sock", "loop", ) def __init__( self, addr: Optional[str] = None, *, loop: Optional[AbstractEventLoop] = None, ) -> None: self.addr = addr or os.environ["NOTIFY_SOCKET"] self.sock = socket.socket(socket.AF_UNIX, socket.SOCK_DGRAM) self.loop = loop or asyncio.get_event_loop() async def connect(self) -> None: await self.loop.sock_connect(self.sock, self.addr) async def close(self) -> None: self.sock.close() async def notify(self, state: Union[str, bytes]) -> None: if isinstance(state, str): state = state.encode() await self.loop.sock_sendall(self.sock, state) class DummyNotifier(AbstractNotifier): async def connect(self) -> None: pass async def close(self) -> None: pass async def notify(self, state: Union[str, bytes]) -> None: pass def notifier( addr: Optional[str] = None, *, loop: Optional[AbstractEventLoop] = None, ) -> AbstractNotifier: if addr or "NOTIFY_SOCKET" in os.environ: return SystemdNotifier(addr=addr, loop=loop) else: logger.warning( "Could not determine systemd socket address, " "systemd notifications are disabled", ) return DummyNotifier()
23.111111
68
0.625962
import asyncio import logging import os import socket from abc import ABC, abstractmethod from asyncio.events import AbstractEventLoop from typing import Optional, Union logger = logging.getLogger(__name__) __version__ = "0.1.0" class AbstractNotifier(ABC): @abstractmethod async def connect(self) -> None: pass @abstractmethod async def close(self) -> None: pass @abstractmethod async def notify(self, state: Union[str, bytes]) -> None: pass async def __aenter__(self): await self.connect() return self async def __aexit__(self, exc_type, exc_value, traceback): await self.close() class SystemdNotifier(AbstractNotifier): __slots__ = ( "addr", "sock", "loop", ) def __init__( self, addr: Optional[str] = None, *, loop: Optional[AbstractEventLoop] = None, ) -> None: self.addr = addr or os.environ["NOTIFY_SOCKET"] self.sock = socket.socket(socket.AF_UNIX, socket.SOCK_DGRAM) self.loop = loop or asyncio.get_event_loop() async def connect(self) -> None: await self.loop.sock_connect(self.sock, self.addr) async def close(self) -> None: self.sock.close() async def notify(self, state: Union[str, bytes]) -> None: if isinstance(state, str): state = state.encode() await self.loop.sock_sendall(self.sock, state) class DummyNotifier(AbstractNotifier): async def connect(self) -> None: pass async def close(self) -> None: pass async def notify(self, state: Union[str, bytes]) -> None: pass def notifier( addr: Optional[str] = None, *, loop: Optional[AbstractEventLoop] = None, ) -> AbstractNotifier: if addr or "NOTIFY_SOCKET" in os.environ: return SystemdNotifier(addr=addr, loop=loop) else: logger.warning( "Could not determine systemd socket address, " "systemd notifications are disabled", ) return DummyNotifier()
true
true
1c483537f12b82976dc803943a2780c94717e1c9
2,119
py
Python
flaskapp/utils.py
crockmitnic/question-paper-generator
3f5339226aedd4332c562913945a08cdb45983b0
[ "MIT" ]
6
2020-08-02T20:58:34.000Z
2022-03-23T20:33:20.000Z
flaskapp/utils.py
crockmitnic/question-paper-generator
3f5339226aedd4332c562913945a08cdb45983b0
[ "MIT" ]
209
2020-02-12T17:09:15.000Z
2021-06-03T20:34:35.000Z
flaskapp/utils.py
crockmitnic/question-paper-generator
3f5339226aedd4332c562913945a08cdb45983b0
[ "MIT" ]
54
2020-02-18T14:54:35.000Z
2021-09-05T06:31:12.000Z
import os from enum import Enum from flask import url_for from flask_login import current_user from itsdangerous import URLSafeSerializer json_url = URLSafeSerializer(os.environ.get("SECRET_KEY", "secret_key")) class AbstractEnum(Enum): @classmethod def from_string(cls, value): return cls.__members__[value] class CognitiveEnum(AbstractEnum): Knowledge = 1 Comprehension = 2 Application = 3 class DifficultyEnum(AbstractEnum): Easy = 1 Medium = 2 Hard = 3 class QuestionTypeEnum(AbstractEnum): sub = 1 mcq = 2 def profile_path(): """get the profile path of user Returns: URL : if user is authentic then return url of user """ if current_user.is_authenticated: return url_for("static", filename="profile_pics/" + current_user.image_file) return "" default_instructions = [ "Write your name and student number in the space provided", "Make sure your mobile phone is switched off and place it at the front together with\ any bags, books, coats etc. Then find your seat.", "Remember that talking is not allowed at any time in the exam hall.", "Listen carefully to instructions. Students are required to comply with\ the instructions of invigilators at all times.", "You are not permitted to share stationery, \ calculators or any other materials during the examination.", "If you have a question or need more papers, raise your hand and a teacher\ will come to you. Teachers will not give hints or answers, so please do not ask for them.", "Stop writing immediately when the teacher says it is the end of the exam.", "Leave the exam hall quickly and quietly. Remember to take all your belongings with you.\ (Remember to collect all your belongings from holding rooms.)\ You must remain silent until after you have exited the building.", "Remember! Any form of cheating is not allowed and action will be taken.", ]
33.634921
116
0.668712
import os from enum import Enum from flask import url_for from flask_login import current_user from itsdangerous import URLSafeSerializer json_url = URLSafeSerializer(os.environ.get("SECRET_KEY", "secret_key")) class AbstractEnum(Enum): @classmethod def from_string(cls, value): return cls.__members__[value] class CognitiveEnum(AbstractEnum): Knowledge = 1 Comprehension = 2 Application = 3 class DifficultyEnum(AbstractEnum): Easy = 1 Medium = 2 Hard = 3 class QuestionTypeEnum(AbstractEnum): sub = 1 mcq = 2 def profile_path(): if current_user.is_authenticated: return url_for("static", filename="profile_pics/" + current_user.image_file) return "" default_instructions = [ "Write your name and student number in the space provided", "Make sure your mobile phone is switched off and place it at the front together with\ any bags, books, coats etc. Then find your seat.", "Remember that talking is not allowed at any time in the exam hall.", "Listen carefully to instructions. Students are required to comply with\ the instructions of invigilators at all times.", "You are not permitted to share stationery, \ calculators or any other materials during the examination.", "If you have a question or need more papers, raise your hand and a teacher\ will come to you. Teachers will not give hints or answers, so please do not ask for them.", "Stop writing immediately when the teacher says it is the end of the exam.", "Leave the exam hall quickly and quietly. Remember to take all your belongings with you.\ (Remember to collect all your belongings from holding rooms.)\ You must remain silent until after you have exited the building.", "Remember! Any form of cheating is not allowed and action will be taken.", ]
true
true
1c48357c46224c703897df7c26b65ca54e06f218
546
py
Python
layout/serializers/sidebarSubsectionSerializer.py
rankrh/soli
8f19945a175106064591d09a53d07fcbfa26b7da
[ "MIT" ]
null
null
null
layout/serializers/sidebarSubsectionSerializer.py
rankrh/soli
8f19945a175106064591d09a53d07fcbfa26b7da
[ "MIT" ]
null
null
null
layout/serializers/sidebarSubsectionSerializer.py
rankrh/soli
8f19945a175106064591d09a53d07fcbfa26b7da
[ "MIT" ]
2
2019-09-07T15:10:14.000Z
2020-09-04T01:51:19.000Z
from rest_framework import serializers class SidebarSubsectionSerializer(serializers.Serializer): name = serializers.CharField(read_only=True, max_length=32) id = serializers.CharField(read_only=True, max_length=16) icon = serializers.CharField(read_only=True, max_length=32) url = serializers.CharField(read_only=True, max_length=64) url_params = serializers.ListField( allow_empty=True, read_only=True, required=False, max_length=64 ) subsection_name = serializers.CharField(read_only=True, max_length=32)
42
74
0.776557
from rest_framework import serializers class SidebarSubsectionSerializer(serializers.Serializer): name = serializers.CharField(read_only=True, max_length=32) id = serializers.CharField(read_only=True, max_length=16) icon = serializers.CharField(read_only=True, max_length=32) url = serializers.CharField(read_only=True, max_length=64) url_params = serializers.ListField( allow_empty=True, read_only=True, required=False, max_length=64 ) subsection_name = serializers.CharField(read_only=True, max_length=32)
true
true
1c4836932fda8321a41ab0f347753291a7915da9
2,595
py
Python
venv/Lib/site-packages/pyrogram/raw/types/input_photo.py
iamgeorgiy/heroku-userbot
5a92417d16f8ead949d88cb38da213fc2da5d3a4
[ "Apache-2.0" ]
null
null
null
venv/Lib/site-packages/pyrogram/raw/types/input_photo.py
iamgeorgiy/heroku-userbot
5a92417d16f8ead949d88cb38da213fc2da5d3a4
[ "Apache-2.0" ]
null
null
null
venv/Lib/site-packages/pyrogram/raw/types/input_photo.py
iamgeorgiy/heroku-userbot
5a92417d16f8ead949d88cb38da213fc2da5d3a4
[ "Apache-2.0" ]
null
null
null
# Pyrogram - Telegram MTProto API Client Library for Python # Copyright (C) 2017-2020 Dan <https://github.com/delivrance> # # This file is part of Pyrogram. # # Pyrogram is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Pyrogram is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with Pyrogram. If not, see <http://www.gnu.org/licenses/>. from io import BytesIO from pyrogram.raw.core.primitives import Int, Long, Int128, Int256, Bool, Bytes, String, Double, Vector from pyrogram.raw.core import TLObject from pyrogram import raw from typing import List, Union, Any # # # # # # # # # # # # # # # # # # # # # # # # # !!! WARNING !!! # # This is a generated file! # # All changes made in this file will be lost! # # # # # # # # # # # # # # # # # # # # # # # # # class InputPhoto(TLObject): # type: ignore """This object is a constructor of the base type :obj:`~pyrogram.raw.base.InputPhoto`. Details: - Layer: ``117`` - ID: ``0x3bb3b94a`` Parameters: id: ``int`` ``64-bit`` access_hash: ``int`` ``64-bit`` file_reference: ``bytes`` """ __slots__: List[str] = ["id", "access_hash", "file_reference"] ID = 0x3bb3b94a QUALNAME = "types.InputPhoto" def __init__(self, *, id: int, access_hash: int, file_reference: bytes) -> None: self.id = id # long self.access_hash = access_hash # long self.file_reference = file_reference # bytes @staticmethod def read(data: BytesIO, *args: Any) -> "InputPhoto": # No flags id = Long.read(data) access_hash = Long.read(data) file_reference = Bytes.read(data) return InputPhoto(id=id, access_hash=access_hash, file_reference=file_reference) def write(self) -> bytes: data = BytesIO() data.write(Int(self.ID, False)) # No flags data.write(Long(self.id)) data.write(Long(self.access_hash)) data.write(Bytes(self.file_reference)) return data.getvalue()
32.037037
103
0.614644
from io import BytesIO from pyrogram.raw.core.primitives import Int, Long, Int128, Int256, Bool, Bytes, String, Double, Vector from pyrogram.raw.core import TLObject from pyrogram import raw from typing import List, Union, Any
true
true
1c4836bb5e77d9e5039fcadff0c472241dcdc667
11,501
py
Python
parser/parser.py
samkhal/course-eval-viewer
d8afdd48247964a8a1a0e8fde6a5c1e09a82ca34
[ "MIT" ]
1
2016-12-19T08:49:10.000Z
2016-12-19T08:49:10.000Z
parser/parser.py
samkhal/course-eval-viewer
d8afdd48247964a8a1a0e8fde6a5c1e09a82ca34
[ "MIT" ]
null
null
null
parser/parser.py
samkhal/course-eval-viewer
d8afdd48247964a8a1a0e8fde6a5c1e09a82ca34
[ "MIT" ]
null
null
null
#!/usr/bin/python from __future__ import print_function from eval_db import Database, Evaluation, QuestionInstance, AnswerField, FailedScrape from bs4 import BeautifulSoup from MySQLdb import DatabaseError import re import sys import requests import getpass import traceback import getopt # Python 3 improvements: flush prints insted of using -u # super lookes nicer # print doesn't need future import # better re-raise with raise from, tracebacks # There can be two courses with same year and CRN (and name, code, section, etc), but different profs # (ex: 2015 ADLN, CRN=30071) # Don't trust means on course selection page #!! notes: dont go by prof, get all at once by not giving prof param # There's a Summer term # Sections can be 1 digit # Which E term is hard to figure out. Usually E01 or E02, sometimes E04 # Use siblings to do the find_all, or just check IN_CRN for each href #!! Add weights for text answers # Use Postgre Materialized Views to avoid recomputing views pages = { 'home': 'https://bannerweb.wpi.edu', 'login': 'https://bannerweb.wpi.edu/pls/prod/twbkwbis.P_WWWLogin', 'validate': 'https://bannerweb.wpi.edu/pls/prod/twbkwbis.P_ValLogin', 'sel_year_adln': 'https://bannerweb.wpi.edu/pls/prod/hwwkscevrp.P_Select_Year', 'sel_course_inst': 'https://bannerweb.wpi.edu/pls/prod/hwwkscevrp.P_Select_CrseInst', 'sel_section': 'https://bannerweb.wpi.edu/pls/prod/hwwkscevrp.P_Select_CrseSect' } # Allow reraising other exceptions as ParseErrors while preserving info class ParseError(Exception): def __init__(self, message="", cause=None): cause_str = "caused by " + repr(cause) if cause else "" punc = ", " if message and cause else "" full_message = message + punc + cause_str if full_message: super(ParseError, self).__init__(full_message) self.cause = cause class LoginError(Exception): pass class Downloader(object): def __init__(self, user, passwd): self.user = user self.passwd = passwd self.sess = requests.Session() def login(self): print("\nLogging in... ", end='') self.sess.get(pages['home']) response = self.sess.post(pages['validate'], params = { 'sid': self.user, 'PIN': self.passwd}, headers = { 'referer': pages['login'] }) # Check cookies for success if 'SESSID' not in self.sess.cookies: raise LoginError('Login unsuccessful') else: print("Success") def download_eval(self, url): if 'SESSID' not in self.sess.cookies: self.login() response = self.sess.get(pages['home'] + url, timeout=60) return response.text # Return list of urls for year, adln, and crn filters. crns is list of strings. def get_urls_by_year(self, year, adln=False, crns=[]): if 'SESSID' not in self.sess.cookies: self.login() print("Downloading list of {} classes for {}-{}... ".format( "ADLN" if adln else "non-ADLN", year-1,year), end='') response = self.sess.post(pages['sel_section'], params = { 'IN_SUBCRSE': '', 'IN_PIDM': '', 'IN_ACYR': year, 'IN_ADLN_OIX': "O" if adln else "X" }, timeout=600) print("Done") soup = BeautifulSoup(response.text, 'lxml') if crns: # Filter urls by provided list of crns regex_str = r".*IN_CRN=(?:" + "|".join(crns) + ")" else: # If no list provided, get all crns regex_str = r".*IN_TYPE=C" class_links = (soup.find('table', class_='datadisplaytable') .find_all('a', href=re.compile(regex_str))) # Return a list so we have its length available return [link['href'] for link in class_links] def parse_eval(text): # Replace any &nbsp; in the html with spaces #text = unicode(text, 'utf-8') # Expects unicode text text = text.replace(u'\xa0',' ') soup = BeautifulSoup(text, "lxml") # Extract all class info from the header ev = Evaluation() try: ev.AcademicYear = re.compile(r"Academic Year \d{4}-(\d{4})").search(text).group(1) crn_re = re.compile(r"CRN (\d+)") crn_str = soup.find(string=crn_re) header_row = str(crn_str.parent.parent) ev.CRN = crn_re.search(crn_str).group(1) # Decompose parts of full course name course = re.compile(r"<b>([A-Z]{2,4})-([\dX]{3,4}) (.+?)</b>").search(header_row) ev.Department = course.group(1) ev.Code = course.group(2) ev.Name = course.group(3) # Identify term from section or season ev.Section = section = re.compile(r"Section (\w+)").search(header_row).group(1) if section[0] in 'ABCD': ev.TermName = section[0] #elif section[0:2] in ['E1','E2']: # ev.TermName = section[0:2] else: ev.TermName = re.compile(r"(Spring|Fall|Summer) \d{4}").search(header_row).group(1) ev.Instructor = re.compile(r"<TH.*?>Prof\. (.*?)</TH>").search(text).group(1) except (AttributeError,IndexError) as e: raise ParseError("Error parsing course info",e), None, sys.exc_info()[2] # Extract all questions, possible answers, and answer values try: question_re = re.compile(r"(\d{1,2}[A-Z]?)\. (.+)") question_tags = soup.find_all("td", class_="dddefault", string=question_re) for q_tag in question_tags: # Pull out question number and text question_num, question_str = question_re.search(q_tag.string).groups() question = QuestionInstance(Num=question_num, FullString=question_str) # Pull out answer text (if available) and responses row_cells = q_tag.parent.find_all("td", class_="dddefault") if len(row_cells)==8: # All data on one row, 1-5 response question.answers = [AnswerField( Weight=i+1, Respondents=row_cells[i+1].p.contents[1]) for i in range(5)] elif len(row_cells)==1: # Answers are one per row on next 5 rows for row in q_tag.parent.next_siblings: if row.name=="tr": #Break when we get to the next question header if row.find("th", class_="ddtitle"): break cells = row.find_all("td", class_="dddefault") answer = AnswerField(AnswerText=cells[0].string, Respondents=cells[1].string) question.answers.append(answer) else: raise ParseError("Unexpected number of cells in row") ev.questions.append(question) except (AttributeError,IndexError) as e: raise ParseError("Error parsing evaluation results",e), None, sys.exc_info()[2] return ev # Only call options: # process_evals(db, dl, year, adln) # process_evals(db, dl, year, adln, crns) # process_evals(db, dl, urls) def process_evals(db, dl, year=None, adln=False, crns=[], urls=[]): use_urls = bool(urls) if not urls: urls = dl.get_urls_by_year(year, adln, crns) for idx, url in enumerate(urls): crn = re.compile('IN_CRN=(\d*)').search(url).group(1) if use_urls: #extract year, adln from url if needed year = re.compile('IN_ACYR=(\d*)').search(url).group(1) adln_char = re.compile('IN_ADLN_OIX=(\d*)').search(url).group(1) if adln_char not in 'OX': raise ParseError("ADLN in URL must be O or X") adln = adln_char=='O' print("\rDownloading and parsing class {}/{} (CRN {})".format(idx+1,len(urls),crn), end='') try: eval_text = dl.download_eval(url) ev = parse_eval(eval_text) ev.ADLN = adln db.store(ev) except ParseError as e: #_, _, tb = sys.exc_info() print("\nError parsing CRN {}:".format(crn)) traceback.print_exc() print() db.store(FailedScrape(CRN=crn, AcademicYear=year, ADLN=adln)) #print(crn, year, adln, file=failed_crns) except DatabaseError as e: print("\nError storing CRN {}: {}\n".format(crn, repr(e))) db.store(FailedScrape(CRN=crn, AcademicYear=year, ADLN=adln)) #print(crn, file=failed_crns) except requests.exceptions.Timeout: print("\nTimed out while loading CRN {}".format(crn)) db.store(FailedScrape(CRN=crn, AcademicYear=year, ADLN=adln)) #!!Add reason to failed table print("\nDone") def process_failed_scrapes(db, dl): grouped_crns = db.get_col_grouped_by(FailedScrape,'CRN',['AcademicYear','ADLN']) db.clear_table(FailedScrape) for group in grouped_crns: process_evals(db, dl, group['AcademicYear'], group['ADLN'], group['CRN']) #!! # check all courses for uniqueness of crns def validate_assumptions(db, dl): pass # Check that all available courses are accounted for in db or failures # Does not identify duplicate CRNS #!! # untested #!! def validate_and_fix(db, dl, fix=True): #years = dl.get_years() #!!! years = range(2006,2017) years = [2010,2013,2016] #!!! remove missing_urls = [] for year in years: missing_in_year = [] urls = dl.get_urls_by_year(year, adln=True) urls += dl.get_urls_by_year(year, adln=False) for url in urls: #!! get this from func crn = re.compile('IN_CRN=(\d*)').search(url).group(1) #!! adln_char = re.compile('IN_ADLN_OIX=(\d*)').search(url).group(1) if adln_char not in 'OX': raise ParseError("ADLN in URL must be O or X") adln = adln_char=='O' db.cur.execute("""SELECT 1 FROM ( SELECT CRN FROM Classes WHERE AcademicYear = {year} AND ADLN={adln} UNION ALL SELECT CRN FROM FailedScrapes WHERE AcademicYear = {year} AND ADLN={adln} ) a WHERE CRN = {crn}""".format(year=year,adln=adln,crn=crn)) if not db.cur.fetchone(): # Entry not found missing_in_year.append(url) print("{} missing entries in {}-{} out of {}".format(len(missing_in_year),year-1,year,len(urls))) missing_urls += missing_in_year if missing_urls: print("{} total missing entries".format(len(missing_urls))) if fix: process_evals(db, dl, urls=missing_urls) else: print("No missing entries found") #!! notify about failed scrapes def main(argv): try: opts, args = getopt.getopt(argv, "ry:fvb") except getopt.GetoptError: print("parser.py [-r] [-y YEAR | -f] [-v]") #!! return reset = False rerun_fails = False year = None validate = False build_sums = False for opt, arg in opts: if opt =='-r': reset = True elif opt =='-y': year = arg elif opt =='-f': rerun_fails = True elif opt == '-v': validate = True elif opt == '-b': build_sums = True db = Database() dl = Downloader(user=raw_input("Username: "), passwd=getpass.getpass()) dl.login() if reset: db.reset() if rerun_fails: print("Re-running failed scrapes") process_failed_scrapes(db, dl) elif year: process_evals(db, dl, 2015, adln=True) process_evals(db, dl, 2015, adln=False) elif validate: validate_and_fix(db, dl) elif build_sums: pass else: for year in range(2016,2017): process_evals(db, dl, year, adln=True) process_evals(db, dl, year, adln=False) #!! process all years db.build_class_summaries() # with open("sampleeval.html") as f: # ev = parse_eval(f.read()) # db.store(ev) # with open("sampleeval2.html") as f: # ev = parse_eval(f.read()) # db.store(ev) # print("Complete") if __name__ == "__main__": main(sys.argv[1:])
32.125698
101
0.63151
from __future__ import print_function from eval_db import Database, Evaluation, QuestionInstance, AnswerField, FailedScrape from bs4 import BeautifulSoup from MySQLdb import DatabaseError import re import sys import requests import getpass import traceback import getopt # better re-raise with raise from, tracebacks # There can be two courses with same year and CRN (and name, code, section, etc), but different profs # (ex: 2015 ADLN, CRN=30071) # Don't trust means on course selection page # Sections can be 1 digit # Which E term is hard to figure out. Usually E01 or E02, sometimes E04 # Use siblings to do the find_all, or just check IN_CRN for each href #!! Add weights for text answers # Use Postgre Materialized Views to avoid recomputing views pages = { 'home': 'https://bannerweb.wpi.edu', 'login': 'https://bannerweb.wpi.edu/pls/prod/twbkwbis.P_WWWLogin', 'validate': 'https://bannerweb.wpi.edu/pls/prod/twbkwbis.P_ValLogin', 'sel_year_adln': 'https://bannerweb.wpi.edu/pls/prod/hwwkscevrp.P_Select_Year', 'sel_course_inst': 'https://bannerweb.wpi.edu/pls/prod/hwwkscevrp.P_Select_CrseInst', 'sel_section': 'https://bannerweb.wpi.edu/pls/prod/hwwkscevrp.P_Select_CrseSect' } # Allow reraising other exceptions as ParseErrors while preserving info class ParseError(Exception): def __init__(self, message="", cause=None): cause_str = "caused by " + repr(cause) if cause else "" punc = ", " if message and cause else "" full_message = message + punc + cause_str if full_message: super(ParseError, self).__init__(full_message) self.cause = cause class LoginError(Exception): pass class Downloader(object): def __init__(self, user, passwd): self.user = user self.passwd = passwd self.sess = requests.Session() def login(self): print("\nLogging in... ", end='') self.sess.get(pages['home']) response = self.sess.post(pages['validate'], params = { 'sid': self.user, 'PIN': self.passwd}, headers = { 'referer': pages['login'] }) # Check cookies for success if 'SESSID' not in self.sess.cookies: raise LoginError('Login unsuccessful') else: print("Success") def download_eval(self, url): if 'SESSID' not in self.sess.cookies: self.login() response = self.sess.get(pages['home'] + url, timeout=60) return response.text # Return list of urls for year, adln, and crn filters. crns is list of strings. def get_urls_by_year(self, year, adln=False, crns=[]): if 'SESSID' not in self.sess.cookies: self.login() print("Downloading list of {} classes for {}-{}... ".format( "ADLN" if adln else "non-ADLN", year-1,year), end='') response = self.sess.post(pages['sel_section'], params = { 'IN_SUBCRSE': '', 'IN_PIDM': '', 'IN_ACYR': year, 'IN_ADLN_OIX': "O" if adln else "X" }, timeout=600) print("Done") soup = BeautifulSoup(response.text, 'lxml') if crns: # Filter urls by provided list of crns regex_str = r".*IN_CRN=(?:" + "|".join(crns) + ")" else: # If no list provided, get all crns regex_str = r".*IN_TYPE=C" class_links = (soup.find('table', class_='datadisplaytable') .find_all('a', href=re.compile(regex_str))) # Return a list so we have its length available return [link['href'] for link in class_links] def parse_eval(text): # Replace any &nbsp; in the html with spaces #text = unicode(text, 'utf-8') # Expects unicode text text = text.replace(u'\xa0',' ') soup = BeautifulSoup(text, "lxml") # Extract all class info from the header ev = Evaluation() try: ev.AcademicYear = re.compile(r"Academic Year \d{4}-(\d{4})").search(text).group(1) crn_re = re.compile(r"CRN (\d+)") crn_str = soup.find(string=crn_re) header_row = str(crn_str.parent.parent) ev.CRN = crn_re.search(crn_str).group(1) # Decompose parts of full course name course = re.compile(r"<b>([A-Z]{2,4})-([\dX]{3,4}) (.+?)</b>").search(header_row) ev.Department = course.group(1) ev.Code = course.group(2) ev.Name = course.group(3) # Identify term from section or season ev.Section = section = re.compile(r"Section (\w+)").search(header_row).group(1) if section[0] in 'ABCD': ev.TermName = section[0] #elif section[0:2] in ['E1','E2']: # ev.TermName = section[0:2] else: ev.TermName = re.compile(r"(Spring|Fall|Summer) \d{4}").search(header_row).group(1) ev.Instructor = re.compile(r"<TH.*?>Prof\. (.*?)</TH>").search(text).group(1) except (AttributeError,IndexError) as e: raise ParseError("Error parsing course info",e), None, sys.exc_info()[2] # Extract all questions, possible answers, and answer values try: question_re = re.compile(r"(\d{1,2}[A-Z]?)\. (.+)") question_tags = soup.find_all("td", class_="dddefault", string=question_re) for q_tag in question_tags: # Pull out question number and text question_num, question_str = question_re.search(q_tag.string).groups() question = QuestionInstance(Num=question_num, FullString=question_str) # Pull out answer text (if available) and responses row_cells = q_tag.parent.find_all("td", class_="dddefault") if len(row_cells)==8: # All data on one row, 1-5 response question.answers = [AnswerField( Weight=i+1, Respondents=row_cells[i+1].p.contents[1]) for i in range(5)] elif len(row_cells)==1: # Answers are one per row on next 5 rows for row in q_tag.parent.next_siblings: if row.name=="tr": #Break when we get to the next question header if row.find("th", class_="ddtitle"): break cells = row.find_all("td", class_="dddefault") answer = AnswerField(AnswerText=cells[0].string, Respondents=cells[1].string) question.answers.append(answer) else: raise ParseError("Unexpected number of cells in row") ev.questions.append(question) except (AttributeError,IndexError) as e: raise ParseError("Error parsing evaluation results",e), None, sys.exc_info()[2] return ev # Only call options: # process_evals(db, dl, year, adln) # process_evals(db, dl, year, adln, crns) # process_evals(db, dl, urls) def process_evals(db, dl, year=None, adln=False, crns=[], urls=[]): use_urls = bool(urls) if not urls: urls = dl.get_urls_by_year(year, adln, crns) for idx, url in enumerate(urls): crn = re.compile('IN_CRN=(\d*)').search(url).group(1) if use_urls: #extract year, adln from url if needed year = re.compile('IN_ACYR=(\d*)').search(url).group(1) adln_char = re.compile('IN_ADLN_OIX=(\d*)').search(url).group(1) if adln_char not in 'OX': raise ParseError("ADLN in URL must be O or X") adln = adln_char=='O' print("\rDownloading and parsing class {}/{} (CRN {})".format(idx+1,len(urls),crn), end='') try: eval_text = dl.download_eval(url) ev = parse_eval(eval_text) ev.ADLN = adln db.store(ev) except ParseError as e: #_, _, tb = sys.exc_info() print("\nError parsing CRN {}:".format(crn)) traceback.print_exc() print() db.store(FailedScrape(CRN=crn, AcademicYear=year, ADLN=adln)) #print(crn, year, adln, file=failed_crns) except DatabaseError as e: print("\nError storing CRN {}: {}\n".format(crn, repr(e))) db.store(FailedScrape(CRN=crn, AcademicYear=year, ADLN=adln)) #print(crn, file=failed_crns) except requests.exceptions.Timeout: print("\nTimed out while loading CRN {}".format(crn)) db.store(FailedScrape(CRN=crn, AcademicYear=year, ADLN=adln)) #!!Add reason to failed table print("\nDone") def process_failed_scrapes(db, dl): grouped_crns = db.get_col_grouped_by(FailedScrape,'CRN',['AcademicYear','ADLN']) db.clear_table(FailedScrape) for group in grouped_crns: process_evals(db, dl, group['AcademicYear'], group['ADLN'], group['CRN']) #!! # check all courses for uniqueness of crns def validate_assumptions(db, dl): pass # Check that all available courses are accounted for in db or failures # Does not identify duplicate CRNS #!! # untested #!! def validate_and_fix(db, dl, fix=True): #years = dl.get_years() #!!! years = range(2006,2017) years = [2010,2013,2016] #!!! remove missing_urls = [] for year in years: missing_in_year = [] urls = dl.get_urls_by_year(year, adln=True) urls += dl.get_urls_by_year(year, adln=False) for url in urls: #!! get this from func crn = re.compile('IN_CRN=(\d*)').search(url).group(1) #!! adln_char = re.compile('IN_ADLN_OIX=(\d*)').search(url).group(1) if adln_char not in 'OX': raise ParseError("ADLN in URL must be O or X") adln = adln_char=='O' db.cur.execute("""SELECT 1 FROM ( SELECT CRN FROM Classes WHERE AcademicYear = {year} AND ADLN={adln} UNION ALL SELECT CRN FROM FailedScrapes WHERE AcademicYear = {year} AND ADLN={adln} ) a WHERE CRN = {crn}""".format(year=year,adln=adln,crn=crn)) if not db.cur.fetchone(): # Entry not found missing_in_year.append(url) print("{} missing entries in {}-{} out of {}".format(len(missing_in_year),year-1,year,len(urls))) missing_urls += missing_in_year if missing_urls: print("{} total missing entries".format(len(missing_urls))) if fix: process_evals(db, dl, urls=missing_urls) else: print("No missing entries found") #!! notify about failed scrapes def main(argv): try: opts, args = getopt.getopt(argv, "ry:fvb") except getopt.GetoptError: print("parser.py [-r] [-y YEAR | -f] [-v]") #!! return reset = False rerun_fails = False year = None validate = False build_sums = False for opt, arg in opts: if opt =='-r': reset = True elif opt =='-y': year = arg elif opt =='-f': rerun_fails = True elif opt == '-v': validate = True elif opt == '-b': build_sums = True db = Database() dl = Downloader(user=raw_input("Username: "), passwd=getpass.getpass()) dl.login() if reset: db.reset() if rerun_fails: print("Re-running failed scrapes") process_failed_scrapes(db, dl) elif year: process_evals(db, dl, 2015, adln=True) process_evals(db, dl, 2015, adln=False) elif validate: validate_and_fix(db, dl) elif build_sums: pass else: for year in range(2016,2017): process_evals(db, dl, year, adln=True) process_evals(db, dl, year, adln=False) #!! process all years db.build_class_summaries() # with open("sampleeval.html") as f: # ev = parse_eval(f.read()) # db.store(ev) # with open("sampleeval2.html") as f: # ev = parse_eval(f.read()) # db.store(ev) # print("Complete") if __name__ == "__main__": main(sys.argv[1:])
false
true
1c4836e8d6c54e7d96ea2a3dadcfa5b15943f85b
238
py
Python
ex010.py
Vhassan/Python-Cheatsheet
526f5fcfc8e93d0aca139ca6d8d4f20851ab16f5
[ "MIT" ]
null
null
null
ex010.py
Vhassan/Python-Cheatsheet
526f5fcfc8e93d0aca139ca6d8d4f20851ab16f5
[ "MIT" ]
null
null
null
ex010.py
Vhassan/Python-Cheatsheet
526f5fcfc8e93d0aca139ca6d8d4f20851ab16f5
[ "MIT" ]
null
null
null
#Crie um programa que leia quanto dinheiro uma pessoa tem na carteira e mostre quantos dólares ela pode comprar. real = float(input('quanto vc tem na carteira: R$')) print('A conversão para moeda dolar é : US${:.2f}'.format((real/3.27)))
59.5
112
0.735294
real = float(input('quanto vc tem na carteira: R$')) print('A conversão para moeda dolar é : US${:.2f}'.format((real/3.27)))
true
true
1c4837316def2f2de591b95262f04bd0a307b76b
3,392
py
Python
ExtendedAIModule/rhombus_services/arg_parser.py
Bricktheworld/rhombus-api-examples-python
b4778c3a635786070ee10a3131b1a1f7f6ebac36
[ "MIT" ]
null
null
null
ExtendedAIModule/rhombus_services/arg_parser.py
Bricktheworld/rhombus-api-examples-python
b4778c3a635786070ee10a3131b1a1f7f6ebac36
[ "MIT" ]
20
2021-06-08T22:29:20.000Z
2022-01-15T19:51:46.000Z
ExtendedAIModule/rhombus_services/arg_parser.py
Bricktheworld/rhombus-api-examples-python
b4778c3a635786070ee10a3131b1a1f7f6ebac36
[ "MIT" ]
9
2021-06-08T22:15:35.000Z
2022-03-03T05:19:58.000Z
################################################################################### # Copyright (c) 2021 Rhombus Systems # # # # Permission is hereby granted, free of charge, to any person obtaining a copy # # of this software and associated documentation files (the "Software"), to deal # # in the Software without restriction, including without limitation the rights # # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # # copies of the Software, and to permit persons to whom the Software is # # furnished to do so, subject to the following conditions: # # # # The above copyright notice and this permission notice shall be included in all # # copies or substantial portions of the Software. # # # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # # SOFTWARE. # ################################################################################### # Import type hints from typing import List # Import argparse to parse our arguments for us easily import argparse def parse_arguments(argv: List[str]) -> argparse.Namespace: """Parse the command line args. :param argv: The Commandline arguments from the user, which can be retrieved via sys.argv[1:] """ # Create our parser parser = argparse.ArgumentParser(description='Pulls footage from a camera on LAN and stores it to the filesystem.') # The --api_key or -a param will hold our API key parser.add_argument('--api_key', '-a', type=str, required=True, help='Rhombus API key') # The --camera_uuid or -c param will hold the UUID of the camera which will be processed parser.add_argument('--camera_uuid', '-c', type=str, required=True, help='Device Id to pull footage from') # The --interval or -i param will hold how often to poll the camera for new footage in seconds, by default 10 seconds parser.add_argument('--interval', '-i', type=int, required=False, help='How often to poll the camera for new footage in seconds, by default 10 seconds', default=10) # The --connection_type or -t param will hold the ConnectionType to the camera. It is not recommended to run in WAN mode unless this python server is running on a separate network from the camera parser.add_argument('--connection_type', '-t', type=str, required=False, help='The connection type to the camera, either LAN or WAN (default LAN)', default="LAN") # Return all of our arguments return parser.parse_args(argv)
60.571429
199
0.583137
true
true
1c48374373ae16db6dbcfd16316661e717dab9fc
5,230
py
Python
tests/input/pdf/test_pdf.py
asweeney86/preview-generator
354cbac1c131ebbb81cd9cfd9b4bc0c184d10103
[ "MIT" ]
null
null
null
tests/input/pdf/test_pdf.py
asweeney86/preview-generator
354cbac1c131ebbb81cd9cfd9b4bc0c184d10103
[ "MIT" ]
null
null
null
tests/input/pdf/test_pdf.py
asweeney86/preview-generator
354cbac1c131ebbb81cd9cfd9b4bc0c184d10103
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import os import re import shutil import typing from PIL import Image from PyPDF2 import PdfFileReader import PyPDF2.utils import pytest from preview_generator.exception import UnavailablePreviewType from preview_generator.manager import PreviewManager from tests import test_utils CURRENT_DIR = os.path.dirname(os.path.abspath(__file__)) CACHE_DIR = "/tmp/preview-generator-tests/cache" PDF_FILE_PATH = os.path.join(CURRENT_DIR, "the_pdf.pdf") PDF_FILE_PATH__ENCRYPTED = os.path.join(CURRENT_DIR, "the_pdf.encrypted.pdf") PDF_FILE_PATH__A4 = os.path.join(CURRENT_DIR, "qpdfconvert.pdf") def setup_function(function: typing.Callable) -> None: shutil.rmtree(CACHE_DIR, ignore_errors=True) def test_to_jpeg() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH) is True path_to_file = manager.get_jpeg_preview( file_path=PDF_FILE_PATH, height=512, width=321, force=True ) assert os.path.exists(path_to_file) is True assert os.path.getsize(path_to_file) > 0 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN__JPEG, path_to_file) with Image.open(path_to_file) as jpeg: assert jpeg.height in range(453, 455) assert jpeg.width == 321 def test_to_jpeg__encrypted_pdf() -> None: with pytest.raises(PyPDF2.utils.PdfReadError): #  ensure file is encrpyted pdf = PdfFileReader(PDF_FILE_PATH__ENCRYPTED) pdf.getPage(0) manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH) is True path_to_file = manager.get_jpeg_preview( file_path=PDF_FILE_PATH__ENCRYPTED, height=512, width=321, force=True ) assert os.path.exists(path_to_file) is True assert os.path.getsize(path_to_file) > 0 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN__JPEG, path_to_file) with Image.open(path_to_file) as jpeg: assert jpeg.height in range(453, 455) assert jpeg.width == 321 def test_to_jpeg_no_size() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH) is True path_to_file = manager.get_jpeg_preview(file_path=PDF_FILE_PATH, force=True) assert os.path.exists(path_to_file) is True assert os.path.getsize(path_to_file) > 0 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN__JPEG, path_to_file) with Image.open(path_to_file) as jpeg: assert jpeg.height == 256 assert jpeg.width in range(180, 182) def test_to_text() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_text_preview(file_path=PDF_FILE_PATH) is False with pytest.raises(UnavailablePreviewType): manager.get_text_preview(file_path=PDF_FILE_PATH, force=True) def test_to_json() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_json_preview(file_path=PDF_FILE_PATH) is True manager.get_json_preview(file_path=PDF_FILE_PATH, force=True) # TODO - G.M - 2018-11-06 - To be completed def test_to_pdf() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_pdf_preview(file_path=PDF_FILE_PATH) is True manager.get_pdf_preview(file_path=PDF_FILE_PATH, force=True) # TODO - G.M - 2018-11-06 - To be completed def test_to_pdf_one_page() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_pdf_preview(file_path=PDF_FILE_PATH) is True path_0 = manager.get_pdf_preview(file_path=PDF_FILE_PATH, page=0, force=True) assert os.path.exists(path_0) is True assert os.path.getsize(path_0) > 1000 # verify if the size of the pdf refer to a normal content assert re.match(test_utils.CACHE_FILE_PATH_PATTERN_WITH_PAGE__PDF, path_0) pdf = PdfFileReader(open(path_0, "rb")) assert pdf.getNumPages() == 1 path_1 = manager.get_pdf_preview(file_path=PDF_FILE_PATH, page=1, force=True) assert os.path.exists(path_1) is True assert os.path.getsize(path_1) > 1000 # verify if the size of the pdf refer to a normal content assert re.match(test_utils.CACHE_FILE_PATH_PATTERN_WITH_PAGE__PDF, path_1) pdf = PdfFileReader(open(path_1, "rb")) assert pdf.getNumPages() == 1 def test_algorithm4() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH__A4) is True path_to_file = manager.get_jpeg_preview(file_path=PDF_FILE_PATH__A4, force=True) with Image.open(path_to_file) as jpeg: assert jpeg.height == 256 assert jpeg.width in range(180, 182) def test_get_nb_page() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) nb_page = manager.get_page_nb(file_path=PDF_FILE_PATH) assert nb_page == 2 nb_page = manager.get_page_nb(file_path=PDF_FILE_PATH__ENCRYPTED) assert nb_page == 2 nb_page = manager.get_page_nb(file_path=PDF_FILE_PATH__A4) assert nb_page == 2
39.621212
100
0.759656
import os import re import shutil import typing from PIL import Image from PyPDF2 import PdfFileReader import PyPDF2.utils import pytest from preview_generator.exception import UnavailablePreviewType from preview_generator.manager import PreviewManager from tests import test_utils CURRENT_DIR = os.path.dirname(os.path.abspath(__file__)) CACHE_DIR = "/tmp/preview-generator-tests/cache" PDF_FILE_PATH = os.path.join(CURRENT_DIR, "the_pdf.pdf") PDF_FILE_PATH__ENCRYPTED = os.path.join(CURRENT_DIR, "the_pdf.encrypted.pdf") PDF_FILE_PATH__A4 = os.path.join(CURRENT_DIR, "qpdfconvert.pdf") def setup_function(function: typing.Callable) -> None: shutil.rmtree(CACHE_DIR, ignore_errors=True) def test_to_jpeg() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH) is True path_to_file = manager.get_jpeg_preview( file_path=PDF_FILE_PATH, height=512, width=321, force=True ) assert os.path.exists(path_to_file) is True assert os.path.getsize(path_to_file) > 0 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN__JPEG, path_to_file) with Image.open(path_to_file) as jpeg: assert jpeg.height in range(453, 455) assert jpeg.width == 321 def test_to_jpeg__encrypted_pdf() -> None: with pytest.raises(PyPDF2.utils.PdfReadError): pdf = PdfFileReader(PDF_FILE_PATH__ENCRYPTED) pdf.getPage(0) manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH) is True path_to_file = manager.get_jpeg_preview( file_path=PDF_FILE_PATH__ENCRYPTED, height=512, width=321, force=True ) assert os.path.exists(path_to_file) is True assert os.path.getsize(path_to_file) > 0 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN__JPEG, path_to_file) with Image.open(path_to_file) as jpeg: assert jpeg.height in range(453, 455) assert jpeg.width == 321 def test_to_jpeg_no_size() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH) is True path_to_file = manager.get_jpeg_preview(file_path=PDF_FILE_PATH, force=True) assert os.path.exists(path_to_file) is True assert os.path.getsize(path_to_file) > 0 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN__JPEG, path_to_file) with Image.open(path_to_file) as jpeg: assert jpeg.height == 256 assert jpeg.width in range(180, 182) def test_to_text() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_text_preview(file_path=PDF_FILE_PATH) is False with pytest.raises(UnavailablePreviewType): manager.get_text_preview(file_path=PDF_FILE_PATH, force=True) def test_to_json() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_json_preview(file_path=PDF_FILE_PATH) is True manager.get_json_preview(file_path=PDF_FILE_PATH, force=True) def test_to_pdf() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_pdf_preview(file_path=PDF_FILE_PATH) is True manager.get_pdf_preview(file_path=PDF_FILE_PATH, force=True) def test_to_pdf_one_page() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_pdf_preview(file_path=PDF_FILE_PATH) is True path_0 = manager.get_pdf_preview(file_path=PDF_FILE_PATH, page=0, force=True) assert os.path.exists(path_0) is True assert os.path.getsize(path_0) > 1000 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN_WITH_PAGE__PDF, path_0) pdf = PdfFileReader(open(path_0, "rb")) assert pdf.getNumPages() == 1 path_1 = manager.get_pdf_preview(file_path=PDF_FILE_PATH, page=1, force=True) assert os.path.exists(path_1) is True assert os.path.getsize(path_1) > 1000 assert re.match(test_utils.CACHE_FILE_PATH_PATTERN_WITH_PAGE__PDF, path_1) pdf = PdfFileReader(open(path_1, "rb")) assert pdf.getNumPages() == 1 def test_algorithm4() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) assert manager.has_jpeg_preview(file_path=PDF_FILE_PATH__A4) is True path_to_file = manager.get_jpeg_preview(file_path=PDF_FILE_PATH__A4, force=True) with Image.open(path_to_file) as jpeg: assert jpeg.height == 256 assert jpeg.width in range(180, 182) def test_get_nb_page() -> None: manager = PreviewManager(cache_folder_path=CACHE_DIR, create_folder=True) nb_page = manager.get_page_nb(file_path=PDF_FILE_PATH) assert nb_page == 2 nb_page = manager.get_page_nb(file_path=PDF_FILE_PATH__ENCRYPTED) assert nb_page == 2 nb_page = manager.get_page_nb(file_path=PDF_FILE_PATH__A4) assert nb_page == 2
true
true
1c48380c9cbf94328974481e6bfb12901edaac59
4,307
py
Python
tests/app/service/test_sender.py
cds-snc/notification-api
b1c1064f291eb860b494c3fa65ac256ad70bf47c
[ "MIT" ]
41
2019-11-28T16:58:41.000Z
2022-01-28T21:11:16.000Z
tests/app/service/test_sender.py
cds-snc/notification-api
b1c1064f291eb860b494c3fa65ac256ad70bf47c
[ "MIT" ]
1,083
2019-07-08T12:57:24.000Z
2022-03-08T18:53:40.000Z
tests/app/service/test_sender.py
cds-snc/notification-api
b1c1064f291eb860b494c3fa65ac256ad70bf47c
[ "MIT" ]
9
2020-01-24T19:56:43.000Z
2022-01-27T21:36:53.000Z
import pytest from flask import current_app from app.dao.services_dao import dao_add_user_to_service from app.models import EMAIL_TYPE, SMS_TYPE, Notification from app.service.sender import send_notification_to_service_users from tests.app.conftest import notify_service as create_notify_service from tests.app.conftest import sample_service as create_sample_service from tests.app.db import create_template, create_user @pytest.mark.parametrize("notification_type", [EMAIL_TYPE, SMS_TYPE]) def test_send_notification_to_service_users_persists_notifications_correctly( notify_db, notify_db_session, notification_type, sample_user, mocker ): mocker.patch("app.service.sender.send_notification_to_queue") notify_service, user = create_notify_service(notify_db, notify_db_session) service = create_sample_service(notify_db, notify_db_session, user=sample_user) template = create_template(service, template_type=notification_type) send_notification_to_service_users(service_id=service.id, template_id=template.id) to = sample_user.email_address if notification_type == EMAIL_TYPE else sample_user.mobile_number notification = Notification.query.one() assert Notification.query.count() == 1 assert notification.to == to assert str(notification.service_id) == current_app.config["NOTIFY_SERVICE_ID"] assert notification.template.id == template.id assert notification.template.template_type == notification_type assert notification.notification_type == notification_type assert notification.reply_to_text == notify_service.get_default_reply_to_email_address() def test_send_notification_to_service_users_sends_to_queue(notify_db, notify_db_session, sample_user, mocker): send_mock = mocker.patch("app.service.sender.send_notification_to_queue") create_notify_service(notify_db, notify_db_session) service = create_sample_service(notify_db, notify_db_session, user=sample_user) template = create_template(service, template_type=EMAIL_TYPE) send_notification_to_service_users(service_id=service.id, template_id=template.id) assert send_mock.called assert send_mock.call_count == 1 def test_send_notification_to_service_users_includes_user_fields_in_personalisation( notify_db, notify_db_session, sample_user, mocker ): persist_mock = mocker.patch("app.service.sender.persist_notification") mocker.patch("app.service.sender.send_notification_to_queue") create_notify_service(notify_db, notify_db_session) service = create_sample_service(notify_db, notify_db_session, user=sample_user) template = create_template(service, template_type=EMAIL_TYPE) send_notification_to_service_users( service_id=service.id, template_id=template.id, include_user_fields=["name", "email_address", "state"], ) persist_call = persist_mock.call_args_list[0][1] assert len(persist_mock.call_args_list) == 1 assert persist_call["personalisation"] == { "name": sample_user.name, "email_address": sample_user.email_address, "state": sample_user.state, } def test_send_notification_to_service_users_sends_to_active_users_only(notify_db, notify_db_session, mocker): mocker.patch("app.service.sender.send_notification_to_queue") create_notify_service(notify_db, notify_db_session) first_active_user = create_user(email="foo@bar.com", state="active") second_active_user = create_user(email="foo1@bar.com", state="active") pending_user = create_user(email="foo2@bar.com", state="pending") service = create_sample_service(notify_db, notify_db_session, user=first_active_user) dao_add_user_to_service(service, second_active_user) dao_add_user_to_service(service, pending_user) template = create_template(service, template_type=EMAIL_TYPE) send_notification_to_service_users(service_id=service.id, template_id=template.id) notifications = Notification.query.all() notifications_recipients = [notification.to for notification in notifications] assert Notification.query.count() == 2 assert pending_user.email_address not in notifications_recipients assert first_active_user.email_address in notifications_recipients assert second_active_user.email_address in notifications_recipients
46.311828
110
0.80404
import pytest from flask import current_app from app.dao.services_dao import dao_add_user_to_service from app.models import EMAIL_TYPE, SMS_TYPE, Notification from app.service.sender import send_notification_to_service_users from tests.app.conftest import notify_service as create_notify_service from tests.app.conftest import sample_service as create_sample_service from tests.app.db import create_template, create_user @pytest.mark.parametrize("notification_type", [EMAIL_TYPE, SMS_TYPE]) def test_send_notification_to_service_users_persists_notifications_correctly( notify_db, notify_db_session, notification_type, sample_user, mocker ): mocker.patch("app.service.sender.send_notification_to_queue") notify_service, user = create_notify_service(notify_db, notify_db_session) service = create_sample_service(notify_db, notify_db_session, user=sample_user) template = create_template(service, template_type=notification_type) send_notification_to_service_users(service_id=service.id, template_id=template.id) to = sample_user.email_address if notification_type == EMAIL_TYPE else sample_user.mobile_number notification = Notification.query.one() assert Notification.query.count() == 1 assert notification.to == to assert str(notification.service_id) == current_app.config["NOTIFY_SERVICE_ID"] assert notification.template.id == template.id assert notification.template.template_type == notification_type assert notification.notification_type == notification_type assert notification.reply_to_text == notify_service.get_default_reply_to_email_address() def test_send_notification_to_service_users_sends_to_queue(notify_db, notify_db_session, sample_user, mocker): send_mock = mocker.patch("app.service.sender.send_notification_to_queue") create_notify_service(notify_db, notify_db_session) service = create_sample_service(notify_db, notify_db_session, user=sample_user) template = create_template(service, template_type=EMAIL_TYPE) send_notification_to_service_users(service_id=service.id, template_id=template.id) assert send_mock.called assert send_mock.call_count == 1 def test_send_notification_to_service_users_includes_user_fields_in_personalisation( notify_db, notify_db_session, sample_user, mocker ): persist_mock = mocker.patch("app.service.sender.persist_notification") mocker.patch("app.service.sender.send_notification_to_queue") create_notify_service(notify_db, notify_db_session) service = create_sample_service(notify_db, notify_db_session, user=sample_user) template = create_template(service, template_type=EMAIL_TYPE) send_notification_to_service_users( service_id=service.id, template_id=template.id, include_user_fields=["name", "email_address", "state"], ) persist_call = persist_mock.call_args_list[0][1] assert len(persist_mock.call_args_list) == 1 assert persist_call["personalisation"] == { "name": sample_user.name, "email_address": sample_user.email_address, "state": sample_user.state, } def test_send_notification_to_service_users_sends_to_active_users_only(notify_db, notify_db_session, mocker): mocker.patch("app.service.sender.send_notification_to_queue") create_notify_service(notify_db, notify_db_session) first_active_user = create_user(email="foo@bar.com", state="active") second_active_user = create_user(email="foo1@bar.com", state="active") pending_user = create_user(email="foo2@bar.com", state="pending") service = create_sample_service(notify_db, notify_db_session, user=first_active_user) dao_add_user_to_service(service, second_active_user) dao_add_user_to_service(service, pending_user) template = create_template(service, template_type=EMAIL_TYPE) send_notification_to_service_users(service_id=service.id, template_id=template.id) notifications = Notification.query.all() notifications_recipients = [notification.to for notification in notifications] assert Notification.query.count() == 2 assert pending_user.email_address not in notifications_recipients assert first_active_user.email_address in notifications_recipients assert second_active_user.email_address in notifications_recipients
true
true
1c48388e6a7603d89677a9f9449e120b5b428b22
15,266
py
Python
Original_Codes/GDL_code-master/models/WGANGP.py
TeaKatz/Generative_Deep_Learning
1f499e482f78b3d1146b24213e5d558226b8fc6e
[ "MIT" ]
2
2021-07-09T16:45:51.000Z
2021-10-30T18:00:06.000Z
models/WGANGP.py
steveive8/Study-Generative-Deep-Learning
f62b9150a5e18240dd22816918f2ce6abf807d58
[ "MIT" ]
10
2020-09-26T01:22:18.000Z
2022-03-12T00:42:42.000Z
models/WGANGP.py
steveive8/Study-Generative-Deep-Learning
f62b9150a5e18240dd22816918f2ce6abf807d58
[ "MIT" ]
null
null
null
from keras.layers import Input, Conv2D, Flatten, Dense, Conv2DTranspose, Reshape, Lambda, Activation, BatchNormalization, LeakyReLU, Dropout, ZeroPadding2D, UpSampling2D from keras.layers.merge import _Merge from keras.models import Model, Sequential from keras import backend as K from keras.optimizers import Adam, RMSprop from keras.callbacks import ModelCheckpoint from keras.utils import plot_model from keras.initializers import RandomNormal from functools import partial import numpy as np import json import os import pickle import matplotlib.pyplot as plt class RandomWeightedAverage(_Merge): def __init__(self, batch_size): super().__init__() self.batch_size = batch_size """Provides a (random) weighted average between real and generated image samples""" def _merge_function(self, inputs): alpha = K.random_uniform((self.batch_size, 1, 1, 1)) return (alpha * inputs[0]) + ((1 - alpha) * inputs[1]) class WGANGP(): def __init__(self , input_dim , critic_conv_filters , critic_conv_kernel_size , critic_conv_strides , critic_batch_norm_momentum , critic_activation , critic_dropout_rate , critic_learning_rate , generator_initial_dense_layer_size , generator_upsample , generator_conv_filters , generator_conv_kernel_size , generator_conv_strides , generator_batch_norm_momentum , generator_activation , generator_dropout_rate , generator_learning_rate , optimiser , grad_weight , z_dim , batch_size ): self.name = 'gan' self.input_dim = input_dim self.critic_conv_filters = critic_conv_filters self.critic_conv_kernel_size = critic_conv_kernel_size self.critic_conv_strides = critic_conv_strides self.critic_batch_norm_momentum = critic_batch_norm_momentum self.critic_activation = critic_activation self.critic_dropout_rate = critic_dropout_rate self.critic_learning_rate = critic_learning_rate self.generator_initial_dense_layer_size = generator_initial_dense_layer_size self.generator_upsample = generator_upsample self.generator_conv_filters = generator_conv_filters self.generator_conv_kernel_size = generator_conv_kernel_size self.generator_conv_strides = generator_conv_strides self.generator_batch_norm_momentum = generator_batch_norm_momentum self.generator_activation = generator_activation self.generator_dropout_rate = generator_dropout_rate self.generator_learning_rate = generator_learning_rate self.optimiser = optimiser self.z_dim = z_dim self.n_layers_critic = len(critic_conv_filters) self.n_layers_generator = len(generator_conv_filters) self.weight_init = RandomNormal(mean=0., stddev=0.02) # 'he_normal' #RandomNormal(mean=0., stddev=0.02) self.grad_weight = grad_weight self.batch_size = batch_size self.d_losses = [] self.g_losses = [] self.epoch = 0 self._build_critic() self._build_generator() self._build_adversarial() def gradient_penalty_loss(self, y_true, y_pred, interpolated_samples): """ Computes gradient penalty based on prediction and weighted real / fake samples """ gradients = K.gradients(y_pred, interpolated_samples)[0] # compute the euclidean norm by squaring ... gradients_sqr = K.square(gradients) # ... summing over the rows ... gradients_sqr_sum = K.sum(gradients_sqr, axis=np.arange(1, len(gradients_sqr.shape))) # ... and sqrt gradient_l2_norm = K.sqrt(gradients_sqr_sum) # compute lambda * (1 - ||grad||)^2 still for each single sample gradient_penalty = K.square(1 - gradient_l2_norm) # return the mean as loss over all the batch samples return K.mean(gradient_penalty) def wasserstein(self, y_true, y_pred): return -K.mean(y_true * y_pred) def get_activation(self, activation): if activation == 'leaky_relu': layer = LeakyReLU(alpha = 0.2) else: layer = Activation(activation) return layer def _build_critic(self): ### THE critic critic_input = Input(shape=self.input_dim, name='critic_input') x = critic_input for i in range(self.n_layers_critic): x = Conv2D( filters = self.critic_conv_filters[i] , kernel_size = self.critic_conv_kernel_size[i] , strides = self.critic_conv_strides[i] , padding = 'same' , name = 'critic_conv_' + str(i) , kernel_initializer = self.weight_init )(x) if self.critic_batch_norm_momentum and i > 0: x = BatchNormalization(momentum = self.critic_batch_norm_momentum)(x) x = self.get_activation(self.critic_activation)(x) if self.critic_dropout_rate: x = Dropout(rate = self.critic_dropout_rate)(x) x = Flatten()(x) # x = Dense(512, kernel_initializer = self.weight_init)(x) # x = self.get_activation(self.critic_activation)(x) critic_output = Dense(1, activation=None , kernel_initializer = self.weight_init )(x) self.critic = Model(critic_input, critic_output) def _build_generator(self): ### THE generator generator_input = Input(shape=(self.z_dim,), name='generator_input') x = generator_input x = Dense(np.prod(self.generator_initial_dense_layer_size), kernel_initializer = self.weight_init)(x) if self.generator_batch_norm_momentum: x = BatchNormalization(momentum = self.generator_batch_norm_momentum)(x) x = self.get_activation(self.generator_activation)(x) x = Reshape(self.generator_initial_dense_layer_size)(x) if self.generator_dropout_rate: x = Dropout(rate = self.generator_dropout_rate)(x) for i in range(self.n_layers_generator): if self.generator_upsample[i] == 2: x = UpSampling2D()(x) x = Conv2D( filters = self.generator_conv_filters[i] , kernel_size = self.generator_conv_kernel_size[i] , padding = 'same' , name = 'generator_conv_' + str(i) , kernel_initializer = self.weight_init )(x) else: x = Conv2DTranspose( filters = self.generator_conv_filters[i] , kernel_size = self.generator_conv_kernel_size[i] , padding = 'same' , strides = self.generator_conv_strides[i] , name = 'generator_conv_' + str(i) , kernel_initializer = self.weight_init )(x) if i < self.n_layers_generator - 1: if self.generator_batch_norm_momentum: x = BatchNormalization(momentum = self.generator_batch_norm_momentum)(x) x = self.get_activation(self.generator_activation)(x) else: x = Activation('tanh')(x) generator_output = x self.generator = Model(generator_input, generator_output) def get_opti(self, lr): if self.optimiser == 'adam': opti = Adam(lr=lr, beta_1=0.5) elif self.optimiser == 'rmsprop': opti = RMSprop(lr=lr) else: opti = Adam(lr=lr) return opti def set_trainable(self, m, val): m.trainable = val for l in m.layers: l.trainable = val def _build_adversarial(self): #------------------------------- # Construct Computational Graph # for the Critic #------------------------------- # Freeze generator's layers while training critic self.set_trainable(self.generator, False) # Image input (real sample) real_img = Input(shape=self.input_dim) # Fake image z_disc = Input(shape=(self.z_dim,)) fake_img = self.generator(z_disc) # critic determines validity of the real and fake images fake = self.critic(fake_img) valid = self.critic(real_img) # Construct weighted average between real and fake images interpolated_img = RandomWeightedAverage(self.batch_size)([real_img, fake_img]) # Determine validity of weighted sample validity_interpolated = self.critic(interpolated_img) # Use Python partial to provide loss function with additional # 'interpolated_samples' argument partial_gp_loss = partial(self.gradient_penalty_loss, interpolated_samples=interpolated_img) partial_gp_loss.__name__ = 'gradient_penalty' # Keras requires function names self.critic_model = Model(inputs=[real_img, z_disc], outputs=[valid, fake, validity_interpolated]) self.critic_model.compile( loss=[self.wasserstein,self.wasserstein, partial_gp_loss] ,optimizer=self.get_opti(self.critic_learning_rate) ,loss_weights=[1, 1, self.grad_weight] ) #------------------------------- # Construct Computational Graph # for Generator #------------------------------- # For the generator we freeze the critic's layers self.set_trainable(self.critic, False) self.set_trainable(self.generator, True) # Sampled noise for input to generator model_input = Input(shape=(self.z_dim,)) # Generate images based of noise img = self.generator(model_input) # Discriminator determines validity model_output = self.critic(img) # Defines generator model self.model = Model(model_input, model_output) self.model.compile(optimizer=self.get_opti(self.generator_learning_rate) , loss=self.wasserstein ) self.set_trainable(self.critic, True) def train_critic(self, x_train, batch_size, using_generator): valid = np.ones((batch_size,1), dtype=np.float32) fake = -np.ones((batch_size,1), dtype=np.float32) dummy = np.zeros((batch_size, 1), dtype=np.float32) # Dummy gt for gradient penalty if using_generator: true_imgs = next(x_train)[0] if true_imgs.shape[0] != batch_size: true_imgs = next(x_train)[0] else: idx = np.random.randint(0, x_train.shape[0], batch_size) true_imgs = x_train[idx] noise = np.random.normal(0, 1, (batch_size, self.z_dim)) d_loss = self.critic_model.train_on_batch([true_imgs, noise], [valid, fake, dummy]) return d_loss def train_generator(self, batch_size): valid = np.ones((batch_size,1), dtype=np.float32) noise = np.random.normal(0, 1, (batch_size, self.z_dim)) return self.model.train_on_batch(noise, valid) def train(self, x_train, batch_size, epochs, run_folder, print_every_n_batches = 10 , n_critic = 5 , using_generator = False): for epoch in range(self.epoch, self.epoch + epochs): if epoch % 100 == 0: critic_loops = 5 else: critic_loops = n_critic for _ in range(critic_loops): d_loss = self.train_critic(x_train, batch_size, using_generator) g_loss = self.train_generator(batch_size) print ("%d (%d, %d) [D loss: (%.1f)(R %.1f, F %.1f, G %.1f)] [G loss: %.1f]" % (epoch, critic_loops, 1, d_loss[0], d_loss[1],d_loss[2],d_loss[3],g_loss)) self.d_losses.append(d_loss) self.g_losses.append(g_loss) # If at save interval => save generated image samples if epoch % print_every_n_batches == 0: self.sample_images(run_folder) self.model.save_weights(os.path.join(run_folder, 'weights/weights-%d.h5' % (epoch))) self.model.save_weights(os.path.join(run_folder, 'weights/weights.h5')) self.save_model(run_folder) self.epoch+=1 def sample_images(self, run_folder): r, c = 5, 5 noise = np.random.normal(0, 1, (r * c, self.z_dim)) gen_imgs = self.generator.predict(noise) #Rescale images 0 - 1 gen_imgs = 0.5 * (gen_imgs + 1) gen_imgs = np.clip(gen_imgs, 0, 1) fig, axs = plt.subplots(r, c, figsize=(15,15)) cnt = 0 for i in range(r): for j in range(c): axs[i,j].imshow(np.squeeze(gen_imgs[cnt, :,:,:]), cmap = 'gray_r') axs[i,j].axis('off') cnt += 1 fig.savefig(os.path.join(run_folder, "images/sample_%d.png" % self.epoch)) plt.close() def plot_model(self, run_folder): plot_model(self.model, to_file=os.path.join(run_folder ,'viz/model.png'), show_shapes = True, show_layer_names = True) plot_model(self.critic, to_file=os.path.join(run_folder ,'viz/critic.png'), show_shapes = True, show_layer_names = True) plot_model(self.generator, to_file=os.path.join(run_folder ,'viz/generator.png'), show_shapes = True, show_layer_names = True) def save(self, folder): with open(os.path.join(folder, 'params.pkl'), 'wb') as f: pickle.dump([ self.input_dim , self.critic_conv_filters , self.critic_conv_kernel_size , self.critic_conv_strides , self.critic_batch_norm_momentum , self.critic_activation , self.critic_dropout_rate , self.critic_learning_rate , self.generator_initial_dense_layer_size , self.generator_upsample , self.generator_conv_filters , self.generator_conv_kernel_size , self.generator_conv_strides , self.generator_batch_norm_momentum , self.generator_activation , self.generator_dropout_rate , self.generator_learning_rate , self.optimiser , self.grad_weight , self.z_dim , self.batch_size ], f) self.plot_model(folder) def save_model(self, run_folder): self.model.save(os.path.join(run_folder, 'model.h5')) self.critic.save(os.path.join(run_folder, 'critic.h5')) self.generator.save(os.path.join(run_folder, 'generator.h5')) pickle.dump(self, open( os.path.join(run_folder, "obj.pkl"), "wb" )) def load_weights(self, filepath): self.model.load_weights(filepath)
35.502326
169
0.60435
from keras.layers import Input, Conv2D, Flatten, Dense, Conv2DTranspose, Reshape, Lambda, Activation, BatchNormalization, LeakyReLU, Dropout, ZeroPadding2D, UpSampling2D from keras.layers.merge import _Merge from keras.models import Model, Sequential from keras import backend as K from keras.optimizers import Adam, RMSprop from keras.callbacks import ModelCheckpoint from keras.utils import plot_model from keras.initializers import RandomNormal from functools import partial import numpy as np import json import os import pickle import matplotlib.pyplot as plt class RandomWeightedAverage(_Merge): def __init__(self, batch_size): super().__init__() self.batch_size = batch_size def _merge_function(self, inputs): alpha = K.random_uniform((self.batch_size, 1, 1, 1)) return (alpha * inputs[0]) + ((1 - alpha) * inputs[1]) class WGANGP(): def __init__(self , input_dim , critic_conv_filters , critic_conv_kernel_size , critic_conv_strides , critic_batch_norm_momentum , critic_activation , critic_dropout_rate , critic_learning_rate , generator_initial_dense_layer_size , generator_upsample , generator_conv_filters , generator_conv_kernel_size , generator_conv_strides , generator_batch_norm_momentum , generator_activation , generator_dropout_rate , generator_learning_rate , optimiser , grad_weight , z_dim , batch_size ): self.name = 'gan' self.input_dim = input_dim self.critic_conv_filters = critic_conv_filters self.critic_conv_kernel_size = critic_conv_kernel_size self.critic_conv_strides = critic_conv_strides self.critic_batch_norm_momentum = critic_batch_norm_momentum self.critic_activation = critic_activation self.critic_dropout_rate = critic_dropout_rate self.critic_learning_rate = critic_learning_rate self.generator_initial_dense_layer_size = generator_initial_dense_layer_size self.generator_upsample = generator_upsample self.generator_conv_filters = generator_conv_filters self.generator_conv_kernel_size = generator_conv_kernel_size self.generator_conv_strides = generator_conv_strides self.generator_batch_norm_momentum = generator_batch_norm_momentum self.generator_activation = generator_activation self.generator_dropout_rate = generator_dropout_rate self.generator_learning_rate = generator_learning_rate self.optimiser = optimiser self.z_dim = z_dim self.n_layers_critic = len(critic_conv_filters) self.n_layers_generator = len(generator_conv_filters) self.weight_init = RandomNormal(mean=0., stddev=0.02) ight self.batch_size = batch_size self.d_losses = [] self.g_losses = [] self.epoch = 0 self._build_critic() self._build_generator() self._build_adversarial() def gradient_penalty_loss(self, y_true, y_pred, interpolated_samples): gradients = K.gradients(y_pred, interpolated_samples)[0] gradients_sqr = K.square(gradients) gradients_sqr_sum = K.sum(gradients_sqr, axis=np.arange(1, len(gradients_sqr.shape))) gradient_l2_norm = K.sqrt(gradients_sqr_sum) gradient_penalty = K.square(1 - gradient_l2_norm) return K.mean(gradient_penalty) def wasserstein(self, y_true, y_pred): return -K.mean(y_true * y_pred) def get_activation(self, activation): if activation == 'leaky_relu': layer = LeakyReLU(alpha = 0.2) else: layer = Activation(activation) return layer def _build_critic(self): nput(shape=self.input_dim, name='critic_input') x = critic_input for i in range(self.n_layers_critic): x = Conv2D( filters = self.critic_conv_filters[i] , kernel_size = self.critic_conv_kernel_size[i] , strides = self.critic_conv_strides[i] , padding = 'same' , name = 'critic_conv_' + str(i) , kernel_initializer = self.weight_init )(x) if self.critic_batch_norm_momentum and i > 0: x = BatchNormalization(momentum = self.critic_batch_norm_momentum)(x) x = self.get_activation(self.critic_activation)(x) if self.critic_dropout_rate: x = Dropout(rate = self.critic_dropout_rate)(x) x = Flatten()(x) critic_output = Dense(1, activation=None , kernel_initializer = self.weight_init )(x) self.critic = Model(critic_input, critic_output) def _build_generator(self): ut(shape=(self.z_dim,), name='generator_input') x = generator_input x = Dense(np.prod(self.generator_initial_dense_layer_size), kernel_initializer = self.weight_init)(x) if self.generator_batch_norm_momentum: x = BatchNormalization(momentum = self.generator_batch_norm_momentum)(x) x = self.get_activation(self.generator_activation)(x) x = Reshape(self.generator_initial_dense_layer_size)(x) if self.generator_dropout_rate: x = Dropout(rate = self.generator_dropout_rate)(x) for i in range(self.n_layers_generator): if self.generator_upsample[i] == 2: x = UpSampling2D()(x) x = Conv2D( filters = self.generator_conv_filters[i] , kernel_size = self.generator_conv_kernel_size[i] , padding = 'same' , name = 'generator_conv_' + str(i) , kernel_initializer = self.weight_init )(x) else: x = Conv2DTranspose( filters = self.generator_conv_filters[i] , kernel_size = self.generator_conv_kernel_size[i] , padding = 'same' , strides = self.generator_conv_strides[i] , name = 'generator_conv_' + str(i) , kernel_initializer = self.weight_init )(x) if i < self.n_layers_generator - 1: if self.generator_batch_norm_momentum: x = BatchNormalization(momentum = self.generator_batch_norm_momentum)(x) x = self.get_activation(self.generator_activation)(x) else: x = Activation('tanh')(x) generator_output = x self.generator = Model(generator_input, generator_output) def get_opti(self, lr): if self.optimiser == 'adam': opti = Adam(lr=lr, beta_1=0.5) elif self.optimiser == 'rmsprop': opti = RMSprop(lr=lr) else: opti = Adam(lr=lr) return opti def set_trainable(self, m, val): m.trainable = val for l in m.layers: l.trainable = val def _build_adversarial(self): self.set_trainable(self.generator, False) # Image input (real sample) real_img = Input(shape=self.input_dim) # Fake image z_disc = Input(shape=(self.z_dim,)) fake_img = self.generator(z_disc) # critic determines validity of the real and fake images fake = self.critic(fake_img) valid = self.critic(real_img) # Construct weighted average between real and fake images interpolated_img = RandomWeightedAverage(self.batch_size)([real_img, fake_img]) # Determine validity of weighted sample validity_interpolated = self.critic(interpolated_img) # Use Python partial to provide loss function with additional # 'interpolated_samples' argument partial_gp_loss = partial(self.gradient_penalty_loss, interpolated_samples=interpolated_img) partial_gp_loss.__name__ = 'gradient_penalty' # Keras requires function names self.critic_model = Model(inputs=[real_img, z_disc], outputs=[valid, fake, validity_interpolated]) self.critic_model.compile( loss=[self.wasserstein,self.wasserstein, partial_gp_loss] ,optimizer=self.get_opti(self.critic_learning_rate) ,loss_weights=[1, 1, self.grad_weight] ) #------------------------------- # Construct Computational Graph # for Generator #------------------------------- # For the generator we freeze the critic's layers self.set_trainable(self.critic, False) self.set_trainable(self.generator, True) model_input = Input(shape=(self.z_dim,)) img = self.generator(model_input) model_output = self.critic(img) self.model = Model(model_input, model_output) self.model.compile(optimizer=self.get_opti(self.generator_learning_rate) , loss=self.wasserstein ) self.set_trainable(self.critic, True) def train_critic(self, x_train, batch_size, using_generator): valid = np.ones((batch_size,1), dtype=np.float32) fake = -np.ones((batch_size,1), dtype=np.float32) dummy = np.zeros((batch_size, 1), dtype=np.float32) if using_generator: true_imgs = next(x_train)[0] if true_imgs.shape[0] != batch_size: true_imgs = next(x_train)[0] else: idx = np.random.randint(0, x_train.shape[0], batch_size) true_imgs = x_train[idx] noise = np.random.normal(0, 1, (batch_size, self.z_dim)) d_loss = self.critic_model.train_on_batch([true_imgs, noise], [valid, fake, dummy]) return d_loss def train_generator(self, batch_size): valid = np.ones((batch_size,1), dtype=np.float32) noise = np.random.normal(0, 1, (batch_size, self.z_dim)) return self.model.train_on_batch(noise, valid) def train(self, x_train, batch_size, epochs, run_folder, print_every_n_batches = 10 , n_critic = 5 , using_generator = False): for epoch in range(self.epoch, self.epoch + epochs): if epoch % 100 == 0: critic_loops = 5 else: critic_loops = n_critic for _ in range(critic_loops): d_loss = self.train_critic(x_train, batch_size, using_generator) g_loss = self.train_generator(batch_size) print ("%d (%d, %d) [D loss: (%.1f)(R %.1f, F %.1f, G %.1f)] [G loss: %.1f]" % (epoch, critic_loops, 1, d_loss[0], d_loss[1],d_loss[2],d_loss[3],g_loss)) self.d_losses.append(d_loss) self.g_losses.append(g_loss) if epoch % print_every_n_batches == 0: self.sample_images(run_folder) self.model.save_weights(os.path.join(run_folder, 'weights/weights-%d.h5' % (epoch))) self.model.save_weights(os.path.join(run_folder, 'weights/weights.h5')) self.save_model(run_folder) self.epoch+=1 def sample_images(self, run_folder): r, c = 5, 5 noise = np.random.normal(0, 1, (r * c, self.z_dim)) gen_imgs = self.generator.predict(noise) gen_imgs = 0.5 * (gen_imgs + 1) gen_imgs = np.clip(gen_imgs, 0, 1) fig, axs = plt.subplots(r, c, figsize=(15,15)) cnt = 0 for i in range(r): for j in range(c): axs[i,j].imshow(np.squeeze(gen_imgs[cnt, :,:,:]), cmap = 'gray_r') axs[i,j].axis('off') cnt += 1 fig.savefig(os.path.join(run_folder, "images/sample_%d.png" % self.epoch)) plt.close() def plot_model(self, run_folder): plot_model(self.model, to_file=os.path.join(run_folder ,'viz/model.png'), show_shapes = True, show_layer_names = True) plot_model(self.critic, to_file=os.path.join(run_folder ,'viz/critic.png'), show_shapes = True, show_layer_names = True) plot_model(self.generator, to_file=os.path.join(run_folder ,'viz/generator.png'), show_shapes = True, show_layer_names = True) def save(self, folder): with open(os.path.join(folder, 'params.pkl'), 'wb') as f: pickle.dump([ self.input_dim , self.critic_conv_filters , self.critic_conv_kernel_size , self.critic_conv_strides , self.critic_batch_norm_momentum , self.critic_activation , self.critic_dropout_rate , self.critic_learning_rate , self.generator_initial_dense_layer_size , self.generator_upsample , self.generator_conv_filters , self.generator_conv_kernel_size , self.generator_conv_strides , self.generator_batch_norm_momentum , self.generator_activation , self.generator_dropout_rate , self.generator_learning_rate , self.optimiser , self.grad_weight , self.z_dim , self.batch_size ], f) self.plot_model(folder) def save_model(self, run_folder): self.model.save(os.path.join(run_folder, 'model.h5')) self.critic.save(os.path.join(run_folder, 'critic.h5')) self.generator.save(os.path.join(run_folder, 'generator.h5')) pickle.dump(self, open( os.path.join(run_folder, "obj.pkl"), "wb" )) def load_weights(self, filepath): self.model.load_weights(filepath)
true
true
1c4839afde50eb8dd507b972de44d105bb02aea1
1,077
py
Python
tests/application/cms/test_filters.py
AlexKouzy/ethnicity-facts-and-figures-publisher
18ab2495a8633f585e18e607c7f75daa564a053d
[ "MIT" ]
1
2021-10-06T13:48:36.000Z
2021-10-06T13:48:36.000Z
tests/application/cms/test_filters.py
AlexKouzy/ethnicity-facts-and-figures-publisher
18ab2495a8633f585e18e607c7f75daa564a053d
[ "MIT" ]
116
2018-11-02T17:20:47.000Z
2022-02-09T11:06:22.000Z
tests/application/cms/test_filters.py
racedisparityaudit/rd_cms
a12f0e3f5461cc41eed0077ed02e11efafc5dd76
[ "MIT" ]
2
2018-11-09T16:47:35.000Z
2020-04-09T13:06:48.000Z
import pytest from application.cms.filters import index_of_last_initial_zero, yesno class TestYesNo: @pytest.mark.parametrize( "input_value, expected_output", ((True, "yes"), (False, "no"), (1, 1), (0, 0), ("true", "true"), ("false", "false"), ("abc", "abc")), ) def test_yesno_converts_boolean_true_and_false_only(self, input_value, expected_output): assert yesno(input_value) == expected_output class TestIndexOfLastInitialZero: def test_when_only_one_zero(self): assert index_of_last_initial_zero([0, 10, 20]) == 0 def test_when_many_zeros(self): assert index_of_last_initial_zero([0, 0, 0, 0, 1, 2]) == 3 def test_when_later_zeros_are_present(self): assert index_of_last_initial_zero([0, 0, 1, 2, 1, 0]) == 1 def test_when_no_zeros_are_present(self): with pytest.raises(ValueError): index_of_last_initial_zero([1, 2, 3, 4]) def test_when_array_contains_strings(self): with pytest.raises(ValueError): index_of_last_initial_zero(["0", "1", "2"])
33.65625
109
0.673166
import pytest from application.cms.filters import index_of_last_initial_zero, yesno class TestYesNo: @pytest.mark.parametrize( "input_value, expected_output", ((True, "yes"), (False, "no"), (1, 1), (0, 0), ("true", "true"), ("false", "false"), ("abc", "abc")), ) def test_yesno_converts_boolean_true_and_false_only(self, input_value, expected_output): assert yesno(input_value) == expected_output class TestIndexOfLastInitialZero: def test_when_only_one_zero(self): assert index_of_last_initial_zero([0, 10, 20]) == 0 def test_when_many_zeros(self): assert index_of_last_initial_zero([0, 0, 0, 0, 1, 2]) == 3 def test_when_later_zeros_are_present(self): assert index_of_last_initial_zero([0, 0, 1, 2, 1, 0]) == 1 def test_when_no_zeros_are_present(self): with pytest.raises(ValueError): index_of_last_initial_zero([1, 2, 3, 4]) def test_when_array_contains_strings(self): with pytest.raises(ValueError): index_of_last_initial_zero(["0", "1", "2"])
true
true
1c483aecbbbdbbb994f33b24f66067faffd38da9
17,014
py
Python
modules/templates/CCC/menus.py
himansu1997/eden
1e2cf2b00f55da46b1ce3e6b7ad44b5345d7a1dc
[ "MIT" ]
205
2015-01-20T08:26:09.000Z
2022-03-27T19:59:33.000Z
modules/templates/CCC/menus.py
himansu1997/eden
1e2cf2b00f55da46b1ce3e6b7ad44b5345d7a1dc
[ "MIT" ]
249
2015-02-10T09:56:35.000Z
2022-03-23T19:54:36.000Z
modules/templates/CCC/menus.py
himansu1997/eden
1e2cf2b00f55da46b1ce3e6b7ad44b5345d7a1dc
[ "MIT" ]
231
2015-02-10T09:33:17.000Z
2022-02-18T19:56:05.000Z
# -*- coding: utf-8 -*- from gluon import current, URL #from s3 import IS_ISO639_2_LANGUAGE_CODE from s3layouts import M, MM try: from .layouts import * except ImportError: pass import s3menus as default # ============================================================================= class S3MainMenu(default.S3MainMenu): """ Custom Application Main Menu """ # ------------------------------------------------------------------------- @classmethod def menu(cls): """ Compose Menu """ # Modules menus main_menu = MM()( cls.menu_modules(), ) # Additional menus current.menu.personal = cls.menu_personal() #current.menu.lang = cls.menu_lang() current.menu.about = cls.menu_about() return main_menu # ------------------------------------------------------------------------- @classmethod def menu_modules(cls): """ Custom Modules Menu """ auth = current.auth if not auth.is_logged_in(): menu = [MM("Volunteer Your Time", c="default", f="index", args="volunteer"), #MM("Donate Items", c="default", f="index", args="donate"), ] return menu has_role = auth.s3_has_role if has_role("ADMIN"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("All Documents", c="doc", f="document", m="datalist"), MM("Affected People", c="br", f="person")( MM("Import", m="import"), ), MM("Donors", c="pr", f="person", vars={"donors": 1})( MM("Donations", c="supply", f="person_item"), MM("Edit General Information", c="cms", f="post", vars={"~.name": "Donor"}, m="update"), ), MM("Organisations", c="org", f="organisation", m="summary")( MM("Import", m="import"), #MM("Message", c="org", f="organisation", args="message"), ), MM("Volunteers", c="hrm", f="human_resource")( MM("Reserves", c="pr", f="person", vars={"reserves": 1}), MM("Reserve Groups", c="pr", f="group"), MM("Inactives", c="pr", f="person", vars={"inactive": 1}), ), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Messages", c="project", f="task"), ] elif has_role("ORG_ADMIN"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisation Documents", c="doc", f="document", m="datalist"), MM("Donors", c="pr", f="person", vars={"donors": 1})( MM("Donations", c="supply", f="person_item"), ), MM("Organisations", c="org", f="organisation", m="summary")( #MM("Message", c="org", f="organisation", args="message"), ), MM("Volunteers", c="hrm", f="human_resource")( MM("Reserves", c="pr", f="person", vars={"reserves": 1}), #MM("Reserve Groups", c="pr", f="group"), ), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Messages", c="project", f="task"), ] elif has_role("AGENCY"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Documents", c="doc", f="document", m="datalist"), MM("Affected People", c="br", f="person")( MM("Import", c="br", f="person", m="import"), ), MM("Donors", c="pr", f="person", vars={"donors": 1})( MM("Donations", c="supply", f="person_item"), ), MM("Organisations", c="org", f="organisation", m="summary")( #MM("Message", c="org", f="organisation", args="message"), ), MM("Volunteers", c="hrm", f="human_resource")( MM("Reserves", c="pr", f="person", vars={"reserves": 1}), MM("Reserve Groups", c="pr", f="group"), ), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Messages", c="project", f="task")( MM("Contact Organisation Admins", c="project", f="task", m="create"), ), ] elif has_role("VOLUNTEER"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisation Documents", c="doc", f="document", m="datalist"), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Contact Organisation Admins", c="project", f="task", m="create"), ] elif has_role("GROUP_ADMIN"): menu = [#MM("Volunteer Your Time", c="default", f="index", args="volunteer"), #MM("Donate Items", c="default", f="index", args="donate"), MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Group", c="pr", f="group", m="update"), ] elif has_role("DONOR"): menu = [#MM("Volunteer Your Time", c="default", f="index", args="volunteer"), #MM("Donate Items", c="default", f="index", args="donate"), MM("General Information", c="default", f="index", m="donor"), MM("Messages", c="project", f="task"), ] elif has_role("RESERVE"): # Reserve Volunteer menu = [#MM("Volunteer Your Time", c="default", f="index", args="volunteer"), #MM("Donate Items", c="default", f="index", args="donate"), MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisations", c="org", f="organisation", m="summary"), MM("Events", c="hrm", f="training_event"), # They can only see ones they're invited to MM("Opportunities", c="req", f="need"), # They can only see ones they're invited to ] else: # Inactive Volunteer menu = [#MM("Volunteer Your Time", c="default", f="index", args="volunteer"), #MM("Donate Items", c="default", f="index", args="donate"), #MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisations", c="org", f="organisation", m="summary"), #MM("Events", c="hrm", f="training_event"), # They can only see ones they're invited to #MM("Opportunities", c="req", f="need"), # They can only see ones they're invited to ] return menu # ------------------------------------------------------------------------- #@classmethod #def menu_lang(cls): # """ Language Selector """ # languages = current.deployment_settings.get_L10n_languages() # represent_local = IS_ISO639_2_LANGUAGE_CODE.represent_local # menu_lang = ML("Language", right=True) # for code in languages: # # Show each language name in its own language # lang_name = represent_local(code) # menu_lang(ML(lang_name, # translate = False, # lang_code = code, # lang_name = lang_name, # ) # ) # return menu_lang # ------------------------------------------------------------------------- @classmethod def menu_personal(cls): """ Personal Menu """ auth = current.auth #s3 = current.response.s3 #settings = current.deployment_settings if not auth.is_logged_in(): request = current.request login_next = URL(args=request.args, vars=request.vars) if request.controller == "default" and \ request.function == "user" and \ "_next" in request.get_vars: login_next = request.get_vars["_next"] #self_registration = settings.get_security_self_registration() menu_personal = MP()( #MP("Register", c="default", f="user", # m = "register", # check = self_registration, # ), MP("Login", c="default", f="user", m = "login", vars = {"_next": login_next}, ), ) #if settings.get_auth_password_retrieval(): # menu_personal(MP("Lost Password", c="default", f="user", # m = "retrieve_password", # ), # ) else: ADMIN = current.auth.get_system_roles().ADMIN s3_has_role = auth.s3_has_role is_org_admin = lambda i: not s3_has_role(ADMIN) and \ s3_has_role("ORG_ADMIN") menu_personal = MP()( MP("Administration", c="admin", f="index", restrict = ADMIN, ), MP("Administration", c="admin", f="user", check = is_org_admin, ), MP("Profile", c="default", f="person"), MP("Change Password", c="default", f="user", m = "change_password", ), MP("Logout", c="default", f="user", m = "logout", ), ) return menu_personal # ------------------------------------------------------------------------- @classmethod def menu_about(cls): #ADMIN = current.auth.get_system_roles().ADMIN menu_about = MA(c="default")( MA("Help", f="help"), MA("Contact Us", f="contact"), #MA("Version", f="about", restrict = ADMIN), ) return menu_about # ============================================================================= class S3OptionsMenu(default.S3OptionsMenu): """ Custom Controller Menus """ # ------------------------------------------------------------------------- def admin(self): """ ADMIN menu """ if not current.auth.s3_has_role("ADMIN"): # OrgAdmin: No Side-menu return None #settings_messaging = self.settings_messaging() #settings = current.deployment_settings #consent_tracking = lambda i: settings.get_auth_consent_tracking() #is_data_repository = lambda i: settings.get_sync_data_repository() #translate = settings.has_module("translate") # NB: Do not specify a controller for the main menu to allow # re-use of this menu by other controllers return M()( #M("Setup", c="setup", f="deployment")( # #M("Create", m="create"), # #M("Servers", f="server")( # #), # #M("Instances", f="instance")( # #), #), #M("Settings", c="admin", f="setting")( # settings_messaging, #), M("User Management", c="admin", f="user")( M("Create User", m="create"), M("List All Users"), M("Import Users", m="import"), M("List All Roles", f="role"), #M("List All Organization Approvers & Whitelists", f="organisation"), #M("Roles", f="group"), #M("Membership", f="membership"), ), #M("Consent Tracking", c="admin", link=False, check=consent_tracking)( M("Consent Tracking", c="admin", link=False)( M("Processing Types", f="processing_type"), M("Consent Options", f="consent_option"), ), #M("Goods / Services", c="supply", f="item")(), #M("Qualifications", c="hrm", f="certificate")(), M("Organizations", c="org", f="organisation")( M("Types", f="organisation_type"), M("Job Titles", c="hrm", f="job_title"), ), #M("Time Slots", c="pr", f="slot")(), #M("Volunteer Offers", c="hrm", f="skill")(), #M("CMS", c="cms", f="post")( #), M("Database", c="appadmin", f="index")( M("Raw Database access", c="appadmin", f="index") ), M("Error Tickets", c="admin", f="errors"), #M("Monitoring", c="setup", f="server")( # M("Checks", f="monitor_check"), # M("Servers", f="server"), # M("Tasks", f="monitor_task"), # M("Logs", f="monitor_run"), #), M("Scheduler", c="admin", f="task"), #M("Synchronization", c="sync", f="index")( # M("Settings", f="config", args=[1], m="update"), # M("Repositories", f="repository"), # M("Public Data Sets", f="dataset", check=is_data_repository), # M("Log", f="log"), #), #M("Edit Application", a="admin", c="default", f="design", #args=[request.application]), #M("Translation", c="admin", f="translate", check=translate)( # M("Select Modules for translation", c="admin", f="translate", # m="create", vars=dict(opt="1")), # M("Upload translated files", c="admin", f="translate", # m="create", vars=dict(opt="2")), # M("View Translation Percentage", c="admin", f="translate", # m="create", vars=dict(opt="3")), # M("Add strings manually", c="admin", f="translate", # m="create", vars=dict(opt="4")) #), #M("View Test Result Reports", c="admin", f="result"), #M("Portable App", c="admin", f="portable") ) # ------------------------------------------------------------------------- @staticmethod def br(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def cms(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def doc(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def hrm(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def org(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def pr(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def project(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def req(): """ No Side Menu """ return None # ------------------------------------------------------------------------- @staticmethod def supply(): """ No Side Menu """ return None # END =========================================================================
43.514066
111
0.401846
from gluon import current, URL from s3layouts import M, MM try: from .layouts import * except ImportError: pass import s3menus as default class S3MainMenu(default.S3MainMenu): @classmethod def menu(cls): main_menu = MM()( cls.menu_modules(), ) current.menu.personal = cls.menu_personal() current.menu.about = cls.menu_about() return main_menu @classmethod def menu_modules(cls): auth = current.auth if not auth.is_logged_in(): menu = [MM("Volunteer Your Time", c="default", f="index", args="volunteer"), ] return menu has_role = auth.s3_has_role if has_role("ADMIN"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("All Documents", c="doc", f="document", m="datalist"), MM("Affected People", c="br", f="person")( MM("Import", m="import"), ), MM("Donors", c="pr", f="person", vars={"donors": 1})( MM("Donations", c="supply", f="person_item"), MM("Edit General Information", c="cms", f="post", vars={"~.name": "Donor"}, m="update"), ), MM("Organisations", c="org", f="organisation", m="summary")( MM("Import", m="import"), ), MM("Volunteers", c="hrm", f="human_resource")( MM("Reserves", c="pr", f="person", vars={"reserves": 1}), MM("Reserve Groups", c="pr", f="group"), MM("Inactives", c="pr", f="person", vars={"inactive": 1}), ), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Messages", c="project", f="task"), ] elif has_role("ORG_ADMIN"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisation Documents", c="doc", f="document", m="datalist"), MM("Donors", c="pr", f="person", vars={"donors": 1})( MM("Donations", c="supply", f="person_item"), ), MM("Organisations", c="org", f="organisation", m="summary")( ), MM("Volunteers", c="hrm", f="human_resource")( MM("Reserves", c="pr", f="person", vars={"reserves": 1}), ), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Messages", c="project", f="task"), ] elif has_role("AGENCY"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Documents", c="doc", f="document", m="datalist"), MM("Affected People", c="br", f="person")( MM("Import", c="br", f="person", m="import"), ), MM("Donors", c="pr", f="person", vars={"donors": 1})( MM("Donations", c="supply", f="person_item"), ), MM("Organisations", c="org", f="organisation", m="summary")( ), MM("Volunteers", c="hrm", f="human_resource")( MM("Reserves", c="pr", f="person", vars={"reserves": 1}), MM("Reserve Groups", c="pr", f="group"), ), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Messages", c="project", f="task")( MM("Contact Organisation Admins", c="project", f="task", m="create"), ), ] elif has_role("VOLUNTEER"): menu = [MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisation Documents", c="doc", f="document", m="datalist"), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), MM("Contact Organisation Admins", c="project", f="task", m="create"), ] elif has_role("GROUP_ADMIN"): menu = [ MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Group", c="pr", f="group", m="update"), ] elif has_role("DONOR"): menu = [ MM("General Information", c="default", f="index", m="donor"), MM("Messages", c="project", f="task"), ] elif has_role("RESERVE"): menu = [ MM("General Information and Advice", c="cms", f="post", m="datalist"), MM("Organisations", c="org", f="organisation", m="summary"), MM("Events", c="hrm", f="training_event"), MM("Opportunities", c="req", f="need"), # They can only see ones they're invited to ] else: menu = [ MM("Organisations", c="org", f="organisation", m="summary"), ="req", f="need"), # They can only see ones they're invited to ] return menu @classmethod def menu_personal(cls): auth = current.auth if not auth.is_logged_in(): request = current.request login_next = URL(args=request.args, vars=request.vars) if request.controller == "default" and \ request.function == "user" and \ "_next" in request.get_vars: login_next = request.get_vars["_next"] menu_personal = MP()( MP("Login", c="default", f="user", m = "login", vars = {"_next": login_next}, ), ) else: ADMIN = current.auth.get_system_roles().ADMIN s3_has_role = auth.s3_has_role is_org_admin = lambda i: not s3_has_role(ADMIN) and \ s3_has_role("ORG_ADMIN") menu_personal = MP()( MP("Administration", c="admin", f="index", restrict = ADMIN, ), MP("Administration", c="admin", f="user", check = is_org_admin, ), MP("Profile", c="default", f="person"), MP("Change Password", c="default", f="user", m = "change_password", ), MP("Logout", c="default", f="user", m = "logout", ), ) return menu_personal @classmethod def menu_about(cls): menu_about = MA(c="default")( MA("Help", f="help"), MA("Contact Us", f="contact"), ) return menu_about class S3OptionsMenu(default.S3OptionsMenu): def admin(self): if not current.auth.s3_has_role("ADMIN"): return None return M()( M("User Management", c="admin", f="user")( M("Create User", m="create"), M("List All Users"), M("Import Users", m="import"), M("List All Roles", f="role"), ), M("Consent Tracking", c="admin", link=False)( M("Processing Types", f="processing_type"), M("Consent Options", f="consent_option"), ), M("Organizations", c="org", f="organisation")( M("Types", f="organisation_type"), M("Job Titles", c="hrm", f="job_title"), ), M("Database", c="appadmin", f="index")( M("Raw Database access", c="appadmin", f="index") ), M("Error Tickets", c="admin", f="errors"), M("Scheduler", c="admin", f="task"), ) @staticmethod def br(): return None @staticmethod def cms(): return None @staticmethod def doc(): return None @staticmethod def hrm(): return None @staticmethod def org(): return None @staticmethod def pr(): return None @staticmethod def project(): return None @staticmethod def req(): return None @staticmethod def supply(): return None
true
true
1c483ccf406e95ea8d666dbe860d047dfb31581a
3,024
py
Python
flask/lib/python3.6/site-packages/stem/interpreter/autocomplete.py
JOFLIX/grapevines
34576e01184570d79cc140b42ffb71d322132da6
[ "MIT", "Unlicense" ]
null
null
null
flask/lib/python3.6/site-packages/stem/interpreter/autocomplete.py
JOFLIX/grapevines
34576e01184570d79cc140b42ffb71d322132da6
[ "MIT", "Unlicense" ]
3
2019-07-29T09:47:34.000Z
2019-07-29T09:47:35.000Z
flask/lib/python3.6/site-packages/stem/interpreter/autocomplete.py
JOFLIX/grapevines
34576e01184570d79cc140b42ffb71d322132da6
[ "MIT", "Unlicense" ]
null
null
null
# Copyright 2014-2017, Damian Johnson and The Tor Project # See LICENSE for licensing information """ Tab completion for our interpreter prompt. """ from stem.interpreter import uses_settings try: # added in python 3.2 from functools import lru_cache except ImportError: from stem.util.lru_cache import lru_cache @uses_settings def _get_commands(controller, config): """ Provides commands recognized by tor. """ commands = config.get('autocomplete', []) if controller is None: return commands # GETINFO commands. Lines are of the form '[option] -- [description]'. This # strips '*' from options that accept values. results = controller.get_info('info/names', None) if results: for line in results.splitlines(): option = line.split(' ', 1)[0].rstrip('*') commands.append('GETINFO %s' % option) else: commands.append('GETINFO ') # GETCONF, SETCONF, and RESETCONF commands. Lines are of the form # '[option] [type]'. results = controller.get_info('config/names', None) if results: for line in results.splitlines(): option = line.split(' ', 1)[0] commands.append('GETCONF %s' % option) commands.append('SETCONF %s' % option) commands.append('RESETCONF %s' % option) else: commands += ['GETCONF ', 'SETCONF ', 'RESETCONF '] # SETEVENT, USEFEATURE, and SIGNAL commands. For each of these the GETINFO # results are simply a space separated lists of the values they can have. options = ( ('SETEVENTS ', 'events/names'), ('USEFEATURE ', 'features/names'), ('SIGNAL ', 'signal/names'), ) for prefix, getinfo_cmd in options: results = controller.get_info(getinfo_cmd, None) if results: commands += [prefix + value for value in results.split()] else: commands.append(prefix) # Adds /help commands. usage_info = config.get('help.usage', {}) for cmd in usage_info.keys(): commands.append('/help ' + cmd) return commands class Autocompleter(object): def __init__(self, controller): self._commands = _get_commands(controller) @lru_cache() def matches(self, text): """ Provides autocompletion matches for the given text. :param str text: text to check for autocompletion matches with :returns: **list** with possible matches """ lowercase_text = text.lower() return [cmd for cmd in self._commands if cmd.lower().startswith(lowercase_text)] def complete(self, text, state): """ Provides case insensetive autocompletion options, acting as a functor for the readlines set_completer function. :param str text: text to check for autocompletion matches with :param int state: index of result to be provided, readline fetches matches until this function provides None :returns: **str** with the autocompletion match, **None** if eithe none exists or state is higher than our number of matches """ try: return self.matches(text)[state] except IndexError: return None
26.068966
84
0.680886
from stem.interpreter import uses_settings try: from functools import lru_cache except ImportError: from stem.util.lru_cache import lru_cache @uses_settings def _get_commands(controller, config): commands = config.get('autocomplete', []) if controller is None: return commands results = controller.get_info('info/names', None) if results: for line in results.splitlines(): option = line.split(' ', 1)[0].rstrip('*') commands.append('GETINFO %s' % option) else: commands.append('GETINFO ') results = controller.get_info('config/names', None) if results: for line in results.splitlines(): option = line.split(' ', 1)[0] commands.append('GETCONF %s' % option) commands.append('SETCONF %s' % option) commands.append('RESETCONF %s' % option) else: commands += ['GETCONF ', 'SETCONF ', 'RESETCONF '] options = ( ('SETEVENTS ', 'events/names'), ('USEFEATURE ', 'features/names'), ('SIGNAL ', 'signal/names'), ) for prefix, getinfo_cmd in options: results = controller.get_info(getinfo_cmd, None) if results: commands += [prefix + value for value in results.split()] else: commands.append(prefix) usage_info = config.get('help.usage', {}) for cmd in usage_info.keys(): commands.append('/help ' + cmd) return commands class Autocompleter(object): def __init__(self, controller): self._commands = _get_commands(controller) @lru_cache() def matches(self, text): lowercase_text = text.lower() return [cmd for cmd in self._commands if cmd.lower().startswith(lowercase_text)] def complete(self, text, state): try: return self.matches(text)[state] except IndexError: return None
true
true
1c483ce4f303e0026de5ff70630090340b35fd96
836
py
Python
backend/tasks.py
ioxio-nexus/mycompany-consent-demo
aefa69375c14dfb345e81aad203db223cec6afe8
[ "BSD-3-Clause" ]
null
null
null
backend/tasks.py
ioxio-nexus/mycompany-consent-demo
aefa69375c14dfb345e81aad203db223cec6afe8
[ "BSD-3-Clause" ]
null
null
null
backend/tasks.py
ioxio-nexus/mycompany-consent-demo
aefa69375c14dfb345e81aad203db223cec6afe8
[ "BSD-3-Clause" ]
null
null
null
from os import environ import uvicorn from invoke import task from uvicorn.supervisors import ChangeReload DEV_ENV = {"FIRESTORE_EMULATOR_HOST": "127.0.0.1:8686"} @task def dev(ctx): environ.update(DEV_ENV) port = environ.get("PORT", 8000) host = "0.0.0.0" # nosec, it's not a mistake config = uvicorn.Config(app="main:app", host=host, port=int(port), debug=True) server = uvicorn.Server(config) from app.log import logger # noqa, must be imported before running supervisor supervisor = ChangeReload(config, target=server.run, sockets=[config.bind_socket()]) supervisor.run() @task def serve(ctx): server = uvicorn.Server( uvicorn.Config( app="main:app", uds="/run/nginx/uvicorn.sock", forwarded_allow_ips="*", ), ) server.run()
23.885714
88
0.65311
from os import environ import uvicorn from invoke import task from uvicorn.supervisors import ChangeReload DEV_ENV = {"FIRESTORE_EMULATOR_HOST": "127.0.0.1:8686"} @task def dev(ctx): environ.update(DEV_ENV) port = environ.get("PORT", 8000) host = "0.0.0.0" config = uvicorn.Config(app="main:app", host=host, port=int(port), debug=True) server = uvicorn.Server(config) from app.log import logger # noqa, must be imported before running supervisor supervisor = ChangeReload(config, target=server.run, sockets=[config.bind_socket()]) supervisor.run() @task def serve(ctx): server = uvicorn.Server( uvicorn.Config( app="main:app", uds="/run/nginx/uvicorn.sock", forwarded_allow_ips="*", ), ) server.run()
true
true
1c483d9e661792fda761dec9b82a4f18dbf7a9aa
1,126
py
Python
tools/trainercard/trainercard.py
stoiandan/OpenPokemonRed
3ce2483d4620255c7fe182012f2821be3121c375
[ "MIT" ]
204
2020-11-04T07:32:28.000Z
2022-01-16T20:39:22.000Z
tools/trainercard/trainercard.py
stoiandan/OpenPokemonRed
3ce2483d4620255c7fe182012f2821be3121c375
[ "MIT" ]
11
2020-10-26T07:53:24.000Z
2021-01-07T19:03:09.000Z
tools/trainercard/trainercard.py
stoiandan/OpenPokemonRed
3ce2483d4620255c7fe182012f2821be3121c375
[ "MIT" ]
14
2020-11-21T22:02:28.000Z
2022-02-15T15:26:55.000Z
import cv2 import os import shutil if os.path.exists("result"): shutil.rmtree("result") os.mkdir("result") # https://www.spriters-resource.com/fullview/8733/ img = cv2.imread("trainercard.png") leader = [ "brock", "misty", "lt_surge", "erika", "koga", "sabrina", "blaine", "giovanni" ] width = 16 height = 16 face = [ [31, 103], [31+32, 103], [31+32+32, 103], [31+32+32+32, 103], [31, 127], [31+32, 127], [31+32+32, 127], [31+32+32+32, 127], ] badge = [ [31, 168], [31+32, 168], [31+32+32, 168], [31+32+32+32, 168], [31, 192], [31+32, 192], [31+32+32, 192], [31+32+32+32, 192], ] # face for i in range(8): name = leader[i] x0 = face[i][0] y0 = face[i][1] x1 = x0 + width y1 = y0 + height tile = img[y0:y1, x0:x1] cv2.imwrite("./result/{}_face.png".format(name), tile) # badge for i in range(8): name = leader[i] x0 = badge[i][0] y0 = badge[i][1] x1 = x0 + width y1 = y0 + height tile = img[y0:y1, x0:x1] cv2.imwrite("./result/{}_badge.png".format(name), tile)
16.318841
59
0.519538
import cv2 import os import shutil if os.path.exists("result"): shutil.rmtree("result") os.mkdir("result") img = cv2.imread("trainercard.png") leader = [ "brock", "misty", "lt_surge", "erika", "koga", "sabrina", "blaine", "giovanni" ] width = 16 height = 16 face = [ [31, 103], [31+32, 103], [31+32+32, 103], [31+32+32+32, 103], [31, 127], [31+32, 127], [31+32+32, 127], [31+32+32+32, 127], ] badge = [ [31, 168], [31+32, 168], [31+32+32, 168], [31+32+32+32, 168], [31, 192], [31+32, 192], [31+32+32, 192], [31+32+32+32, 192], ] for i in range(8): name = leader[i] x0 = face[i][0] y0 = face[i][1] x1 = x0 + width y1 = y0 + height tile = img[y0:y1, x0:x1] cv2.imwrite("./result/{}_face.png".format(name), tile) for i in range(8): name = leader[i] x0 = badge[i][0] y0 = badge[i][1] x1 = x0 + width y1 = y0 + height tile = img[y0:y1, x0:x1] cv2.imwrite("./result/{}_badge.png".format(name), tile)
true
true
1c483dcfb05352b30e44f9812d4f220140dfff77
23,987
py
Python
AgentRun.py
zhangtjtongxue/DL_RL_Zoo
fe8393a941a8c22205b9dc5534f399cf7860f409
[ "Apache-2.0" ]
1
2021-06-08T08:20:31.000Z
2021-06-08T08:20:31.000Z
AgentRun.py
zhangtjtongxue/DL_RL_Zoo
fe8393a941a8c22205b9dc5534f399cf7860f409
[ "Apache-2.0" ]
null
null
null
AgentRun.py
zhangtjtongxue/DL_RL_Zoo
fe8393a941a8c22205b9dc5534f399cf7860f409
[ "Apache-2.0" ]
null
null
null
import os import sys import gym import torch import numpy as np from AgentZoo import Recorder from AgentZoo import BufferArray, initial_exploration """ 2019-07-01 Zen4Jia1Hao2, GitHub: YonV1943 DL_RL_Zoo RL 2019-11-11 Issay-0.0 [Essay Consciousness] 2020-02-02 Issay-0.1 Deep Learning Techniques (spectral norm, DenseNet, etc.) 2020-04-04 Issay-0.1 [An Essay of Consciousness by YonV1943], IntelAC 2020-04-20 Issay-0.2 SN_AC, IntelAC_UnitedLoss 2020-04-22 Issay-0.2 [Essay, LongDear's Cerebellum (Little Brain)] 2020-06-06 Issay-0.3 check PPO, SAC. Plan to add discrete SAC, EBM(soft-q-learning) I consider that Reinforcement Learning Algorithms before 2020 have not consciousness They feel more like a Cerebellum (Little Brain) for Machines. In my opinion, before 2020, the policy gradient algorithm agent didn't learn s policy. Actually, they "learn game feel" or "get a soft touch". In Chinese "shou3 gan3". Learn more about policy gradient algorithms in: https://lilianweng.github.io/lil-log/2018/04/08/policy-gradient-algorithms.html 2020-04-28 Add Discrete Env CartPole, Pendulum """ class Arguments: # default working setting and hyper-parameter def __init__(self, class_agent): self.class_agent = class_agent self.net_dim = 2 ** 7 # the network width self.max_step = 2 ** 10 # max steps in one epoch self.max_memo = 2 ** 17 # memories capacity (memories: replay buffer) self.max_epoch = 2 ** 10 # max times of train_epoch self.batch_size = 2 ** 7 # num of transitions sampled from replay buffer. self.repeat_times = 1 # Two-time Update Rule (TTUR) self.reward_scale = 2 ** 0 # an approximate target reward usually be closed to 256 self.gamma = 0.99 # discount factor of future rewards self.gpu_id = 0 self.random_seed = 19430 self.is_remove = True # remove the pre-training data? (True, False, None:ask me) self.env_name = "LunarLanderContinuous-v2" self.cwd = 'AC_Methods_LL' # current work directory self.show_gap = 2 ** 7 # show the Reward and Loss of actor and critic per show_gap seconds def init_for_training(self): # remove cwd, choose GPU, set random seed, set CPU threads print('GPU: {} | CWD: {}'.format(self.gpu_id, self.cwd)) whether_remove_history(self.cwd, self.is_remove) os.environ['CUDA_VISIBLE_DEVICES'] = str(self.gpu_id) # env.seed() # env has random seed too. np.random.seed(self.random_seed) torch.manual_seed(self.random_seed) torch.set_default_dtype(torch.float32) torch.set_num_threads(8) def train_agent__off_policy( class_agent, net_dim, batch_size, repeat_times, gamma, reward_scale, cwd, env_name, max_step, max_memo, max_epoch, **_kwargs): # 2020-06-01 env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=False) assert not is_discrete '''init''' agent = class_agent(state_dim, action_dim, net_dim) # training agent agent.state = env.reset() buffer = BufferArray(max_memo, state_dim, action_dim) # experiment replay buffer recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) # unnecessary '''loop''' with torch.no_grad(): # update replay buffer # rewards, steps = agent.update_buffer(env, buffer, max_step, max_action, reward_scale, gamma) rewards, steps = initial_exploration(env, buffer, max_step, max_action, reward_scale, gamma, action_dim) recorder.show_reward(rewards, steps, loss_a=0, loss_c=0) try: for epoch in range(max_epoch): # update replay buffer by interact with environment with torch.no_grad(): # for saving the GPU buffer rewards, steps = agent.update_buffer( env, buffer, max_step, max_action, reward_scale, gamma) # update network parameters by random sampling buffer for gradient descent buffer.init_before_sample() loss_a, loss_c = agent.update_parameters( buffer, max_step, batch_size, repeat_times) # show/check the reward, save the max reward actor with torch.no_grad(): # for saving the GPU buffer # NOTICE! Recorder saves the agent with max reward automatically. recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("| raise KeyboardInterrupt and break training loop") # except AssertionError: # for BipedWalker BUG 2020-03-03 # print("AssertionError: OpenAI gym r.LengthSquared() > 0.0f ??? Please run again.") train_time = recorder.print_and_save_npy(env_name, cwd) if is_solved: agent.save_or_load_model(cwd, is_save=True) draw_plot_with_npy(cwd, train_time) def train_agent__on_policy( class_agent, net_dim, batch_size, repeat_times, gamma, reward_scale, cwd, env_name, max_step, max_memo, max_epoch, **_kwargs): # 2020-0430 env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=True) agent = class_agent(state_dim, action_dim, net_dim) agent.save_or_load_model(cwd, is_save=False) recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) try: for epoch in range(max_epoch): with torch.no_grad(): # just the GPU memory rewards, steps, buffer = agent.update_buffer_online( env, max_step, max_memo, max_action, reward_scale, gamma) loss_a, loss_c = agent.update_parameters_online( buffer, batch_size, repeat_times) with torch.no_grad(): # just the GPU memory recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("raise KeyboardInterrupt while training.") except AssertionError: # for BipedWalker BUG 2020-03-03 print("AssertionError: OpenAI gym r.LengthSquared() > 0.0f ??? Please run again.") return False train_time = recorder.print_and_save_npy(env_name, cwd) draw_plot_with_npy(cwd, train_time) return True def train_agent_discrete( class_agent, net_dim, batch_size, repeat_times, gamma, reward_scale, cwd, env_name, max_step, max_memo, max_epoch, **_kwargs): # 2020-05-20 env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=True) assert is_discrete '''init''' agent = class_agent(state_dim, action_dim, net_dim) # training agent agent.state = env.reset() buffer = BufferArray(max_memo, state_dim, action_dim=1) # experiment replay buffer recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) '''loop''' with torch.no_grad(): # update replay buffer rewards, steps = initial_exploration( env, buffer, max_step, max_action, reward_scale, gamma, action_dim) recorder.show_reward(rewards, steps, loss_a=0, loss_c=0) try: for epoch in range(max_epoch): # update replay buffer by interact with environment with torch.no_grad(): # for saving the GPU buffer rewards, steps = agent.update_buffer( env, buffer, max_step, max_action, reward_scale, gamma) # update network parameters by random sampling buffer for gradient descent buffer.init_before_sample() loss_a, loss_c = agent.update_parameters(buffer, max_step, batch_size, repeat_times) # show/check the reward, save the max reward actor with torch.no_grad(): # for saving the GPU buffer # NOTICE! Recorder saves the agent with max reward automatically. recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("| raise KeyboardInterrupt and break training loop") # except AssertionError: # for BipedWalker BUG 2020-03-03 # print("AssertionError: OpenAI gym r.LengthSquared() > 0.0f ??? Please run again.") train_time = recorder.print_and_save_npy(env_name, cwd) if is_solved: agent.save_or_load_model(cwd, is_save=True) draw_plot_with_npy(cwd, train_time) """utils""" def get_env_info(env, is_print): # 2020-06-06 state_dim = env.observation_space.shape[0] try: is_discrete = isinstance(env.action_space, gym.spaces.Discrete) if is_discrete: # discrete action_dim = env.action_space.n action_max = int(1) elif isinstance(env.action_space, gym.spaces.Box): # make sure it is continuous action space action_dim = env.action_space.shape[0] action_max = float(env.action_space.high[0]) else: raise AttributeError except AttributeError: print("| Could you assign these value manually? \n" "| I need: state_dim, action_dim, action_max, target_reward, is_discrete") raise AttributeError target_reward = env.spec.reward_threshold if target_reward is None: print("| Could you assign these value manually? \n" "| I need: target_reward") raise ValueError if is_print: print("| env_name: {}, action space: {}".format(repr(env)[10:-1], 'Discrete' if is_discrete else 'Continuous')) print("| state_dim: {}, action_dim: {}, action_max: {}, target_reward: {}".format( state_dim, action_dim, action_max, target_reward)) return state_dim, action_dim, action_max, target_reward, is_discrete def draw_plot_with_npy(mod_dir, train_time): # 2020-04-40 record_epoch = np.load('%s/record_epoch.npy' % mod_dir) # , allow_pickle=True) # record_epoch.append((epoch_reward, actor_loss, critic_loss, iter_num)) record_eval = np.load('%s/record_eval.npy' % mod_dir) # , allow_pickle=True) # record_eval.append((epoch, eval_reward, eval_std)) # print(';record_epoch:', record_epoch.shape) # print(';record_eval:', record_eval.shape) # print(record_epoch) # # print(record_eval) # exit() if len(record_eval.shape) == 1: record_eval = np.array([[0., 0., 0.]]) train_time = int(train_time) iter_num = int(sum(record_epoch[:, -1])) epoch_num = int(record_eval[-1, 0]) save_title = "plot_{:04}E_{}T_{}s".format(epoch_num, iter_num, train_time) save_path = "{}/{}.png".format(mod_dir, save_title) """plot""" import matplotlib as mpl # draw figure in Terminal mpl.use('Agg') import matplotlib.pyplot as plt # plt.style.use('ggplot') fig, axs = plt.subplots(2) plt.title(save_title, y=2.3) ax13 = axs[0].twinx() ax13.fill_between(np.arange(record_epoch.shape[0]), record_epoch[:, 3], facecolor='grey', alpha=0.1, ) ax11 = axs[0] ax11_color = 'royalblue' ax11_label = 'Epo R' ax11.set_ylabel(ylabel=ax11_label, color=ax11_color) ax11.tick_params(axis='y', labelcolor=ax11_color) ax11.plot(record_epoch[:, 0], label=ax11_label, color=ax11_color) ax12 = axs[0] ax12_color = 'lightcoral' ax12_label = 'Epoch R' ax12.set_ylabel(ylabel=ax12_label, color=ax12_color) ax12.tick_params(axis='y', labelcolor=ax12_color) xs = record_eval[:, 0] r_avg = record_eval[:, 1] r_std = record_eval[:, 2] ax12.plot(xs, r_avg, label=ax12_label, color=ax12_color) ax12.fill_between(xs, r_avg - r_std, r_avg + r_std, facecolor=ax12_color, alpha=0.3, ) ax21 = axs[1] ax21_color = 'darkcyan' ax21_label = '- loss A' ax21.set_ylabel(ax21_label, color=ax21_color) ax21.plot(-record_epoch[:, 1], label=ax21_label, color=ax21_color) # negative loss A ax21.tick_params(axis='y', labelcolor=ax21_color) ax22 = axs[1].twinx() ax22_color = 'darkcyan' ax22_label = 'loss C' ax22.set_ylabel(ax22_label, color=ax22_color) ax22.fill_between(np.arange(record_epoch.shape[0]), record_epoch[:, 2], facecolor=ax22_color, alpha=0.2, ) ax22.tick_params(axis='y', labelcolor=ax22_color) plt.savefig(save_path) # plt.show() # plt.ion() # plt.pause(4) def whether_remove_history(cwd, is_remove=None): # 2020-03-04 import shutil if is_remove is None: is_remove = bool(input("PRESS 'y' to REMOVE: {}? ".format(cwd)) == 'y') if is_remove: shutil.rmtree(cwd, ignore_errors=True) print("| Remove") os.makedirs(cwd, exist_ok=True) # shutil.copy(sys.argv[-1], "{}/AgentRun-py-backup".format(cwd)) # copy *.py to cwd # shutil.copy('AgentZoo.py', "{}/AgentZoo-py-backup".format(cwd)) # copy *.py to cwd # shutil.copy('AgentNet.py', "{}/AgentNetwork-py-backup".format(cwd)) # copy *.py to cwd del shutil """demo""" def run__demo(gpu_id, cwd='AC_BasicAC'): from AgentZoo import AgentSNAC as AgentClass args = Arguments(AgentClass) args.gpu_id = gpu_id args.env_name = "LunarLanderContinuous-v2" args.cwd = './{}/LL_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) args.env_name = "BipedalWalker-v3" args.cwd = './{}/BW_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) def run__zoo(gpu_id, cwd='AC_Zoo'): import AgentZoo as Zoo class_agent = Zoo.AgentDeepSAC assert class_agent in { Zoo.AgentDDPG, Zoo.AgentTD3, Zoo.ActorSAC, Zoo.AgentDeepSAC, Zoo.AgentBasicAC, Zoo.AgentSNAC, Zoo.AgentInterAC, Zoo.AgentInterSAC, } # you can't run PPO here. goto run__ppo(). PPO need its hyper-parameters args = Arguments(class_agent) args.gpu_id = gpu_id args.env_name = "LunarLanderContinuous-v2" args.cwd = './{}/LL_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) args.env_name = "BipedalWalker-v3" args.cwd = './{}/BW_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) # args.env_name = "BipedalWalkerHardcore-v3" # args.cwd = './{}/BWHC_{}'.format(cwd, gpu_id) # args.net_dim = int(2 ** 8.5) # args.max_memo = int(2 ** 20) # args.batch_size = int(2 ** 9) # args.max_epoch = 2 ** 14 # args.reward_scale = int(2 ** 6.5) # args.is_remove = None # args.init_for_training() # while not train_agent(**vars(args)): # args.random_seed += 42 # import pybullet_envs # for python-bullet-gym # dir(pybullet_envs) # args.env_name = "MinitaurBulletEnv-v0" # args.cwd = './{}/Minitaur_{}'.format(cwd, args.gpu_id) # args.max_epoch = 2 ** 13 # args.max_memo = 2 ** 20 # args.net_dim = 2 ** 9 # args.max_step = 2 ** 12 # args.batch_size = 2 ** 8 # args.reward_scale = 2 ** 3 # args.is_remove = True # args.eva_size = 2 ** 5 # for Recorder # args.show_gap = 2 ** 8 # for Recorder # args.init_for_training() # while not train_agent(**vars(args)): # args.random_seed += 42 # import pybullet_envs # for python-bullet-gym # dir(pybullet_envs) # args.env_name = "AntBulletEnv-v0" # args.cwd = './{}/Ant_{}'.format(cwd, args.gpu_id) # args.max_epoch = 2 ** 13 # args.max_memo = 2 ** 20 # args.max_step = 2 ** 10 # args.net_dim = 2 ** 8 # args.batch_size = 2 ** 8 # args.reward_scale = 2 ** -3 # args.is_remove = True # args.eva_size = 2 ** 5 # for Recorder # args.show_gap = 2 ** 8 # for Recorder # args.init_for_training() # while not train_agent(**vars(args)): # args.random_seed += 42 def run__ppo(gpu_id, cwd): import AgentZoo as Zoo class_agent = Zoo.AgentGAE assert class_agent in {Zoo.AgentPPO, Zoo.AgentGAE} args = Arguments(class_agent) args.gpu_id = gpu_id args.max_memo = 2 ** 12 args.batch_size = 2 ** 9 args.repeat_times = 2 ** 4 args.net_dim = 2 ** 8 args.gamma = 0.99 args.env_name = "LunarLanderContinuous-v2" args.cwd = './{}/LL_{}'.format(cwd, gpu_id) args.init_for_training() while not train_agent__on_policy(**vars(args)): args.random_seed += 42 args.env_name = "BipedalWalker-v3" args.cwd = './{}/BW_{}'.format(cwd, gpu_id) args.init_for_training() while not train_agent__on_policy(**vars(args)): args.random_seed += 42 def run__dqn(gpu_id, cwd='RL_DQN'): from AgentZoo import AgentDQN # from AgentZoo import AgentNoisyDQN # from AgentZoo import AgentDoubleDQN args = Arguments(AgentDQN) args.gpu_id = gpu_id args.show_gap = 2 ** 5 args.env_name = "CartPole-v0" args.cwd = '{}/{}'.format(cwd, args.env_name) args.init_for_training() train_agent_discrete(**vars(args)) args.env_name = "LunarLander-v2" args.cwd = '{}/{}'.format(cwd, args.env_name) args.init_for_training() train_agent_discrete(**vars(args)) def run__multi_process(target_func, gpu_tuple=(0, 1), cwd='RL_MP'): os.makedirs(cwd, exist_ok=True) # all the files save in here '''run in multiprocessing''' import multiprocessing as mp processes = [mp.Process(target=target_func, args=(gpu_id, cwd)) for gpu_id in gpu_tuple] [process.start() for process in processes] [process.join() for process in processes] def process__buffer(q_aggr, qs_dist, args, **_kwargs): max_memo = args.max_memo env_name = args.env_name max_step = args.max_step batch_size = args.batch_size repeat_times = 2 # reward_scale = args.reward_scale # gamma = args.gamma '''init''' env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=False) buffer = BufferArray(max_memo, state_dim, action_dim) # experiment replay buffer workers_num = len(qs_dist) '''loop''' is_training = True while is_training: for i in range(workers_num): memo_array, is_solved = q_aggr.get() buffer.extend_memo(memo_array) if is_solved: is_training = False buffer.init_before_sample() for i in range(max_step * repeat_times): # batch_arrays = buffer.random_sample(batch_size, device=None) # faster but worse for q_dist in qs_dist: batch_arrays = buffer.random_sample(batch_size, device=None) # slower but better q_dist.put(batch_arrays) print('|| Exit: process__buffer') def process__workers(gpu_id, root_cwd, q_aggr, q_dist, args, **_kwargs): class_agent = args.class_agent env_name = args.env_name cwd = args.cwd net_dim = args.net_dim max_step = args.max_step # max_memo = args.max_memo max_epoch = args.max_epoch batch_size = args.batch_size * 1.5 gamma = args.gamma update_gap = args.update_gap reward_scale = args.reward_scale cwd = '{}/{}_{}'.format(root_cwd, cwd, gpu_id) os.makedirs(cwd, exist_ok=True) os.environ['CUDA_VISIBLE_DEVICES'] = str(gpu_id) random_seed = 42 + gpu_id np.random.seed(random_seed) torch.manual_seed(random_seed) torch.set_default_dtype(torch.float32) torch.set_num_threads(8) env = gym.make(env_name) is_solved = False class BufferArrayMP(BufferArray): def init_before_sample(self): q_aggr.put((self.memories, is_solved)) # self.now_len = self.max_len if self.is_full else self.next_idx def random_sample(self, _batch_size, device=None): batch_arrays = q_dist.get() '''convert array into torch.tensor''' tensors = [torch.tensor(ary, device=device) for ary in batch_arrays] return tensors '''init''' state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=True) agent = class_agent(env, state_dim, action_dim, net_dim) # training agent buffer = BufferArrayMP(max_step, state_dim, action_dim) # experiment replay buffer recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) '''loop''' # with torch.no_grad(): # update replay buffer # # rewards, steps = agent.update_buffer( # # env, buffer, max_step, max_action, reward_scale, gamma) # rewards, steps = initial_exploration( # env, buffer, max_step, max_action, reward_scale, gamma, action_dim) # recorder.show_reward(rewards, steps, 0, 0) try: for epoch in range(max_epoch): '''update replay buffer by interact with environment''' with torch.no_grad(): # for saving the GPU buffer rewards, steps = agent.update_buffer(env, buffer, max_step, max_action, reward_scale, gamma) '''update network parameters by random sampling buffer for stochastic gradient descent''' loss_a, loss_c = agent.update_parameters(buffer, max_step, batch_size, update_gap) '''show/check the reward, save the max reward actor''' with torch.no_grad(): # for saving the GPU buffer '''NOTICE! Recorder saves the agent with max reward automatically. ''' recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("raise KeyboardInterrupt while training.") # except AssertionError: # for BipedWalker BUG 2020-03-03 # print("AssertionError: OpenAI gym r.LengthSquared() > 0.0f ??? Please run again.") # return False train_time = recorder.print_and_save_npy(env_name, cwd) # agent.save_or_load_model(cwd, is_save=True) # save max reward agent in Recorder # buffer.save_or_load_memo(cwd, is_save=True) draw_plot_with_npy(cwd, train_time) return True def run__multi_workers(gpu_tuple=(0, 1), root_cwd='RL_MP'): print('GPU: {} | CWD: {}'.format(gpu_tuple, root_cwd)) whether_remove_history(root_cwd, is_remove=True) from AgentZoo import AgentSAC args = Arguments(AgentSAC) args.env_name = "BipedalWalker-v3" # args.env_name = "LunarLanderContinuous-v2" args.show_gap = 2 ** 8 # for Recorder '''run in multiprocessing''' import multiprocessing as mp workers_num = len(gpu_tuple) queue_aggr = mp.Queue(maxsize=workers_num) # queue of aggregation queues_dist = [mp.Queue(maxsize=args.max_step) for _ in range(workers_num)] # queue of distribution processes = [mp.Process(target=process__buffer, args=(queue_aggr, queues_dist, args))] processes.extend([mp.Process(target=process__workers, args=(gpu_id, root_cwd, queue_aggr, queue_dist, args)) for gpu_id, queue_dist in zip(gpu_tuple, queues_dist)]) [process.start() for process in processes] # [process.join() for process in processes] [process.close() for process in processes] if __name__ == '__main__': # run__demo(gpu_id=0, cwd='AC_BasicAC') run__zoo(gpu_id=0, cwd='AC_SAC') # run__ppo(gpu_id=1, cwd='AC_PPO') # run__multi_process(run__zoo, gpu_tuple=(0, 1, 2, 3), cwd='AC_ZooMP') # run__multi_process(run__ppo, gpu_tuple=(2, 3), cwd='AC_PPO') # run__multi_workers(gpu_tuple=(2, 3), root_cwd='AC_SAC_MP') # '''Discrete action space''' # run__dqn(gpu_id=sys.argv[-1][-4], cwd='RL_DQN') # '''multi worker''' # run__multi_workers(gpu_tuple=(2, 3), root_cwd='AC_SAC_MP') print('Finish:', sys.argv[-1])
38.074603
119
0.6614
import os import sys import gym import torch import numpy as np from AgentZoo import Recorder from AgentZoo import BufferArray, initial_exploration class Arguments: def __init__(self, class_agent): self.class_agent = class_agent self.net_dim = 2 ** 7 self.max_step = 2 ** 10 self.max_memo = 2 ** 17 self.max_epoch = 2 ** 10 self.batch_size = 2 ** 7 self.repeat_times = 1 self.reward_scale = 2 ** 0 self.gamma = 0.99 self.gpu_id = 0 self.random_seed = 19430 self.is_remove = True self.env_name = "LunarLanderContinuous-v2" self.cwd = 'AC_Methods_LL' self.show_gap = 2 ** 7 def init_for_training(self): print('GPU: {} | CWD: {}'.format(self.gpu_id, self.cwd)) whether_remove_history(self.cwd, self.is_remove) os.environ['CUDA_VISIBLE_DEVICES'] = str(self.gpu_id) lf.random_seed) torch.manual_seed(self.random_seed) torch.set_default_dtype(torch.float32) torch.set_num_threads(8) def train_agent__off_policy( class_agent, net_dim, batch_size, repeat_times, gamma, reward_scale, cwd, env_name, max_step, max_memo, max_epoch, **_kwargs): env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=False) assert not is_discrete agent = class_agent(state_dim, action_dim, net_dim) agent.state = env.reset() buffer = BufferArray(max_memo, state_dim, action_dim) recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) with torch.no_grad(): rewards, steps = initial_exploration(env, buffer, max_step, max_action, reward_scale, gamma, action_dim) recorder.show_reward(rewards, steps, loss_a=0, loss_c=0) try: for epoch in range(max_epoch): with torch.no_grad(): rewards, steps = agent.update_buffer( env, buffer, max_step, max_action, reward_scale, gamma) buffer.init_before_sample() loss_a, loss_c = agent.update_parameters( buffer, max_step, batch_size, repeat_times) with torch.no_grad(): recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("| raise KeyboardInterrupt and break training loop") .print_and_save_npy(env_name, cwd) if is_solved: agent.save_or_load_model(cwd, is_save=True) draw_plot_with_npy(cwd, train_time) def train_agent__on_policy( class_agent, net_dim, batch_size, repeat_times, gamma, reward_scale, cwd, env_name, max_step, max_memo, max_epoch, **_kwargs): env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=True) agent = class_agent(state_dim, action_dim, net_dim) agent.save_or_load_model(cwd, is_save=False) recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) try: for epoch in range(max_epoch): with torch.no_grad(): rewards, steps, buffer = agent.update_buffer_online( env, max_step, max_memo, max_action, reward_scale, gamma) loss_a, loss_c = agent.update_parameters_online( buffer, batch_size, repeat_times) with torch.no_grad(): recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("raise KeyboardInterrupt while training.") except AssertionError: print("AssertionError: OpenAI gym r.LengthSquared() > 0.0f ??? Please run again.") return False train_time = recorder.print_and_save_npy(env_name, cwd) draw_plot_with_npy(cwd, train_time) return True def train_agent_discrete( class_agent, net_dim, batch_size, repeat_times, gamma, reward_scale, cwd, env_name, max_step, max_memo, max_epoch, **_kwargs): env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=True) assert is_discrete agent = class_agent(state_dim, action_dim, net_dim) agent.state = env.reset() buffer = BufferArray(max_memo, state_dim, action_dim=1) recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) with torch.no_grad(): rewards, steps = initial_exploration( env, buffer, max_step, max_action, reward_scale, gamma, action_dim) recorder.show_reward(rewards, steps, loss_a=0, loss_c=0) try: for epoch in range(max_epoch): with torch.no_grad(): rewards, steps = agent.update_buffer( env, buffer, max_step, max_action, reward_scale, gamma) buffer.init_before_sample() loss_a, loss_c = agent.update_parameters(buffer, max_step, batch_size, repeat_times) with torch.no_grad(): recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("| raise KeyboardInterrupt and break training loop") .print_and_save_npy(env_name, cwd) if is_solved: agent.save_or_load_model(cwd, is_save=True) draw_plot_with_npy(cwd, train_time) def get_env_info(env, is_print): state_dim = env.observation_space.shape[0] try: is_discrete = isinstance(env.action_space, gym.spaces.Discrete) if is_discrete: action_dim = env.action_space.n action_max = int(1) elif isinstance(env.action_space, gym.spaces.Box): action_dim = env.action_space.shape[0] action_max = float(env.action_space.high[0]) else: raise AttributeError except AttributeError: print("| Could you assign these value manually? \n" "| I need: state_dim, action_dim, action_max, target_reward, is_discrete") raise AttributeError target_reward = env.spec.reward_threshold if target_reward is None: print("| Could you assign these value manually? \n" "| I need: target_reward") raise ValueError if is_print: print("| env_name: {}, action space: {}".format(repr(env)[10:-1], 'Discrete' if is_discrete else 'Continuous')) print("| state_dim: {}, action_dim: {}, action_max: {}, target_reward: {}".format( state_dim, action_dim, action_max, target_reward)) return state_dim, action_dim, action_max, target_reward, is_discrete def draw_plot_with_npy(mod_dir, train_time): record_epoch = np.load('%s/record_epoch.npy' % mod_dir) record_eval = np.load('%s/record_eval.npy' % mod_dir) cord_eval.shape) == 1: record_eval = np.array([[0., 0., 0.]]) train_time = int(train_time) iter_num = int(sum(record_epoch[:, -1])) epoch_num = int(record_eval[-1, 0]) save_title = "plot_{:04}E_{}T_{}s".format(epoch_num, iter_num, train_time) save_path = "{}/{}.png".format(mod_dir, save_title) import matplotlib as mpl mpl.use('Agg') import matplotlib.pyplot as plt fig, axs = plt.subplots(2) plt.title(save_title, y=2.3) ax13 = axs[0].twinx() ax13.fill_between(np.arange(record_epoch.shape[0]), record_epoch[:, 3], facecolor='grey', alpha=0.1, ) ax11 = axs[0] ax11_color = 'royalblue' ax11_label = 'Epo R' ax11.set_ylabel(ylabel=ax11_label, color=ax11_color) ax11.tick_params(axis='y', labelcolor=ax11_color) ax11.plot(record_epoch[:, 0], label=ax11_label, color=ax11_color) ax12 = axs[0] ax12_color = 'lightcoral' ax12_label = 'Epoch R' ax12.set_ylabel(ylabel=ax12_label, color=ax12_color) ax12.tick_params(axis='y', labelcolor=ax12_color) xs = record_eval[:, 0] r_avg = record_eval[:, 1] r_std = record_eval[:, 2] ax12.plot(xs, r_avg, label=ax12_label, color=ax12_color) ax12.fill_between(xs, r_avg - r_std, r_avg + r_std, facecolor=ax12_color, alpha=0.3, ) ax21 = axs[1] ax21_color = 'darkcyan' ax21_label = '- loss A' ax21.set_ylabel(ax21_label, color=ax21_color) ax21.plot(-record_epoch[:, 1], label=ax21_label, color=ax21_color) ax21.tick_params(axis='y', labelcolor=ax21_color) ax22 = axs[1].twinx() ax22_color = 'darkcyan' ax22_label = 'loss C' ax22.set_ylabel(ax22_label, color=ax22_color) ax22.fill_between(np.arange(record_epoch.shape[0]), record_epoch[:, 2], facecolor=ax22_color, alpha=0.2, ) ax22.tick_params(axis='y', labelcolor=ax22_color) plt.savefig(save_path) def whether_remove_history(cwd, is_remove=None): import shutil if is_remove is None: is_remove = bool(input("PRESS 'y' to REMOVE: {}? ".format(cwd)) == 'y') if is_remove: shutil.rmtree(cwd, ignore_errors=True) print("| Remove") os.makedirs(cwd, exist_ok=True) d='AC_BasicAC'): from AgentZoo import AgentSNAC as AgentClass args = Arguments(AgentClass) args.gpu_id = gpu_id args.env_name = "LunarLanderContinuous-v2" args.cwd = './{}/LL_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) args.env_name = "BipedalWalker-v3" args.cwd = './{}/BW_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) def run__zoo(gpu_id, cwd='AC_Zoo'): import AgentZoo as Zoo class_agent = Zoo.AgentDeepSAC assert class_agent in { Zoo.AgentDDPG, Zoo.AgentTD3, Zoo.ActorSAC, Zoo.AgentDeepSAC, Zoo.AgentBasicAC, Zoo.AgentSNAC, Zoo.AgentInterAC, Zoo.AgentInterSAC, } args = Arguments(class_agent) args.gpu_id = gpu_id args.env_name = "LunarLanderContinuous-v2" args.cwd = './{}/LL_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) args.env_name = "BipedalWalker-v3" args.cwd = './{}/BW_{}'.format(cwd, gpu_id) args.init_for_training() train_agent__off_policy(**vars(args)) # args.env_name = "BipedalWalkerHardcore-v3" # args.cwd = './{}/BWHC_{}'.format(cwd, gpu_id) # args.net_dim = int(2 ** 8.5) # args.max_memo = int(2 ** 20) # args.batch_size = int(2 ** 9) # args.max_epoch = 2 ** 14 # args.reward_scale = int(2 ** 6.5) # args.is_remove = None # args.init_for_training() # while not train_agent(**vars(args)): # args.random_seed += 42 # import pybullet_envs # for python-bullet-gym # dir(pybullet_envs) # args.env_name = "MinitaurBulletEnv-v0" # args.cwd = './{}/Minitaur_{}'.format(cwd, args.gpu_id) # args.max_epoch = 2 ** 13 # args.max_memo = 2 ** 20 # args.net_dim = 2 ** 9 # args.max_step = 2 ** 12 # args.batch_size = 2 ** 8 # args.reward_scale = 2 ** 3 # args.is_remove = True # args.eva_size = 2 ** 5 # for Recorder # args.show_gap = 2 ** 8 # for Recorder # args.init_for_training() # while not train_agent(**vars(args)): # args.random_seed += 42 # import pybullet_envs # for python-bullet-gym # dir(pybullet_envs) # args.env_name = "AntBulletEnv-v0" # args.cwd = './{}/Ant_{}'.format(cwd, args.gpu_id) # args.max_epoch = 2 ** 13 # args.max_memo = 2 ** 20 # args.max_step = 2 ** 10 # args.net_dim = 2 ** 8 # args.batch_size = 2 ** 8 # args.reward_scale = 2 ** -3 # args.is_remove = True # args.eva_size = 2 ** 5 # for Recorder # args.show_gap = 2 ** 8 # for Recorder # args.init_for_training() # while not train_agent(**vars(args)): # args.random_seed += 42 def run__ppo(gpu_id, cwd): import AgentZoo as Zoo class_agent = Zoo.AgentGAE assert class_agent in {Zoo.AgentPPO, Zoo.AgentGAE} args = Arguments(class_agent) args.gpu_id = gpu_id args.max_memo = 2 ** 12 args.batch_size = 2 ** 9 args.repeat_times = 2 ** 4 args.net_dim = 2 ** 8 args.gamma = 0.99 args.env_name = "LunarLanderContinuous-v2" args.cwd = './{}/LL_{}'.format(cwd, gpu_id) args.init_for_training() while not train_agent__on_policy(**vars(args)): args.random_seed += 42 args.env_name = "BipedalWalker-v3" args.cwd = './{}/BW_{}'.format(cwd, gpu_id) args.init_for_training() while not train_agent__on_policy(**vars(args)): args.random_seed += 42 def run__dqn(gpu_id, cwd='RL_DQN'): from AgentZoo import AgentDQN # from AgentZoo import AgentNoisyDQN # from AgentZoo import AgentDoubleDQN args = Arguments(AgentDQN) args.gpu_id = gpu_id args.show_gap = 2 ** 5 args.env_name = "CartPole-v0" args.cwd = '{}/{}'.format(cwd, args.env_name) args.init_for_training() train_agent_discrete(**vars(args)) args.env_name = "LunarLander-v2" args.cwd = '{}/{}'.format(cwd, args.env_name) args.init_for_training() train_agent_discrete(**vars(args)) def run__multi_process(target_func, gpu_tuple=(0, 1), cwd='RL_MP'): os.makedirs(cwd, exist_ok=True) # all the files save in here import multiprocessing as mp processes = [mp.Process(target=target_func, args=(gpu_id, cwd)) for gpu_id in gpu_tuple] [process.start() for process in processes] [process.join() for process in processes] def process__buffer(q_aggr, qs_dist, args, **_kwargs): max_memo = args.max_memo env_name = args.env_name max_step = args.max_step batch_size = args.batch_size repeat_times = 2 # reward_scale = args.reward_scale # gamma = args.gamma env = gym.make(env_name) state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=False) buffer = BufferArray(max_memo, state_dim, action_dim) # experiment replay buffer workers_num = len(qs_dist) is_training = True while is_training: for i in range(workers_num): memo_array, is_solved = q_aggr.get() buffer.extend_memo(memo_array) if is_solved: is_training = False buffer.init_before_sample() for i in range(max_step * repeat_times): # batch_arrays = buffer.random_sample(batch_size, device=None) # faster but worse for q_dist in qs_dist: batch_arrays = buffer.random_sample(batch_size, device=None) # slower but better q_dist.put(batch_arrays) print('|| Exit: process__buffer') def process__workers(gpu_id, root_cwd, q_aggr, q_dist, args, **_kwargs): class_agent = args.class_agent env_name = args.env_name cwd = args.cwd net_dim = args.net_dim max_step = args.max_step # max_memo = args.max_memo max_epoch = args.max_epoch batch_size = args.batch_size * 1.5 gamma = args.gamma update_gap = args.update_gap reward_scale = args.reward_scale cwd = '{}/{}_{}'.format(root_cwd, cwd, gpu_id) os.makedirs(cwd, exist_ok=True) os.environ['CUDA_VISIBLE_DEVICES'] = str(gpu_id) random_seed = 42 + gpu_id np.random.seed(random_seed) torch.manual_seed(random_seed) torch.set_default_dtype(torch.float32) torch.set_num_threads(8) env = gym.make(env_name) is_solved = False class BufferArrayMP(BufferArray): def init_before_sample(self): q_aggr.put((self.memories, is_solved)) # self.now_len = self.max_len if self.is_full else self.next_idx def random_sample(self, _batch_size, device=None): batch_arrays = q_dist.get() tensors = [torch.tensor(ary, device=device) for ary in batch_arrays] return tensors state_dim, action_dim, max_action, target_reward, is_discrete = get_env_info(env, is_print=True) agent = class_agent(env, state_dim, action_dim, net_dim) # training agent buffer = BufferArrayMP(max_step, state_dim, action_dim) # experiment replay buffer recorder = Recorder(agent, max_step, max_action, target_reward, env_name, **_kwargs) # with torch.no_grad(): # update replay buffer # # rewards, steps = agent.update_buffer( # # env, buffer, max_step, max_action, reward_scale, gamma) # rewards, steps = initial_exploration( # env, buffer, max_step, max_action, reward_scale, gamma, action_dim) # recorder.show_reward(rewards, steps, 0, 0) try: for epoch in range(max_epoch): with torch.no_grad(): # for saving the GPU buffer rewards, steps = agent.update_buffer(env, buffer, max_step, max_action, reward_scale, gamma) loss_a, loss_c = agent.update_parameters(buffer, max_step, batch_size, update_gap) with torch.no_grad(): # for saving the GPU buffer recorder.show_reward(rewards, steps, loss_a, loss_c) is_solved = recorder.check_reward(cwd, loss_a, loss_c) if is_solved: break except KeyboardInterrupt: print("raise KeyboardInterrupt while training.") # except AssertionError: # for BipedWalker BUG 2020-03-03 # print("AssertionError: OpenAI gym r.LengthSquared() > 0.0f ??? Please run again.") # return False train_time = recorder.print_and_save_npy(env_name, cwd) # agent.save_or_load_model(cwd, is_save=True) # save max reward agent in Recorder # buffer.save_or_load_memo(cwd, is_save=True) draw_plot_with_npy(cwd, train_time) return True def run__multi_workers(gpu_tuple=(0, 1), root_cwd='RL_MP'): print('GPU: {} | CWD: {}'.format(gpu_tuple, root_cwd)) whether_remove_history(root_cwd, is_remove=True) from AgentZoo import AgentSAC args = Arguments(AgentSAC) args.env_name = "BipedalWalker-v3" # args.env_name = "LunarLanderContinuous-v2" args.show_gap = 2 ** 8 # for Recorder import multiprocessing as mp workers_num = len(gpu_tuple) queue_aggr = mp.Queue(maxsize=workers_num) # queue of aggregation queues_dist = [mp.Queue(maxsize=args.max_step) for _ in range(workers_num)] # queue of distribution processes = [mp.Process(target=process__buffer, args=(queue_aggr, queues_dist, args))] processes.extend([mp.Process(target=process__workers, args=(gpu_id, root_cwd, queue_aggr, queue_dist, args)) for gpu_id, queue_dist in zip(gpu_tuple, queues_dist)]) [process.start() for process in processes] # [process.join() for process in processes] [process.close() for process in processes] if __name__ == '__main__': # run__demo(gpu_id=0, cwd='AC_BasicAC') run__zoo(gpu_id=0, cwd='AC_SAC') # run__ppo(gpu_id=1, cwd='AC_PPO') # run__multi_process(run__zoo, gpu_tuple=(0, 1, 2, 3), cwd='AC_ZooMP') # run__multi_process(run__ppo, gpu_tuple=(2, 3), cwd='AC_PPO') # run__multi_workers(gpu_tuple=(2, 3), root_cwd='AC_SAC_MP') # '''Discrete action space''' # run__dqn(gpu_id=sys.argv[-1][-4], cwd='RL_DQN') # '''multi worker''' # run__multi_workers(gpu_tuple=(2, 3), root_cwd='AC_SAC_MP') print('Finish:', sys.argv[-1])
true
true
1c483e974ca788f3d309d21acc49121f28db829a
911
py
Python
searches/double_linear_search_recursion.py
jenia90/Python
696fb4a681ad9e4d84e0d2b894daf449a3e30b24
[ "MIT" ]
145,614
2016-07-21T05:40:05.000Z
2022-03-31T22:17:22.000Z
searches/double_linear_search_recursion.py
Agha-Muqarib/Python
04f156a8973d6156a4357e0717d9eb0aa264d086
[ "MIT" ]
3,987
2016-07-28T17:31:25.000Z
2022-03-30T23:07:46.000Z
searches/double_linear_search_recursion.py
Agha-Muqarib/Python
04f156a8973d6156a4357e0717d9eb0aa264d086
[ "MIT" ]
40,014
2016-07-26T15:14:41.000Z
2022-03-31T22:23:03.000Z
def search(list_data: list, key: int, left: int = 0, right: int = 0) -> int: """ Iterate through the array to find the index of key using recursion. :param list_data: the list to be searched :param key: the key to be searched :param left: the index of first element :param right: the index of last element :return: the index of key value if found, -1 otherwise. >>> search(list(range(0, 11)), 5) 5 >>> search([1, 2, 4, 5, 3], 4) 2 >>> search([1, 2, 4, 5, 3], 6) -1 >>> search([5], 5) 0 >>> search([], 1) -1 """ right = right or len(list_data) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(list_data, key, left + 1, right - 1) if __name__ == "__main__": import doctest doctest.testmod()
25.305556
76
0.567508
def search(list_data: list, key: int, left: int = 0, right: int = 0) -> int: right = right or len(list_data) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(list_data, key, left + 1, right - 1) if __name__ == "__main__": import doctest doctest.testmod()
true
true
1c483eda5ea59fd0903c853ecf78214873dd9e96
652
py
Python
qiskit/providers/aer/backends/__init__.py
derivation/qiskit-aer
d8d77270c745e4c31129ce7f816a93e1efc2e743
[ "Apache-2.0" ]
null
null
null
qiskit/providers/aer/backends/__init__.py
derivation/qiskit-aer
d8d77270c745e4c31129ce7f816a93e1efc2e743
[ "Apache-2.0" ]
29
2018-12-19T10:11:00.000Z
2018-12-19T10:16:18.000Z
qiskit/providers/aer/backends/__init__.py
atilag/qiskit-aer
d964795b0a24b1d3287ba2ba2dda45d1dfed4a5d
[ "Apache-2.0" ]
null
null
null
# This code is part of Qiskit. # # (C) Copyright IBM 2018, 2019. # # This code is licensed under the Apache License, Version 2.0. You may # obtain a copy of this license in the LICENSE.txt file in the root directory # of this source tree or at http://www.apache.org/licenses/LICENSE-2.0. # # Any modifications or derivative works of this code must retain this # copyright notice, and modified files need to carry a notice indicating # that they have been altered from the originals. """Aer Backends.""" from .qasm_simulator import QasmSimulator from .statevector_simulator import StatevectorSimulator from .unitary_simulator import UnitarySimulator
34.315789
77
0.779141
from .qasm_simulator import QasmSimulator from .statevector_simulator import StatevectorSimulator from .unitary_simulator import UnitarySimulator
true
true
1c483f25e50a0b953ab2d539931be326a66b2eb4
18,592
py
Python
src/pyphocorehelpers/gui/Qt/GlobalConnectionManager.py
CommanderPho/pyPhoCoreHelpers
1872cc9779d3ec936077be1df867fc13bc7f177a
[ "MIT" ]
null
null
null
src/pyphocorehelpers/gui/Qt/GlobalConnectionManager.py
CommanderPho/pyPhoCoreHelpers
1872cc9779d3ec936077be1df867fc13bc7f177a
[ "MIT" ]
null
null
null
src/pyphocorehelpers/gui/Qt/GlobalConnectionManager.py
CommanderPho/pyPhoCoreHelpers
1872cc9779d3ec936077be1df867fc13bc7f177a
[ "MIT" ]
null
null
null
# GlobalConnectionManager from indexed import IndexedOrderedDict from qtpy import QtCore, QtWidgets, QtGui """ Requires https://github.com/jazzycamel/PyQt5Singleton.git pip install PyQt5Singleton """ from PyQt5Singleton import Singleton class GlobalConnectionManager(QtCore.QObject, metaclass=Singleton): """ A singleton owned by the QApplication instance that owns connections between widgets/windows and includes tools for discovering widgets to control/be controlled by. """ _currentInstance = None def __init__(self, owning_application: QtWidgets.QApplication, parent=None, **kwargs): super(GlobalConnectionManager, self).__init__(parent, **kwargs) if owning_application is None or not isinstance(owning_application, QtWidgets.QApplication): # app was never constructed is already deleted or is an # QCoreApplication/QGuiApplication and not a full QApplication raise NotImplementedError # Setup member variables: self._registered_available_drivers = IndexedOrderedDict({}) self._registered_available_drivables = IndexedOrderedDict({}) self._active_connections = IndexedOrderedDict({}) # Setup internal connections: # owning_application.aboutToQuit.connect(self.on_application_quit) @property def registered_available_drivers(self): """ an IndexedOrderedDict of widget/objects that can drive a certain property (currently limited to time or time windows) """ return self._registered_available_drivers @property def registered_available_drivables(self): """ an IndexedOrderedDict of widgets/objects that can be driven by a driver.""" return self._registered_available_drivables @property def active_connections(self): """ an IndexedOrderedDict of widgets/objects that can be driven by a driver.""" return self._active_connections #### ================ Registration Methods: def register_driver(self, driver, driver_identifier=None): """Registers a new driver object/widget """ return GlobalConnectionManager.register_control_object(self._registered_available_drivers, driver, driver_identifier) # return the new identifier def register_drivable(self, drivable, drivable_identifier=None): return GlobalConnectionManager.register_control_object(self._registered_available_drivables, drivable, drivable_identifier) # return the new identifier def unregister_object(self, control_object, debug_print=True): # unregisters object from both drivers and drivables # For Driver list: found_driver_key, found_object = GlobalConnectionManager._unregister_object(self._registered_available_drivers, control_object=control_object) if found_driver_key is not None: print(f'removed object with key {found_driver_key} from drivers list.') # For Drivable List: found_drivable_key, found_object = GlobalConnectionManager._unregister_object(self._registered_available_drivables, control_object=control_object) if found_drivable_key is not None: print(f'removed object with key {found_drivable_key} from drivers list.') return found_driver_key, found_drivable_key def connect_drivable_to_driver(self, drivable, driver, custom_connect_function=None): """ attempts to connect the drivable to the driver. drivable/driver can either be a key for a drivable/driver already registered or the drivable/driver itself. Inputs: custom_connect_function: is an optional Callable that takes the driver, drivable as input and returns a connection. """ # Get key for drivable: if isinstance(drivable, str): drivable_key = drivable drivable = self.registered_available_drivables[drivable_key] else: # already have the object, just find the key: drivable_key = GlobalConnectionManager._try_find_object_key(self.registered_available_drivables, control_object=drivable) # Get Key for driver: if isinstance(driver, str): driver_key = driver driver = self.registered_available_drivers[driver_key] else: # already have the object, just find the key: driver_key = GlobalConnectionManager._try_find_object_key(self.registered_available_drivers, control_object=driver) ## Make sure the connection doesn't already exist: extant_connection = self.active_connections.get(drivable, None) if extant_connection is None: ## Make the connection: if custom_connect_function is not None: # Perform the custom connection function: new_connection_obj = custom_connect_function(driver, drivable) else: # Otherwise perform the default: new_connection_obj = GlobalConnectionManager.connect_additional_controlled_plotter(driver, controlled_plt=drivable) self.active_connections[drivable] = new_connection_obj # add the connection object to the self.active_connections array return self.active_connections[drivable] else: print(f'connection already existed!') return extant_connection ## Make the connection: ## Sync ipspikesDataExplorer to raster window: # extra_interactive_spike_behavior_browser_sync_connection = spike_raster_window.connect_additional_controlled_plotter(controlled_plt=ipspikesDataExplorer) # extra_interactive_spike_behavior_browser_sync_connection = _connect_additional_controlled_plotter(spike_raster_window.spike_raster_plt_2d, ipspikesDataExplorer) def disconnect_drivable(self, drivable): """ disconnects the drivable from any drivers. """ self.unregister_object(drivable) #### ================ Access Methods: def get_available_drivers(self): """ gets a list of the available widgets that could be used to drive a time widget. """ return self.registered_available_drivers def get_available_drivables(self): """ gets a list of the available widgets that could be driven via a time widget. """ return self.registered_available_drivables #### ================ Utility Methods: def _disambiguate_driver_name(self, extant_name): """ attempts to create a unique name for the driver that doesn't already exist in the dict and return it """ return GlobalConnectionManager.disambiguate_registered_name(self._registered_available_drivers, extant_name) def _disambiguate_drivable_name(self, extant_name): """ attempts to create a unique name for the drivable that doesn't already exist in the dict and return it """ return GlobalConnectionManager.disambiguate_registered_name(self._registered_available_drivables, extant_name) #### ================ Slots Methods: # @QtCore.Slot() # def on_application_quit(self): # print(f'GlobalConnectionManager.on_application_quit') # GlobalConnectionManager._currentInstance = None #### ================ Static Methods: @classmethod def disambiguate_registered_name(cls, registraction_dict, extant_name): """ attempts to create a unique name for the driver/drivee that doesn't already exist in the dict and return it """ matching_names_with_prefix = list(filter(lambda x: x.startswith(extant_name), list(registraction_dict.keys()))) itr_index = len(matching_names_with_prefix) # get the next number after the previous matching names to build a string like # "RasterPlot2D_1" proposed_driver_identifier = f'{extant_name}_{itr_index}' # Proposed name shouldn't exist: extant_driver_with_identifier = registraction_dict.get(proposed_driver_identifier, None) assert extant_driver_with_identifier is None, f"Driver with new name {extant_driver_with_identifier} already exists too!" # return the new name return proposed_driver_identifier @classmethod def register_control_object(cls, registraction_dict, control_object, control_identifier=None): """Registers a new driver or driven object/widget Args: control_object (_type_): _description_ control_identifier (_type_, optional): _description_. Defaults to None. Returns: _type_: _description_ """ if control_identifier is None: control_identifier = control_object.windowName # 'Spike3DRasterWindow' try: extant_driver_index = list(registraction_dict.values()).index(control_object) # Driver already exists somewhere in the registered drivers: return registraction_dict.keys()[extant_driver_index] # return its key except ValueError as e: # driver doesn't exist anywhere in the registered drivers: pass extant_driver_with_identifier = registraction_dict.get(control_identifier, None) if extant_driver_with_identifier is not None: # driver already exists with this identifier: # check and see if it's the same object if extant_driver_with_identifier == control_object: # driver with this key already exists, but it's the same driver, so it's just attempting to be re-registered for some reason. No problem. return else: print(f'driver with key {control_identifier} already exists and is a different object. Disambiguating name...') # control_identifier = self.disambiguate_driver_name(control_identifier) control_identifier = GlobalConnectionManager.disambiguate_registered_name(registraction_dict, control_identifier) print(f'\t proposed_driver_name is now {control_identifier}') # now has a unique driver identifier # register the driver provided: registraction_dict[control_identifier] = control_object return control_identifier # return the new identifier @classmethod def _try_find_object_key(cls, registraction_dict, control_object): # tries to find the key of the object in the provided registration_dict found_key = None try: extant_item_index = list(registraction_dict.values()).index(control_object) found_key = registraction_dict.keys()[extant_item_index] return found_key except ValueError as e: pass except KeyError as e: pass return found_key @classmethod def _unregister_object(cls, registraction_dict, control_object): # unregisters object from both drivers and drivables found_key = cls._try_find_object_key(registraction_dict, control_object=control_object) found_object = None if found_key is not None: found_object = registraction_dict.pop(found_key) # pop the key ## TODO: tear down any connections that use it. return found_key, found_object #### ================ Static Methods factored out of SyncedTimelineWindowLink.py on 2022-05-25 @classmethod def connect_additional_controlled_plotter(cls, source_spike_raster_plt, controlled_plt): """ allow the window to control InteractivePlaceCellDataExplorer (ipspikesDataExplorer) objects; source_spike_raster_plt: the spike raster plotter to connect to as the source controlled_plt: should be a InteractivePlaceCellDataExplorer object (ipspikesDataExplorer), but can be any function with a valid update_window_start_end @QtCore.Slot(float, float) slot. Requirements: source_spike_raster_plt: .spikes_window.active_time_window .window_scrolled controlled_plt: .disable_ui_window_updating_controls() .update_window_start_end(float, float) Usage: from pyphoplacecellanalysis.GUI.Qt.SpikeRasterWindows.Spike3DRasterWindowWidget import Spike3DRasterWindowWidget # Build the controlled ipspikesDataExplorer: display_output = dict() pActiveSpikesBehaviorPlotter = None display_output = display_output | curr_active_pipeline.display(DefaultDisplayFunctions._display_3d_interactive_spike_and_behavior_browser, active_config_name, extant_plotter=display_output.get('pActiveSpikesBehaviorPlotter', None)) # Works now! ipspikesDataExplorer = display_output['ipspikesDataExplorer'] display_output['pActiveSpikesBehaviorPlotter'] = display_output.pop('plotter') # rename the key from the generic "plotter" to "pActiveSpikesBehaviorPlotter" to avoid collisions with others pActiveSpikesBehaviorPlotter = display_output['pActiveSpikesBehaviorPlotter'] # Build the contolling raster window: spike_raster_window = Spike3DRasterWindowWidget(curr_spikes_df) # Call this function to connect them: extra_interactive_spike_behavior_browser_sync_connection = connect_additional_controlled_plotter(spike_raster_window.spike_raster_plt_2d, ipspikesDataExplorer) """ # Perform Initial (one-time) update from source -> controlled: controlled_plt.disable_ui_window_updating_controls() # disable the GUI for manual updates. controlled_plt.update_window_start_end(source_spike_raster_plt.spikes_window.active_time_window[0], source_spike_raster_plt.spikes_window.active_time_window[1]) # Connect to update self when video window playback position changes sync_connection = source_spike_raster_plt.window_scrolled.connect(controlled_plt.update_window_start_end) return sync_connection @classmethod def connect_controlled_time_synchornized_plotter(cls, source_spike_raster_plt, controlled_plt): """ source_spike_raster_plt: TimeSynchronizedPlotterBase Identical to the connect_additional_controlled_plotter(...) but uses on_window_changed(...) instead of update_window_start_end(...) """ controlled_plt.on_window_changed(source_spike_raster_plt.spikes_window.active_time_window[0], source_spike_raster_plt.spikes_window.active_time_window[1]) sync_connection = source_spike_raster_plt.window_scrolled.connect(controlled_plt.on_window_changed) # connect the window_scrolled event to the _on_window_updated function return sync_connection # @classmethod # def connect_additional_controlled_spike_raster_plotter(cls, spike_raster_plt_2d, controlled_spike_raster_plt): # """ Connect an additional plotter to a source that's driving the update of the data-window: # Requirements: # source_spike_raster_plt: # .spikes_window.active_time_window # .window_scrolled # controlled_spike_raster_plt: # .spikes_window.update_window_start_end(float, float) # Usage: # spike_raster_plt_3d, spike_raster_plt_2d, spike_3d_to_2d_window_connection = build_spike_3d_raster_with_2d_controls(curr_spikes_df) # spike_raster_plt_3d_vedo = Spike3DRaster_Vedo(curr_spikes_df, window_duration=15.0, window_start_time=30.0, neuron_colors=None, neuron_sort_order=None) # extra_vedo_sync_connection = connect_additional_controlled_spike_raster_plotter(spike_raster_plt_2d, spike_raster_plt_3d_vedo) # """ # controlled_spike_raster_plt.spikes_window.update_window_start_end(spike_raster_plt_2d.spikes_window.active_time_window[0], spike_raster_plt_2d.spikes_window.active_time_window[1]) # # Connect to update self when video window playback position changes # sync_connection = spike_raster_plt_2d.window_scrolled.connect(controlled_spike_raster_plt.spikes_window.update_window_start_end) # return sync_connection ### Usesful Examples: ### Checking if application instance exists yet: # if QtGui.QApplication.instance() is None: # return ### Checking if an object is still alive/extant: # from ...Qt import isQObjectAlive # for k in ViewBox.AllViews: # if isQObjectAlive(k) and getConfigOption('crashWarning'): # sys.stderr.write('Warning: ViewBox should be closed before application exit.\n') # try: # k.destroyed.disconnect() # except RuntimeError: ## signal is already disconnected. # pass # except TypeError: ## view has already been deleted (?) # pass # except AttributeError: # PySide has deleted signal # pass class GlobalConnectionManagerAccessingMixin: """ Implementor owns a connection manager instance which it usually uses to register itself or its children as drivers/drivable Required Properties: ._connection_man """ @property def connection_man(self): """The connection_man property.""" return self._connection_man def GlobalConnectionManagerAccessingMixin_on_init(self, owning_application=None): if owning_application is None: owning_application = QtWidgets.QApplication.instance() # <PyQt5.QtWidgets.QApplication at 0x1d44a4891f0> if owning_application is None: print(f'could not get valid QApplication instance!') raise NotImplementedError # Set self._connection_man: self._connection_man = GlobalConnectionManager(owning_application=owning_application) ######################################################## ## For GlobalConnectionManagerAccessingMixin conformance: ######################################################## # @QtCore.pyqtSlot() def GlobalConnectionManagerAccessingMixin_on_setup(self): """ perfrom registration of drivers/drivables:""" ## TODO: register children pass # @QtCore.pyqtSlot() def GlobalConnectionManagerAccessingMixin_on_destroy(self): """ perfrom teardown/destruction of anything that needs to be manually removed or released """ ## TODO: unregister children pass
49.978495
256
0.701592
from indexed import IndexedOrderedDict from qtpy import QtCore, QtWidgets, QtGui from PyQt5Singleton import Singleton class GlobalConnectionManager(QtCore.QObject, metaclass=Singleton): _currentInstance = None def __init__(self, owning_application: QtWidgets.QApplication, parent=None, **kwargs): super(GlobalConnectionManager, self).__init__(parent, **kwargs) if owning_application is None or not isinstance(owning_application, QtWidgets.QApplication): raise NotImplementedError self._registered_available_drivers = IndexedOrderedDict({}) self._registered_available_drivables = IndexedOrderedDict({}) self._active_connections = IndexedOrderedDict({}) @property def registered_available_drivers(self): return self._registered_available_drivers @property def registered_available_drivables(self): return self._registered_available_drivables @property def active_connections(self): return self._active_connections ect(self._registered_available_drivers, driver, driver_identifier) def register_drivable(self, drivable, drivable_identifier=None): return GlobalConnectionManager.register_control_object(self._registered_available_drivables, drivable, drivable_identifier) def unregister_object(self, control_object, debug_print=True): found_driver_key, found_object = GlobalConnectionManager._unregister_object(self._registered_available_drivers, control_object=control_object) if found_driver_key is not None: print(f'removed object with key {found_driver_key} from drivers list.') found_drivable_key, found_object = GlobalConnectionManager._unregister_object(self._registered_available_drivables, control_object=control_object) if found_drivable_key is not None: print(f'removed object with key {found_drivable_key} from drivers list.') return found_driver_key, found_drivable_key def connect_drivable_to_driver(self, drivable, driver, custom_connect_function=None): if isinstance(drivable, str): drivable_key = drivable drivable = self.registered_available_drivables[drivable_key] else: drivable_key = GlobalConnectionManager._try_find_object_key(self.registered_available_drivables, control_object=drivable) if isinstance(driver, str): driver_key = driver driver = self.registered_available_drivers[driver_key] else: driver_key = GlobalConnectionManager._try_find_object_key(self.registered_available_drivers, control_object=driver) ons.get(drivable, None) if extant_connection is None: ## Make the connection: if custom_connect_function is not None: # Perform the custom connection function: new_connection_obj = custom_connect_function(driver, drivable) else: # Otherwise perform the default: new_connection_obj = GlobalConnectionManager.connect_additional_controlled_plotter(driver, controlled_plt=drivable) self.active_connections[drivable] = new_connection_obj # add the connection object to the self.active_connections array return self.active_connections[drivable] else: print(f'connection already existed!') return extant_connection ## Make the connection: ## Sync ipspikesDataExplorer to raster window: # extra_interactive_spike_behavior_browser_sync_connection = spike_raster_window.connect_additional_controlled_plotter(controlled_plt=ipspikesDataExplorer) # extra_interactive_spike_behavior_browser_sync_connection = _connect_additional_controlled_plotter(spike_raster_window.spike_raster_plt_2d, ipspikesDataExplorer) def disconnect_drivable(self, drivable): self.unregister_object(drivable) #### ================ Access Methods: def get_available_drivers(self): return self.registered_available_drivers def get_available_drivables(self): return self.registered_available_drivables #### ================ Utility Methods: def _disambiguate_driver_name(self, extant_name): return GlobalConnectionManager.disambiguate_registered_name(self._registered_available_drivers, extant_name) def _disambiguate_drivable_name(self, extant_name): return GlobalConnectionManager.disambiguate_registered_name(self._registered_available_drivables, extant_name) #### ================ Slots Methods: # @QtCore.Slot() # def on_application_quit(self): # print(f'GlobalConnectionManager.on_application_quit') # GlobalConnectionManager._currentInstance = None #### ================ Static Methods: @classmethod def disambiguate_registered_name(cls, registraction_dict, extant_name): matching_names_with_prefix = list(filter(lambda x: x.startswith(extant_name), list(registraction_dict.keys()))) itr_index = len(matching_names_with_prefix) # get the next number after the previous matching names to build a string like # "RasterPlot2D_1" proposed_driver_identifier = f'{extant_name}_{itr_index}' # Proposed name shouldn't exist: extant_driver_with_identifier = registraction_dict.get(proposed_driver_identifier, None) assert extant_driver_with_identifier is None, f"Driver with new name {extant_driver_with_identifier} already exists too!" return proposed_driver_identifier @classmethod def register_control_object(cls, registraction_dict, control_object, control_identifier=None): if control_identifier is None: control_identifier = control_object.windowName try: extant_driver_index = list(registraction_dict.values()).index(control_object) return registraction_dict.keys()[extant_driver_index] except ValueError as e: pass extant_driver_with_identifier = registraction_dict.get(control_identifier, None) if extant_driver_with_identifier is not None: # driver already exists with this identifier: # check and see if it's the same object if extant_driver_with_identifier == control_object: return else: print(f'driver with key {control_identifier} already exists and is a different object. Disambiguating name...') control_identifier = GlobalConnectionManager.disambiguate_registered_name(registraction_dict, control_identifier) print(f'\t proposed_driver_name is now {control_identifier}') registraction_dict[control_identifier] = control_object return control_identifier @classmethod def _try_find_object_key(cls, registraction_dict, control_object): found_key = None try: extant_item_index = list(registraction_dict.values()).index(control_object) found_key = registraction_dict.keys()[extant_item_index] return found_key except ValueError as e: pass except KeyError as e: pass return found_key @classmethod def _unregister_object(cls, registraction_dict, control_object): found_key = cls._try_find_object_key(registraction_dict, control_object=control_object) found_object = None if found_key is not None: found_object = registraction_dict.pop(found_key) me_window[0], source_spike_raster_plt.spikes_window.active_time_window[1]) sync_connection = source_spike_raster_plt.window_scrolled.connect(controlled_plt.update_window_start_end) return sync_connection @classmethod def connect_controlled_time_synchornized_plotter(cls, source_spike_raster_plt, controlled_plt): controlled_plt.on_window_changed(source_spike_raster_plt.spikes_window.active_time_window[0], source_spike_raster_plt.spikes_window.active_time_window[1]) sync_connection = source_spike_raster_plt.window_scrolled.connect(controlled_plt.on_window_changed) return sync_connection # Requirements: # source_spike_raster_plt: # .spikes_window.active_time_window # .window_scrolled # controlled_spike_raster_plt: # .spikes_window.update_window_start_end(float, float) # Usage: # spike_raster_plt_3d, spike_raster_plt_2d, spike_3d_to_2d_window_connection = build_spike_3d_raster_with_2d_controls(curr_spikes_df) # spike_raster_plt_3d_vedo = Spike3DRaster_Vedo(curr_spikes_df, window_duration=15.0, window_start_time=30.0, neuron_colors=None, neuron_sort_order=None) # extra_vedo_sync_connection = connect_additional_controlled_spike_raster_plotter(spike_raster_plt_2d, spike_raster_plt_3d_vedo) # """ # controlled_spike_raster_plt.spikes_window.update_window_start_end(spike_raster_plt_2d.spikes_window.active_time_window[0], spike_raster_plt_2d.spikes_window.active_time_window[1]) # # Connect to update self when video window playback position changes # sync_connection = spike_raster_plt_2d.window_scrolled.connect(controlled_spike_raster_plt.spikes_window.update_window_start_end) # return sync_connection ### Usesful Examples: ### Checking if application instance exists yet: # if QtGui.QApplication.instance() is None: # return ### Checking if an object is still alive/extant: # from ...Qt import isQObjectAlive # for k in ViewBox.AllViews: # if isQObjectAlive(k) and getConfigOption('crashWarning'): # sys.stderr.write('Warning: ViewBox should be closed before application exit.\n') # try: # k.destroyed.disconnect() # except RuntimeError: ## signal is already disconnected. # pass # except TypeError: ## view has already been deleted (?) # pass # except AttributeError: # PySide has deleted signal # pass class GlobalConnectionManagerAccessingMixin: @property def connection_man(self): return self._connection_man def GlobalConnectionManagerAccessingMixin_on_init(self, owning_application=None): if owning_application is None: owning_application = QtWidgets.QApplication.instance() # <PyQt5.QtWidgets.QApplication at 0x1d44a4891f0> if owning_application is None: print(f'could not get valid QApplication instance!') raise NotImplementedError # Set self._connection_man: self._connection_man = GlobalConnectionManager(owning_application=owning_application) ######################################################## ## For GlobalConnectionManagerAccessingMixin conformance: ######################################################## # @QtCore.pyqtSlot() def GlobalConnectionManagerAccessingMixin_on_setup(self): ## TODO: register children pass # @QtCore.pyqtSlot() def GlobalConnectionManagerAccessingMixin_on_destroy(self): ## TODO: unregister children pass
true
true
1c48402aad47944422211ce05f453f84068fcaf9
14,633
py
Python
tetris.py
JoeIaquinto/Snakes
e2e9dc94279c1204a4c4e2c5e96713635e3c0a00
[ "Apache-2.0" ]
null
null
null
tetris.py
JoeIaquinto/Snakes
e2e9dc94279c1204a4c4e2c5e96713635e3c0a00
[ "Apache-2.0" ]
null
null
null
tetris.py
JoeIaquinto/Snakes
e2e9dc94279c1204a4c4e2c5e96713635e3c0a00
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- #Dan GORMAN #3/12/13 #Tetris import pygame, sys, time, random, datetime from pygame.locals import * # set up pygame pygame.init() # set up the window WINDOWWIDTH = 340 WINDOWHEIGHT = 2*WINDOWWIDTH columnwidth=WINDOWWIDTH/10 rowheight=WINDOWHEIGHT/20 notplaywindow=WINDOWWIDTH+(5*columnwidth) window = pygame.display.set_mode((notplaywindow, WINDOWHEIGHT), 0, 32) pygame.display.set_caption('Tetris') # set up the play variables BLACK = (0, 0, 0) red = (255, 0, 0) green = (0, 255, 0) blue = (0, 0, 255) purple=(120,0,120) WHITE=(255,255,255) yellow=(255,255,0) cyan=(0, 255, 255) grey=(128, 128, 128) orange=(255, 127, 0) points=0 blocks=[[[-1,0],[1,0],[2,0],cyan],[[-1,0],[-1,-1],[1,0],blue],[[-1,0],[1,0],[1,-1],orange],[[0,-1],[1,-1],[1,0],yellow],[[-1,0],[0,-1],[1,-1],green],[[-1,0],[0,-1],[1,0],purple],[[-1,-1],[0,-1],[1,0],red]] blockonscreen=[] startspotx=columnwidth*4 startspoty=-rowheight newblock=1 fix=0 speed=.02 hi=0 blockmovecount=9 # set up the rect data structure def makenewblock(block): return [pygame.Rect(startspotx,startspoty,columnwidth,rowheight),pygame.Rect(startspotx+(block[0][0]*columnwidth),startspoty+(block[0][1]*rowheight),columnwidth,rowheight),pygame.Rect(startspotx+(block[1][0]*columnwidth),startspoty+(block[1][1]*rowheight),columnwidth,rowheight),pygame.Rect(startspotx+(block[2][0]*columnwidth),startspoty+(block[2][1]*rowheight),columnwidth,rowheight),block[3]] def overlap(one,two): for bk in two: if bk!=curblock: for b in range(len(bk)-1): if one==bk[b]: return True return False def harddrop(): fix=0 while True: for i in range(4): curblock[i].top+=(rowheight) if curblock[i].top==WINDOWHEIGHT: fix=1 if overlap(curblock[i],blockonscreen): fix=1 if fix: for i in range(4): curblock[i].top-=(rowheight) fix=0 lineclear() break def makeghost(): fix=0 ghost=[] for i in range(4): ghost.append(pygame.Rect(curblock[i].left,curblock[i].top,curblock[i].width,curblock[i].height)) ghost.append((40,40,40)) while True: for i in range(4): ghost[i].top+=(rowheight) if ghost[i].top==WINDOWHEIGHT: fix=1 if overlap(ghost[i],blockonscreen): fix=1 if fix: for i in range(4): ghost[i].top-=(rowheight) fix=0 break return ghost def rotfix(drec): if drec=='UP': for i in range(4): curblock[i].top-=rowheight elif drec=='LEFT': for i in range(4): curblock[i].left-=columnwidth elif drec=='RIGHT': for i in range(4): curblock[i].left+=columnwidth bag=[0,1,2,3,4,5,6] def shufflefirst(): newbag=[] newbag2=[] for i in range (7): while True: x=random.choice(bag) if x not in newbag: newbag.append(x) break while True: y=random.choice(bag) if y not in newbag2: newbag2.append(x) break newbag3=newbag+newbag2 return newbag3 def shuffle(): newbag1=pieces[7:] newbag=[] for i in range (7): while True: x=random.choice(bag) if x not in newbag: newbag.append(x) break newbag3=newbag1+newbag return newbag3 def copyrect(rect): return pygame.Rect(rect.left,rect.top,rect.width,rect.height) def rotatecheck(curc, ro, col): new=[] for z in range(len(curc)): if z != 4: new.append(copyrect(curc[z])) else: new.append(curc[4]) rotate(new, ro, col) for i in range(len(new)-1): if new[i].left<0 or new[i].left>=WINDOWWIDTH: return False for b in blockonscreen: if b != curc: for x in range(len(b)-1): if new[i] == b[x]: return False return True def rotate(curr,ro,col): if curr[4]==yellow: rotateo(curr,ro,col) if curr[4]==cyan: rotatei(curr,ro,col) else: rotatea(curr,ro,col) def rotatei(curr,ro,col): if curr[2].centerx>curr[0].centerx and curr[3].centerx>curr[0].centerx: horiz=True curr[0].centerx+=col bla=1 elif curr[2].centerx<curr[0].centerx and curr[3].centerx<curr[0].centerx: horiz=True curr[0].centerx-=col bla=-1 elif curr[2].centery>curr[0].centery and curr[3].centery>curr[0].centery: horiz=False curr[0].centery+=ro bla=-1 elif curr[2].centery<curr[0].centery and curr[3].centery<curr[0].centery: horiz=False curr[0].centery-=ro bla=1 for f in range (1,4): if horiz: curr[f].left=curr[0].left curr[f].top=curr[0].top+(blocks[0][f-1][0]*ro*bla) else: curr[f].top=curr[0].top curr[f].right=curr[0].right+(blocks[0][f-1][0]*col*bla) def rotateo(curr,ro,col): pass def rotatea(curr,ro,col): middlex=curr[0].centerx middley=curr[0].centery for r in range(1,4): if curr[r].centery==middley and curr[r].centerx>middlex: curr[r].centery+=ro curr[r].centerx-=col elif curr[r].centery==middley and curr[r].centerx<middlex: curr[r].centery-=ro curr[r].centerx+=col elif curr[r].centerx==middlex and curr[r].centery>middley: curr[r].centery-=ro curr[r].centerx-=col elif curr[r].centerx==middlex and curr[r].centery<middley: curr[r].centery+=ro curr[r].centerx+=col elif curr[r].centerx>middlex and curr[r].centery>middley: curr[r].centerx-=(2*col) elif curr[r].centery>middley and curr[r].centerx<middlex: curr[r].centery-=(2*ro) elif curr[r].centerx<middlex and curr[r].centery<middley: curr[r].centerx+=(2*col) elif curr[r].centerx>middlex and curr[r].centery<middley: curr[r].centery+=(2*ro) def makenextpiece(block,startx,starty,row,column): return [pygame.Rect(startx,starty,column,row),pygame.Rect((block[0][0]*column)+startx,(block[0][1]*row)+starty,column,row),pygame.Rect((block[1][0]*column)+startx,(block[1][1]*row)+starty,column,row),pygame.Rect((block[2][0]*column)+startx,(block[2][1]*row)+starty,column,row),block[3]] def nextpiece(): nextpieces=[] for b in range (4): if b==0: newrowheight=WINDOWHEIGHT/20 newcolumnwidth=WINDOWWIDTH/10 else: newrowheight=WINDOWHEIGHT/40 newcolumnwidth=WINDOWWIDTH/20 nextpiece=makenextpiece(blocks[pieces[piece+b]],notplaywindow-(3*rowheight),startspoty+(rowheight*((3*b)+3)),newrowheight,newcolumnwidth) if nextpiece[4]!=cyan and newrowheight!=WINDOWWIDTH/20: rotate(nextpiece,newrowheight,newcolumnwidth) else: for i in range(4): nextpiece[i].left-=(newcolumnwidth/2) nextpieces.append(nextpiece) return nextpieces def lineclear(): bla=False numd=[] for d in reversed(range(20)): a=0 for c in blockonscreen: for x in range(len(c)-1): if c[x].top==rowheight*d: a+=1 if a>=10: for c in blockonscreen: b=0 for x in range(len(c)-1): x=x-b if c[x].top==rowheight*d: c.remove(c[x]) b+=1 bla=True for c in blockonscreen: for x in range(len(c)-1): if c[x].top<rowheight*d: c[x].top+=rowheight numd.append('') global points points+=10 lineclear() break for bk in blockonscreen: if len(bk)==1: blockonscreen.remove(bk) return len(numd) held=[] def holdblock(): newblock=0 curb='' if len(held)==0: held.append(curblock[4]) for i in range (7): if blocks[i][3]==held[0]: held.append(i) held.remove(held[0]) blockonscreen.remove(curblock) newblock=1 else: blockonscreen.append(makenewblock(blocks[held[0]])) held.remove(held[0]) blockonscreen.remove(curblock) held.append(curblock[4]) for i in range (7): if blocks[i][3]==held[0]: held.append(i) held.remove(held[0]) curb=blockonscreen[len(blockonscreen)-1] return newblock,curb pieces=shufflefirst() piece=0 print datetime.datetime.now() # run the game loop while True: # check for the QUIT event for event in pygame.event.get(): if event.type == QUIT: pygame.quit() sys.exit() #moves piece if event.type == pygame.KEYDOWN: if event.key == pygame.K_LEFT: for i in range(4): curblock[i].left-=columnwidth if curblock[i].left==-columnwidth: fix=1 if overlap(curblock[i],blockonscreen): fix=1 if fix: for i in range(4): curblock[i].left+=columnwidth fix=0 ghost=makeghost() elif event.key == pygame.K_RIGHT: for i in range(4): curblock[i].left+=columnwidth if curblock[i].left==WINDOWWIDTH: fix=1 if overlap(curblock[i],blockonscreen): fix=1 if fix: for i in range(4): curblock[i].left-=columnwidth fix=0 ghost=makeghost() elif event.key == pygame.K_UP: if rotatecheck(curblock,rowheight,columnwidth): rotate(curblock,rowheight,columnwidth) ghost=makeghost() elif event.key == pygame.K_DOWN: origblockmovecount=blockmovecount blockmovecount=2 elif event.key == pygame.K_SPACE: harddrop() newblock=1 hi=0 ghost=ghost[4:] elif event.key == pygame.K_p: brea=False while True: for event in pygame.event.get(): if event.type == QUIT: pygame.quit() sys.exit() if event.type==pygame.KEYDOWN: brea=True if brea: break elif event.key==pygame.K_RSHIFT or event.key==pygame.K_LSHIFT: n=holdblock() newblock=n[0] curblock=n[1] if curblock!='': ghost=makeghost() if event.type == pygame.KEYUP: if event.key == pygame.K_DOWN: blockmovecount=origblockmovecount # draw the black background onto the surface window.fill(BLACK) if newblock: blockonscreen.append(makenewblock(blocks[pieces[piece]])) newblock=0 curblock=blockonscreen[len(blockonscreen)-1] ghost=makeghost() piece+=1 if piece==7: pieces=shuffle() piece=0 nexpiece=nextpiece() print(points) if hi>=blockmovecount: for i in range(4): curblock[i].top+=(rowheight) if curblock[i].bottom>WINDOWHEIGHT: fix=1 newblock=1 if overlap(curblock[i],blockonscreen): fix=1 newblock=1 hi=0 if fix: for i in range(4): curblock[i].top-=(rowheight) fix=0 lineclear() for i in range (len(ghost)-1): pygame.draw.rect(window,ghost[len(ghost)-1],ghost[i]) for block in blockonscreen: for i in range(len(block)-1): pygame.draw.rect(window,block[len(block)-1],block[i]) for n in nexpiece: for i in range (4): pygame.draw.rect(window,n[4],n[i]) if len(held)!=0: heldblock=makenextpiece(blocks[held[0]],notplaywindow-(columnwidth*3),WINDOWHEIGHT-(rowheight*3),rowheight,columnwidth) if heldblock[4]==cyan: blueswitch=columnwidth/2 for i in range(4): heldblock[i].left-=(columnwidth/2) else: blueswitch=0 for i in range (4): pygame.draw.rect(window,heldblock[4],heldblock[i]) for i in range(4): pygame.draw.line(window,BLACK,(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),rowheight*15),(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),(rowheight*20))) for i in range (4): pygame.draw.line(window,BLACK,(WINDOWWIDTH+1,rowheight*(i+1)),(notplaywindow,rowheight*(i+1))) if nexpiece[0][4]==cyan: blueswitch=columnwidth/2 else: blueswitch=0 for i in range(4): pygame.draw.line(window,BLACK,(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),0),(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),(rowheight*4))) for i in range (1,10): pygame.draw.line(window, BLACK, (columnwidth*i,0),(columnwidth*i,WINDOWHEIGHT)) for i in range (1,20): pygame.draw.line(window, BLACK, (0,rowheight*i),(WINDOWWIDTH,rowheight*i)) pygame.draw.line(window, BLACK, (WINDOWWIDTH,rowheight*17),(notplaywindow,rowheight*17)) pygame.draw.line(window, (255,255,255), (WINDOWWIDTH,0), (WINDOWWIDTH,WINDOWHEIGHT),4) hi+=1 pygame.display.update() time.sleep(speed)
33.031603
400
0.517802
import pygame, sys, time, random, datetime from pygame.locals import * pygame.init() WINDOWWIDTH = 340 WINDOWHEIGHT = 2*WINDOWWIDTH columnwidth=WINDOWWIDTH/10 rowheight=WINDOWHEIGHT/20 notplaywindow=WINDOWWIDTH+(5*columnwidth) window = pygame.display.set_mode((notplaywindow, WINDOWHEIGHT), 0, 32) pygame.display.set_caption('Tetris') BLACK = (0, 0, 0) red = (255, 0, 0) green = (0, 255, 0) blue = (0, 0, 255) purple=(120,0,120) WHITE=(255,255,255) yellow=(255,255,0) cyan=(0, 255, 255) grey=(128, 128, 128) orange=(255, 127, 0) points=0 blocks=[[[-1,0],[1,0],[2,0],cyan],[[-1,0],[-1,-1],[1,0],blue],[[-1,0],[1,0],[1,-1],orange],[[0,-1],[1,-1],[1,0],yellow],[[-1,0],[0,-1],[1,-1],green],[[-1,0],[0,-1],[1,0],purple],[[-1,-1],[0,-1],[1,0],red]] blockonscreen=[] startspotx=columnwidth*4 startspoty=-rowheight newblock=1 fix=0 speed=.02 hi=0 blockmovecount=9 def makenewblock(block): return [pygame.Rect(startspotx,startspoty,columnwidth,rowheight),pygame.Rect(startspotx+(block[0][0]*columnwidth),startspoty+(block[0][1]*rowheight),columnwidth,rowheight),pygame.Rect(startspotx+(block[1][0]*columnwidth),startspoty+(block[1][1]*rowheight),columnwidth,rowheight),pygame.Rect(startspotx+(block[2][0]*columnwidth),startspoty+(block[2][1]*rowheight),columnwidth,rowheight),block[3]] def overlap(one,two): for bk in two: if bk!=curblock: for b in range(len(bk)-1): if one==bk[b]: return True return False def harddrop(): fix=0 while True: for i in range(4): curblock[i].top+=(rowheight) if curblock[i].top==WINDOWHEIGHT: fix=1 if overlap(curblock[i],blockonscreen): fix=1 if fix: for i in range(4): curblock[i].top-=(rowheight) fix=0 lineclear() break def makeghost(): fix=0 ghost=[] for i in range(4): ghost.append(pygame.Rect(curblock[i].left,curblock[i].top,curblock[i].width,curblock[i].height)) ghost.append((40,40,40)) while True: for i in range(4): ghost[i].top+=(rowheight) if ghost[i].top==WINDOWHEIGHT: fix=1 if overlap(ghost[i],blockonscreen): fix=1 if fix: for i in range(4): ghost[i].top-=(rowheight) fix=0 break return ghost def rotfix(drec): if drec=='UP': for i in range(4): curblock[i].top-=rowheight elif drec=='LEFT': for i in range(4): curblock[i].left-=columnwidth elif drec=='RIGHT': for i in range(4): curblock[i].left+=columnwidth bag=[0,1,2,3,4,5,6] def shufflefirst(): newbag=[] newbag2=[] for i in range (7): while True: x=random.choice(bag) if x not in newbag: newbag.append(x) break while True: y=random.choice(bag) if y not in newbag2: newbag2.append(x) break newbag3=newbag+newbag2 return newbag3 def shuffle(): newbag1=pieces[7:] newbag=[] for i in range (7): while True: x=random.choice(bag) if x not in newbag: newbag.append(x) break newbag3=newbag1+newbag return newbag3 def copyrect(rect): return pygame.Rect(rect.left,rect.top,rect.width,rect.height) def rotatecheck(curc, ro, col): new=[] for z in range(len(curc)): if z != 4: new.append(copyrect(curc[z])) else: new.append(curc[4]) rotate(new, ro, col) for i in range(len(new)-1): if new[i].left<0 or new[i].left>=WINDOWWIDTH: return False for b in blockonscreen: if b != curc: for x in range(len(b)-1): if new[i] == b[x]: return False return True def rotate(curr,ro,col): if curr[4]==yellow: rotateo(curr,ro,col) if curr[4]==cyan: rotatei(curr,ro,col) else: rotatea(curr,ro,col) def rotatei(curr,ro,col): if curr[2].centerx>curr[0].centerx and curr[3].centerx>curr[0].centerx: horiz=True curr[0].centerx+=col bla=1 elif curr[2].centerx<curr[0].centerx and curr[3].centerx<curr[0].centerx: horiz=True curr[0].centerx-=col bla=-1 elif curr[2].centery>curr[0].centery and curr[3].centery>curr[0].centery: horiz=False curr[0].centery+=ro bla=-1 elif curr[2].centery<curr[0].centery and curr[3].centery<curr[0].centery: horiz=False curr[0].centery-=ro bla=1 for f in range (1,4): if horiz: curr[f].left=curr[0].left curr[f].top=curr[0].top+(blocks[0][f-1][0]*ro*bla) else: curr[f].top=curr[0].top curr[f].right=curr[0].right+(blocks[0][f-1][0]*col*bla) def rotateo(curr,ro,col): pass def rotatea(curr,ro,col): middlex=curr[0].centerx middley=curr[0].centery for r in range(1,4): if curr[r].centery==middley and curr[r].centerx>middlex: curr[r].centery+=ro curr[r].centerx-=col elif curr[r].centery==middley and curr[r].centerx<middlex: curr[r].centery-=ro curr[r].centerx+=col elif curr[r].centerx==middlex and curr[r].centery>middley: curr[r].centery-=ro curr[r].centerx-=col elif curr[r].centerx==middlex and curr[r].centery<middley: curr[r].centery+=ro curr[r].centerx+=col elif curr[r].centerx>middlex and curr[r].centery>middley: curr[r].centerx-=(2*col) elif curr[r].centery>middley and curr[r].centerx<middlex: curr[r].centery-=(2*ro) elif curr[r].centerx<middlex and curr[r].centery<middley: curr[r].centerx+=(2*col) elif curr[r].centerx>middlex and curr[r].centery<middley: curr[r].centery+=(2*ro) def makenextpiece(block,startx,starty,row,column): return [pygame.Rect(startx,starty,column,row),pygame.Rect((block[0][0]*column)+startx,(block[0][1]*row)+starty,column,row),pygame.Rect((block[1][0]*column)+startx,(block[1][1]*row)+starty,column,row),pygame.Rect((block[2][0]*column)+startx,(block[2][1]*row)+starty,column,row),block[3]] def nextpiece(): nextpieces=[] for b in range (4): if b==0: newrowheight=WINDOWHEIGHT/20 newcolumnwidth=WINDOWWIDTH/10 else: newrowheight=WINDOWHEIGHT/40 newcolumnwidth=WINDOWWIDTH/20 nextpiece=makenextpiece(blocks[pieces[piece+b]],notplaywindow-(3*rowheight),startspoty+(rowheight*((3*b)+3)),newrowheight,newcolumnwidth) if nextpiece[4]!=cyan and newrowheight!=WINDOWWIDTH/20: rotate(nextpiece,newrowheight,newcolumnwidth) else: for i in range(4): nextpiece[i].left-=(newcolumnwidth/2) nextpieces.append(nextpiece) return nextpieces def lineclear(): bla=False numd=[] for d in reversed(range(20)): a=0 for c in blockonscreen: for x in range(len(c)-1): if c[x].top==rowheight*d: a+=1 if a>=10: for c in blockonscreen: b=0 for x in range(len(c)-1): x=x-b if c[x].top==rowheight*d: c.remove(c[x]) b+=1 bla=True for c in blockonscreen: for x in range(len(c)-1): if c[x].top<rowheight*d: c[x].top+=rowheight numd.append('') global points points+=10 lineclear() break for bk in blockonscreen: if len(bk)==1: blockonscreen.remove(bk) return len(numd) held=[] def holdblock(): newblock=0 curb='' if len(held)==0: held.append(curblock[4]) for i in range (7): if blocks[i][3]==held[0]: held.append(i) held.remove(held[0]) blockonscreen.remove(curblock) newblock=1 else: blockonscreen.append(makenewblock(blocks[held[0]])) held.remove(held[0]) blockonscreen.remove(curblock) held.append(curblock[4]) for i in range (7): if blocks[i][3]==held[0]: held.append(i) held.remove(held[0]) curb=blockonscreen[len(blockonscreen)-1] return newblock,curb pieces=shufflefirst() piece=0 print datetime.datetime.now() while True: for event in pygame.event.get(): if event.type == QUIT: pygame.quit() sys.exit() if event.type == pygame.KEYDOWN: if event.key == pygame.K_LEFT: for i in range(4): curblock[i].left-=columnwidth if curblock[i].left==-columnwidth: fix=1 if overlap(curblock[i],blockonscreen): fix=1 if fix: for i in range(4): curblock[i].left+=columnwidth fix=0 ghost=makeghost() elif event.key == pygame.K_RIGHT: for i in range(4): curblock[i].left+=columnwidth if curblock[i].left==WINDOWWIDTH: fix=1 if overlap(curblock[i],blockonscreen): fix=1 if fix: for i in range(4): curblock[i].left-=columnwidth fix=0 ghost=makeghost() elif event.key == pygame.K_UP: if rotatecheck(curblock,rowheight,columnwidth): rotate(curblock,rowheight,columnwidth) ghost=makeghost() elif event.key == pygame.K_DOWN: origblockmovecount=blockmovecount blockmovecount=2 elif event.key == pygame.K_SPACE: harddrop() newblock=1 hi=0 ghost=ghost[4:] elif event.key == pygame.K_p: brea=False while True: for event in pygame.event.get(): if event.type == QUIT: pygame.quit() sys.exit() if event.type==pygame.KEYDOWN: brea=True if brea: break elif event.key==pygame.K_RSHIFT or event.key==pygame.K_LSHIFT: n=holdblock() newblock=n[0] curblock=n[1] if curblock!='': ghost=makeghost() if event.type == pygame.KEYUP: if event.key == pygame.K_DOWN: blockmovecount=origblockmovecount window.fill(BLACK) if newblock: blockonscreen.append(makenewblock(blocks[pieces[piece]])) newblock=0 curblock=blockonscreen[len(blockonscreen)-1] ghost=makeghost() piece+=1 if piece==7: pieces=shuffle() piece=0 nexpiece=nextpiece() print(points) if hi>=blockmovecount: for i in range(4): curblock[i].top+=(rowheight) if curblock[i].bottom>WINDOWHEIGHT: fix=1 newblock=1 if overlap(curblock[i],blockonscreen): fix=1 newblock=1 hi=0 if fix: for i in range(4): curblock[i].top-=(rowheight) fix=0 lineclear() for i in range (len(ghost)-1): pygame.draw.rect(window,ghost[len(ghost)-1],ghost[i]) for block in blockonscreen: for i in range(len(block)-1): pygame.draw.rect(window,block[len(block)-1],block[i]) for n in nexpiece: for i in range (4): pygame.draw.rect(window,n[4],n[i]) if len(held)!=0: heldblock=makenextpiece(blocks[held[0]],notplaywindow-(columnwidth*3),WINDOWHEIGHT-(rowheight*3),rowheight,columnwidth) if heldblock[4]==cyan: blueswitch=columnwidth/2 for i in range(4): heldblock[i].left-=(columnwidth/2) else: blueswitch=0 for i in range (4): pygame.draw.rect(window,heldblock[4],heldblock[i]) for i in range(4): pygame.draw.line(window,BLACK,(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),rowheight*15),(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),(rowheight*20))) for i in range (4): pygame.draw.line(window,BLACK,(WINDOWWIDTH+1,rowheight*(i+1)),(notplaywindow,rowheight*(i+1))) if nexpiece[0][4]==cyan: blueswitch=columnwidth/2 else: blueswitch=0 for i in range(4): pygame.draw.line(window,BLACK,(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),0),(WINDOWWIDTH+blueswitch+((i+1)*columnwidth),(rowheight*4))) for i in range (1,10): pygame.draw.line(window, BLACK, (columnwidth*i,0),(columnwidth*i,WINDOWHEIGHT)) for i in range (1,20): pygame.draw.line(window, BLACK, (0,rowheight*i),(WINDOWWIDTH,rowheight*i)) pygame.draw.line(window, BLACK, (WINDOWWIDTH,rowheight*17),(notplaywindow,rowheight*17)) pygame.draw.line(window, (255,255,255), (WINDOWWIDTH,0), (WINDOWWIDTH,WINDOWHEIGHT),4) hi+=1 pygame.display.update() time.sleep(speed)
false
true
1c484057952e765042ba5f556beae1700c93a132
2,106
py
Python
services/users/project/api/utils/response.py
shwetha-manvinkurke/dx-automator
ec01e51d80c8be8f5dea4669baa25d38256b1052
[ "MIT" ]
14
2018-01-04T22:33:54.000Z
2020-03-04T18:38:34.000Z
services/users/project/api/utils/response.py
shwetha-manvinkurke/dx-automator
ec01e51d80c8be8f5dea4669baa25d38256b1052
[ "MIT" ]
87
2018-01-04T22:15:16.000Z
2022-01-06T14:49:07.000Z
services/users/project/api/utils/response.py
shwetha-manvinkurke/dx-automator
ec01e51d80c8be8f5dea4669baa25d38256b1052
[ "MIT" ]
17
2018-01-04T23:33:48.000Z
2021-11-08T18:39:04.000Z
def response_json_ok(json): """Creates a tuple representing the HTTP package to respond the requisition with the given JSON on its body and status code 200 :param json: object to be sent on HTTP body :return response: tuple representing the HTTP response package """ return _make_json_response(json, 200) def response_json_created(json): """Creates a tuple representing the HTTP package to respond the requisition with the given JSON on its body and status code 201 :param json: object to be sent on HTTP body :return response: tuple representing the HTTP response package """ return _make_json_response(json, 201) def response_json_bad_request(json): """Creates a tuple representing the HTTP package to respond the requisition with the given JSON on its body and status code 400 :param json: object to be sent on HTTP body :return response: tuple representing the HTTP response package """ return _make_json_response(json, 400) def response_json_unauthorized(json): """Creates a tuple representing the HTTP package to respond the requisition with the given JSON on its body and status code 401 :param json: object to be sent on HTTP body :return response: tuple representing the HTTP response package """ return _make_json_response(json, 401) def response_json_not_found(json): """Creates a tuple representing the HTTP package to respond the requisition with the given JSON on its body and status code 404 :param json: object to be sent on HTTP body :return response: tuple representing the HTTP response package """ return _make_json_response(json, 404) def _make_json_response(json, status): """Creates a tuple representing the HTTP package to respond the requisition with the given JSON on its body and the given status code. :param json: object to be sent on HTTP body :param status: status code :return response: tuple representing the HTTP response package """ return json, status, {'Content-Type': 'application/json'}
36.310345
66
0.731244
def response_json_ok(json): return _make_json_response(json, 200) def response_json_created(json): return _make_json_response(json, 201) def response_json_bad_request(json): return _make_json_response(json, 400) def response_json_unauthorized(json): return _make_json_response(json, 401) def response_json_not_found(json): return _make_json_response(json, 404) def _make_json_response(json, status): return json, status, {'Content-Type': 'application/json'}
true
true
1c48412fd41a287281f75f9338ade4cb7bd2bfd1
12,087
py
Python
src/relstorage/cache/tests/test_lru_cffiring.py
enfold/relstorage
9fcd526b537cb6537cc2ae33154b63096550f210
[ "ZPL-2.1" ]
40
2015-10-08T05:35:13.000Z
2022-03-28T23:50:06.000Z
src/relstorage/cache/tests/test_lru_cffiring.py
enfold/relstorage
9fcd526b537cb6537cc2ae33154b63096550f210
[ "ZPL-2.1" ]
364
2015-03-23T15:25:42.000Z
2022-03-17T08:41:34.000Z
src/relstorage/cache/tests/test_lru_cffiring.py
enfold/relstorage
9fcd526b537cb6537cc2ae33154b63096550f210
[ "ZPL-2.1" ]
33
2015-06-08T23:03:22.000Z
2022-03-21T08:25:53.000Z
############################################################################## # # Copyright (c) 2009 Zope Foundation and Contributors. # All Rights Reserved. # # This software is subject to the provisions of the Zope Public License, # Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution. # THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED # WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS # FOR A PARTICULAR PURPOSE. # ############################################################################## from __future__ import absolute_import from __future__ import division from __future__ import print_function from hamcrest import assert_that from nti.testing.matchers import validly_provides # The overhead of cache values, and thus how much fits in a ring or the # cache, depends on 32 or 64 bit, whether or not we copy # bytes into native code std::string, and how the compiler lays out # the objects. We only copy strings on PyPy, but we have little control # over the object layout, especially with the various MSVC compilers # we have to deal with. So that explains the tests that have a range of sizes. from relstorage.tests import TestCase from relstorage.cache import interfaces from . import Cache class GenerationTests(TestCase): def _makeCache(self, limit): from . import Cache as BaseCache return BaseCache(limit) def _makeOne(self, limit): return self._makeCache(limit).eden def test_bool(self): cache = self._makeCache(100) lru = cache.eden self.assertFalse(lru) cache[1] = (b'', 0) self.assertTrue(lru) self.assertEqual(1, len(lru)) del cache[1] self.assertFalse(lru) class EdenTests(TestCase): def _makeOne(self, limit): from . import Cache as BaseCache return BaseCache(limit) def test_add_MRUs_empty(self): lru = self._makeOne(100) self.assertEqual((), lru.add_MRUs([])) def test_add_MRUs_too_many(self): lru = self._makeOne(1000) too_many = [(i, (b'a' * i, 0, 0, 1)) for i in range(50)] # They just exceed the limit added = lru.add_MRUs(too_many) # Much less got added self.assertGreaterEqual(len(added), 7) self.assertLessEqual(len(added), 9) class NoOverheadSizeCache(Cache): def __init__(self, byte_limit): Cache.__init__(self, byte_limit) self.base_size = self.weight self[0] = (b'', 0) self.entry_size = self[0].weight del self[0] @property def weight(self): weight = super(NoOverheadSizeCache, self).size weight -= self.base_size weight -= self.entry_size * len(self) return weight @property def size(self): return self.weight def __getitem__(self, oid): entry = Cache.__getitem__(self, oid) if entry is not None: return self.get_item_with_tid(oid, entry.tid) class GenericLRUCacheTests(TestCase): """ Generic LRU caching tests that can be applied to any LRU implementation, using the kind of keys and values that we actually use: ``(oid_int, tid_int)`` and ``(state_bytes, tid_int)``. """ def _getClass(self): return NoOverheadSizeCache def _makeOne(self, limit, kind=None): kind = kind or self._getClass() return kind(limit) def _getIface(self): return interfaces.ILRUCache def test_implements(self): cache = self._makeOne(100) assert_that(cache, validly_provides(self._getIface())) return cache def test_eden_implements(self): cache = self._makeOne(100) assert_that(cache.eden, validly_provides(interfaces.IGeneration)) def test_item_implements(self): cache = self._makeOne(20) cache[1] = (b'', 0) entrya = cache[1] assert_that(entrya, validly_provides(interfaces.ILRUEntry)) def test_add_too_many(self): class _Cache(self._getClass()): pass cache = _Cache(20 + _Cache(20).base_size + (_Cache(20).entry_size * 2)) entry_count = 10 entries = cache.add_MRUs(list(reversed([ # oid, state, frozen, frequency (x, (b'abcde', 0, False, x)) for x in range(1, entry_count) ]))) self.assertEqual( [5 + cache.entry_size] * len(entries), [e.weight for e in entries]) self.assertLessEqual( cache.weight, cache.limit ) self.assertEqual( [e.key for e in entries], [e.frequency for e in entries]) self.assertEqual( [8, 7, 6, 5], [e.key for e in entries] ) self.assertEqual(4, len(cache)) return cache def test_age(self): cache = self._makeOne(100) base_size = cache.base_size entry_size = cache.entry_size cache = self._getClass()(100 + base_size + entry_size) entries = cache.add_MRUs([ (1, (b'abcde', 0, False, 1)), (2, (b'abcde', 0, False, 1)), (3, (b'abcde', 0, False, 1)), (0, (b'abcde', 0, False, 1)), ]) self.assertIn( [e.key for e in entries], ([1, 2, 3, 0], [2, 3, 0]) ) for _ in range(4): for e in entries: _ = cache[e.key] freqs = [e.frequency for e in cache.values()] self.assertEqual([5] * len(entries), freqs) # By half each time cache.age_frequencies() freqs = [e.frequency for e in cache.values()] self.assertEqual([2] * len(entries), freqs) return cache def test_delete(self): cache = self._makeOne(20) cache[1] = (b'abc', 0) self.assertIn(1, cache) self.assertEqual(1, len(cache)) self.assertEqual(3, cache.size) self.assertEqual(cache[1], (b'abc', 0)) self.assertEqual(list(cache), [(1, 0)]) del cache[1] self.assertNotIn(1, cache) self.assertEqual(0, len(cache)) self.assertEqual(0, cache.size) self.assertIsNone(cache[1]) self.assertEqual(list(cache), []) def test_entries(self): cache = self._makeOne(20) cache[1] = (b'abc', 0) entries = list(cache.values()) self.assertEqual(1, len(entries)) entry = entries[0] assert_that(entry, validly_provides(interfaces.ILRUEntry)) self.assertEqual(1, entry.key) self.assertEqual(b'abc', entry.value) self.assertEqual(1, entry.frequency) # Getting it again updates its frequency, not # necessarily on the same object though. self.assertIsNotNone(cache[1]) entries = list(cache.values()) self.assertEqual(1, len(entries)) entry = entries[0] self.assertEqual(1, entry.key) self.assertEqual(b'abc', entry.value) self.assertEqual(2, entry.frequency) def test_add_too_many_MRUs_works_aronud_big_entry(self): cache = self._getClass()(20) base_size = cache.base_size entry_size = cache.entry_size cache = self._getClass()(40 + base_size + entry_size) entries = cache.add_MRUs([ (1, (b'abc', 0, False, 1)), # This entry itself will fit nowhere (2, (b'12345678901234567890' * 20, 0, False, 1)), (3, (b'bcd', 0, False, 1)), (4, (b'cde', 0, False, 1)), (5, (b'dehi', 0, False, 1)), (6, (b'edghijkl', 0, False, 1)), ]) self.assertGreaterEqual(len(cache), 3) self.assertLessEqual(len(cache), 4) self.assertIn( [e.key for e in entries], ([1, 3, 4, 5], [3, 4, 5])) return cache class GenericGenerationalLRUCacheTests(GenericLRUCacheTests): """ Tests for any generational LRU cache. """ def test_implements(self): cache = super(GenericGenerationalLRUCacheTests, self).test_implements() assert_that(cache.eden, validly_provides(interfaces.IGeneration)) assert_that(cache.protected, validly_provides(interfaces.IGeneration)) assert_that(cache.probation, validly_provides(interfaces.IGeneration)) def test_bad_generation_index_attribute_error(self): cache = self._makeOne(20) # Check proper init getattr(cache.generations[1], 'limit') getattr(cache.generations[2], 'limit') getattr(cache.generations[3], 'limit') # Gen 0 should be missing with self.assertRaisesRegex(AttributeError, "Generation 0 has no attribute 'on_hit'"): cache.generations[0].on_hit() def test_add_MRUs_reject_sets_sentinel_values(self): # When we find an item that completely fills the cache, # all the rest of the items are marked as rejected. cache = self._getClass()(20) base_size = cache.base_size entry_size = cache.entry_size cache = self._getClass()(20 + base_size + entry_size) added_entries = cache.add_MRUs([ # over fill eden with item of size 15 (1, (b'012345678901234' * 20, 0, False, 1)), # 1 goes to protected, filling it. eden is also over full with 2. probation is empty (2, (b'012', 0, False, 1)), # 3 fills eden, bumping 2 to probation. But probation is actually overfull now # so we'd like to spill something if we could (but we can't.) (3, (b'0', 0, False, 1)), # 4 should never be added because it won't fit anywhere. (4, (b'ee', 0, False, 1)), ]) def keys(x): return [e.key for e in x] self.assertEqual(keys(cache.protected), [3, 2]) self.assertEqual(keys(cache.probation), []) self.assertEqual(keys(cache.eden), [4]) self.assertEqual( [2, 3, 4], [e.key for e in added_entries]) self.assertEqual(3, len(added_entries)) self.assertEqual(3, len(cache)) self.assertEqual(3, len(list(cache))) class CFFICacheTests(TestCase): """ Tests that are specific to the CFFI implementation of the cache. These can use arbitrary keys and values. """ def _getClass(self): return NoOverheadSizeCache def _makeOne(self, limit, kind=None): self.skipTest("Weights not supported") kind = kind or self._getClass() return kind(limit, key_weight=self.key_weight, value_weight=self.value_weight) def key_weight(self, k): return len(k) def value_weight(self, v): return len(v) def test_free_reuse(self): cache = self._makeOne(20) lru = cache.protected self.assertEqual(lru.limit, 16) entrya = lru.add_MRU('a', b'') entryb = lru.add_MRU('b', b'') entryc = lru.add_MRU('c', b'1') entryd = lru.add_MRU('d', b'1') evicted = lru.update_MRU(entryb, b'1234567890') self.assertEqual(evicted, ()) # Not changing the size is just a hit, it doesnt't # evict anything. evicted = lru.update_MRU(entryb, b'1234567890') self.assertEqual(evicted, ()) evicted = lru.update_MRU(entryc, b'1234567890') # a and d were evicted and placed on the freelist self.assertEqual(entrya.key, None) self.assertEqual(entrya.value, None) self.assertEqual(entryd.key, None) self.assertEqual(entryd.key, None) self.assertEqual(evicted, [('a', b''), ('d', b'1')]) self.assertEqual(2, len(lru.node_free_list)) lru.add_MRU('c', b'1') self.assertEqual(1, len(lru.node_free_list))
32.579515
96
0.587822
[ (1, (b'abc', 0, False, 1)), (2, (b'12345678901234567890' * 20, 0, False, 1)), (3, (b'bcd', 0, False, 1)), (4, (b'cde', 0, False, 1)), (5, (b'dehi', 0, False, 1)), (6, (b'edghijkl', 0, False, 1)), ]) self.assertGreaterEqual(len(cache), 3) self.assertLessEqual(len(cache), 4) self.assertIn( [e.key for e in entries], ([1, 3, 4, 5], [3, 4, 5])) return cache class GenericGenerationalLRUCacheTests(GenericLRUCacheTests): def test_implements(self): cache = super(GenericGenerationalLRUCacheTests, self).test_implements() assert_that(cache.eden, validly_provides(interfaces.IGeneration)) assert_that(cache.protected, validly_provides(interfaces.IGeneration)) assert_that(cache.probation, validly_provides(interfaces.IGeneration)) def test_bad_generation_index_attribute_error(self): cache = self._makeOne(20) getattr(cache.generations[1], 'limit') getattr(cache.generations[2], 'limit') getattr(cache.generations[3], 'limit') with self.assertRaisesRegex(AttributeError, "Generation 0 has no attribute 'on_hit'"): cache.generations[0].on_hit() def test_add_MRUs_reject_sets_sentinel_values(self): cache = self._getClass()(20) base_size = cache.base_size entry_size = cache.entry_size cache = self._getClass()(20 + base_size + entry_size) added_entries = cache.add_MRUs([ (1, (b'012345678901234' * 20, 0, False, 1)), (2, (b'012', 0, False, 1)), (3, (b'0', 0, False, 1)), (4, (b'ee', 0, False, 1)), ]) def keys(x): return [e.key for e in x] self.assertEqual(keys(cache.protected), [3, 2]) self.assertEqual(keys(cache.probation), []) self.assertEqual(keys(cache.eden), [4]) self.assertEqual( [2, 3, 4], [e.key for e in added_entries]) self.assertEqual(3, len(added_entries)) self.assertEqual(3, len(cache)) self.assertEqual(3, len(list(cache))) class CFFICacheTests(TestCase): def _getClass(self): return NoOverheadSizeCache def _makeOne(self, limit, kind=None): self.skipTest("Weights not supported") kind = kind or self._getClass() return kind(limit, key_weight=self.key_weight, value_weight=self.value_weight) def key_weight(self, k): return len(k) def value_weight(self, v): return len(v) def test_free_reuse(self): cache = self._makeOne(20) lru = cache.protected self.assertEqual(lru.limit, 16) entrya = lru.add_MRU('a', b'') entryb = lru.add_MRU('b', b'') entryc = lru.add_MRU('c', b'1') entryd = lru.add_MRU('d', b'1') evicted = lru.update_MRU(entryb, b'1234567890') self.assertEqual(evicted, ()) # Not changing the size is just a hit, it doesnt't evicted = lru.update_MRU(entryb, b'1234567890') self.assertEqual(evicted, ()) evicted = lru.update_MRU(entryc, b'1234567890') self.assertEqual(entrya.key, None) self.assertEqual(entrya.value, None) self.assertEqual(entryd.key, None) self.assertEqual(entryd.key, None) self.assertEqual(evicted, [('a', b''), ('d', b'1')]) self.assertEqual(2, len(lru.node_free_list)) lru.add_MRU('c', b'1') self.assertEqual(1, len(lru.node_free_list))
true
true
1c48417f4536995b8d781890b1514b8e62adaaf0
607
py
Python
apps/static_pages/tests/test_urls.py
ilyukevich/tasks
ba0c8202cfe61d26975c35f388155d36e1c2b856
[ "MIT" ]
null
null
null
apps/static_pages/tests/test_urls.py
ilyukevich/tasks
ba0c8202cfe61d26975c35f388155d36e1c2b856
[ "MIT" ]
null
null
null
apps/static_pages/tests/test_urls.py
ilyukevich/tasks
ba0c8202cfe61d26975c35f388155d36e1c2b856
[ "MIT" ]
null
null
null
from django.test import Client, TestCase class StaticURLTests(TestCase): def setUp(self): self.guest_client = Client() def test_about_url_exists_at_desired_location(self): """Проверка доступности адреса /page/about/.""" response = self.guest_client.get('/page/about/') self.assertEqual(response.status_code, 200) def test_about_url_uses_correct_template(self): """Проверка шаблона для адреса /page/about/.""" response = self.guest_client.get('/page/about/') self.assertTemplateUsed(response, 'static_pages/about.html')
35.705882
69
0.680395
from django.test import Client, TestCase class StaticURLTests(TestCase): def setUp(self): self.guest_client = Client() def test_about_url_exists_at_desired_location(self): response = self.guest_client.get('/page/about/') self.assertEqual(response.status_code, 200) def test_about_url_uses_correct_template(self): response = self.guest_client.get('/page/about/') self.assertTemplateUsed(response, 'static_pages/about.html')
true
true
1c4841829620980ad574c246291fe78ff2d81173
5,779
py
Python
projects/mammography_project/integrate_final_result.py
lanhsincheng/detectron2
45ec85c3bde2a39ed4e870b76442021e8da26ede
[ "Apache-2.0" ]
null
null
null
projects/mammography_project/integrate_final_result.py
lanhsincheng/detectron2
45ec85c3bde2a39ed4e870b76442021e8da26ede
[ "Apache-2.0" ]
null
null
null
projects/mammography_project/integrate_final_result.py
lanhsincheng/detectron2
45ec85c3bde2a39ed4e870b76442021e8da26ede
[ "Apache-2.0" ]
null
null
null
from detectron2.utils.visualizer import ColorMode import cv2 import random from detectron2.utils.visualizer import Visualizer from projects.mammography_project.mammo_dataset import * import operator import xlsxwriter import csv wb_name = 'mammo0824_model_0059999' def mammo_integrate(test_dirname, predictor, dataset_metadata, test_data_csv_path, output_dir): dataset_dicts = get_mammo_dicts(test_dirname,test_data_csv_path) answer_sheet_list = [] for_ensemble_confidence_list = [] score_class_list = [] big_list = [] for d in dataset_dicts: im = cv2.imread(d["file_name"]) outputs = predictor(im) num_instances = len(outputs['instances']) scores_class_dict = {} for_ensemble_dict = {} per_class_dict = {} for s in range(num_instances): scores = outputs['instances']._fields['scores'].T[s].item() pred_classes_num = outputs['instances']._fields['pred_classes'].T[s].item() per_class_dict.update({pred_classes_num : scores}) # if pred_classes_num == 0 or pred_classes_num == 2 or pred_classes_num == 4 or pred_classes_num == 6: # pred_classes = 'benign' # elif pred_classes_num == 1 or pred_classes_num == 3 or pred_classes_num == 5 or pred_classes_num == 7: # pred_classes = 'malignant' if pred_classes_num == 0 : pred_classes = 'benign' elif pred_classes_num == 1 : pred_classes = 'malignant' # if pred_classes_num == 1: # pred_classes = 'benign' # elif pred_classes_num == 0: # pred_classes = 'malignant' scores_class_dict.update( {scores : pred_classes} ) if (pred_classes not in for_ensemble_dict) or (pred_classes in for_ensemble_dict and for_ensemble_dict[pred_classes] < scores): for_ensemble_dict.update( {pred_classes : scores} ) if ('benign' not in for_ensemble_dict): for_ensemble_dict.update({'benign': 0}) if ('malignant' not in for_ensemble_dict): for_ensemble_dict.update({'malignant': 0}) per_class_dict = sorted(per_class_dict.items(), key= lambda per_class_dict: per_class_dict[0]) big_list.append(per_class_dict) if ( not bool(scores_class_dict)==True ): print(scores_class_dict) scores_class_dict.update({0: 'benign', 0: 'malignant'}) print(scores_class_dict) score_class_list.append(scores_class_dict) final_for_ensemble = sorted(for_ensemble_dict.items(), key=lambda for_ensemble_dict: for_ensemble_dict[0]) final_class = max(scores_class_dict.items(), key=lambda scores_class_dict: scores_class_dict[0])[1] answer_sheet_list.append(final_class) for i in range(len(final_for_ensemble)): for_ensemble_confidence_list.append(final_for_ensemble[i][1]) # load golden answer and evaluate accuracy golden = [] golden_sheet = r'D:\Mammograph\golden/balance_golden_v1_852.csv' with open(golden_sheet, newline='') as csvFile: T = 0 F = 0 mm = 0 bb = 0 bm = 0 mb = 0 rows = csv.reader(csvFile) for row in rows: golden.append(row[0]) for i, j in zip(golden, answer_sheet_list): if i == j: if(i=='benign'): bb += 1 else: mm += 1 T += 1 else: if (i == 'benign'): bm += 1 else: mb += 1 F += 1 accuracy = T / (T+F) accuracy_malignant = (mm + bb)/(mm + bb + bm + mb) accuracy_benign = (bb + mm)/(bb + mm + mb + bm) print('T: ', T, ' F: ', F, ' accuracy: ',accuracy, 'malignant to benign : ', mb, 'benign to malignant : ', bm) # write class and confidence per bounding box to the xlsfile(4 classes) big_name = 'big_' + wb_name + '.xlsx' workbook = xlsxwriter.Workbook(big_name) worksheet = workbook.add_worksheet() row = 0 column = 0 for list_ele in big_list: for item in list_ele: worksheet.write(row, column, item[0]) column += 1 worksheet.write(row, column, item[1]) column += 1 column = 0 row += 1 workbook.close() # write class and confidence per bounding box to the xlsfile dict_name = 'dict_' + wb_name + '.xlsx' # workbook = xlsxwriter.Workbook(r'dict.xlsx') workbook = xlsxwriter.Workbook(dict_name) worksheet = workbook.add_worksheet() row = 0 column = 0 for dict_ele in score_class_list: item_list = [] for key, value in dict_ele.items(): item_list.append(key) item_list.append(value) for item in item_list: worksheet.write(row, column, item) column += 1 column = 0 row += 1 workbook.close() # write final_class to the xlsfile ans_name = 'ans_' + wb_name + '.xlsx' workbook = xlsxwriter.Workbook(ans_name) worksheet = workbook.add_worksheet() row = 0 column = 0 # write down answer sheet for item in answer_sheet_list: # write operation perform worksheet.write(row, column, item) row += 1 workbook.close() # write down confidence for every class integrate_name = 'integrate_' + wb_name + '.csv' with open(integrate_name, 'w', newline='') as csvfile: writer = csv.writer(csvfile) # 2 items a row for i in range(0, len(for_ensemble_confidence_list), 2): # step by threes. writer.writerow(for_ensemble_confidence_list[i:i + 2]) return T, F, accuracy, mb, bm
38.526667
139
0.607372
from detectron2.utils.visualizer import ColorMode import cv2 import random from detectron2.utils.visualizer import Visualizer from projects.mammography_project.mammo_dataset import * import operator import xlsxwriter import csv wb_name = 'mammo0824_model_0059999' def mammo_integrate(test_dirname, predictor, dataset_metadata, test_data_csv_path, output_dir): dataset_dicts = get_mammo_dicts(test_dirname,test_data_csv_path) answer_sheet_list = [] for_ensemble_confidence_list = [] score_class_list = [] big_list = [] for d in dataset_dicts: im = cv2.imread(d["file_name"]) outputs = predictor(im) num_instances = len(outputs['instances']) scores_class_dict = {} for_ensemble_dict = {} per_class_dict = {} for s in range(num_instances): scores = outputs['instances']._fields['scores'].T[s].item() pred_classes_num = outputs['instances']._fields['pred_classes'].T[s].item() per_class_dict.update({pred_classes_num : scores}) if pred_classes_num == 0 : pred_classes = 'benign' elif pred_classes_num == 1 : pred_classes = 'malignant' scores_class_dict.update( {scores : pred_classes} ) if (pred_classes not in for_ensemble_dict) or (pred_classes in for_ensemble_dict and for_ensemble_dict[pred_classes] < scores): for_ensemble_dict.update( {pred_classes : scores} ) if ('benign' not in for_ensemble_dict): for_ensemble_dict.update({'benign': 0}) if ('malignant' not in for_ensemble_dict): for_ensemble_dict.update({'malignant': 0}) per_class_dict = sorted(per_class_dict.items(), key= lambda per_class_dict: per_class_dict[0]) big_list.append(per_class_dict) if ( not bool(scores_class_dict)==True ): print(scores_class_dict) scores_class_dict.update({0: 'benign', 0: 'malignant'}) print(scores_class_dict) score_class_list.append(scores_class_dict) final_for_ensemble = sorted(for_ensemble_dict.items(), key=lambda for_ensemble_dict: for_ensemble_dict[0]) final_class = max(scores_class_dict.items(), key=lambda scores_class_dict: scores_class_dict[0])[1] answer_sheet_list.append(final_class) for i in range(len(final_for_ensemble)): for_ensemble_confidence_list.append(final_for_ensemble[i][1]) golden = [] golden_sheet = r'D:\Mammograph\golden/balance_golden_v1_852.csv' with open(golden_sheet, newline='') as csvFile: T = 0 F = 0 mm = 0 bb = 0 bm = 0 mb = 0 rows = csv.reader(csvFile) for row in rows: golden.append(row[0]) for i, j in zip(golden, answer_sheet_list): if i == j: if(i=='benign'): bb += 1 else: mm += 1 T += 1 else: if (i == 'benign'): bm += 1 else: mb += 1 F += 1 accuracy = T / (T+F) accuracy_malignant = (mm + bb)/(mm + bb + bm + mb) accuracy_benign = (bb + mm)/(bb + mm + mb + bm) print('T: ', T, ' F: ', F, ' accuracy: ',accuracy, 'malignant to benign : ', mb, 'benign to malignant : ', bm) big_name = 'big_' + wb_name + '.xlsx' workbook = xlsxwriter.Workbook(big_name) worksheet = workbook.add_worksheet() row = 0 column = 0 for list_ele in big_list: for item in list_ele: worksheet.write(row, column, item[0]) column += 1 worksheet.write(row, column, item[1]) column += 1 column = 0 row += 1 workbook.close() dict_name = 'dict_' + wb_name + '.xlsx' workbook = xlsxwriter.Workbook(dict_name) worksheet = workbook.add_worksheet() row = 0 column = 0 for dict_ele in score_class_list: item_list = [] for key, value in dict_ele.items(): item_list.append(key) item_list.append(value) for item in item_list: worksheet.write(row, column, item) column += 1 column = 0 row += 1 workbook.close() ans_name = 'ans_' + wb_name + '.xlsx' workbook = xlsxwriter.Workbook(ans_name) worksheet = workbook.add_worksheet() row = 0 column = 0 for item in answer_sheet_list: worksheet.write(row, column, item) row += 1 workbook.close() integrate_name = 'integrate_' + wb_name + '.csv' with open(integrate_name, 'w', newline='') as csvfile: writer = csv.writer(csvfile) for i in range(0, len(for_ensemble_confidence_list), 2): writer.writerow(for_ensemble_confidence_list[i:i + 2]) return T, F, accuracy, mb, bm
true
true
1c4842ccf80b14547a6aafc0838a19e7f6e672cc
4,340
py
Python
src/models/hg_3d.py
DNALuo/3Dposes
c5e2ed5fea612318d7715e239176571f593ccf83
[ "MIT" ]
null
null
null
src/models/hg_3d.py
DNALuo/3Dposes
c5e2ed5fea612318d7715e239176571f593ccf83
[ "MIT" ]
null
null
null
src/models/hg_3d.py
DNALuo/3Dposes
c5e2ed5fea612318d7715e239176571f593ccf83
[ "MIT" ]
null
null
null
from .layers.Residual import Residual import torch.nn as nn import math import ref class Hourglass(nn.Module): def __init__(self, n, nModules, nFeats): super(Hourglass, self).__init__() self.n = n self.nModules = nModules self.nFeats = nFeats _up1_, _low1_, _low2_, _low3_ = [], [], [], [] for j in range(self.nModules): _up1_.append(Residual(self.nFeats, self.nFeats)) self.low1 = nn.MaxPool2d(kernel_size = 2, stride = 2) for j in range(self.nModules): _low1_.append(Residual(self.nFeats, self.nFeats)) if self.n > 1: self.low2 = Hourglass(n - 1, self.nModules, self.nFeats) else: for j in range(self.nModules): _low2_.append(Residual(self.nFeats, self.nFeats)) self.low2_ = nn.ModuleList(_low2_) for j in range(self.nModules): _low3_.append(Residual(self.nFeats, self.nFeats)) self.up1_ = nn.ModuleList(_up1_) self.low1_ = nn.ModuleList(_low1_) self.low3_ = nn.ModuleList(_low3_) self.up2 = nn.Upsample(scale_factor = 2) def forward(self, x): up1 = x for j in range(self.nModules): up1 = self.up1_[j](up1) low1 = self.low1(x) for j in range(self.nModules): low1 = self.low1_[j](low1) if self.n > 1: low2 = self.low2(low1) else: low2 = low1 for j in range(self.nModules): low2 = self.low2_[j](low2) low3 = low2 for j in range(self.nModules): low3 = self.low3_[j](low3) up2 = self.up2(low3) return up1 + up2 class HourglassNet3D(nn.Module): def __init__(self, nStack, nModules, nFeats, nRegModules): super(HourglassNet3D, self).__init__() self.nStack = nStack self.nModules = nModules self.nFeats = nFeats self.nRegModules = nRegModules self.conv1_ = nn.Conv2d(3, 64, bias = True, kernel_size = 7, stride = 2, padding = 3) self.bn1 = nn.BatchNorm2d(64) self.relu = nn.ReLU(inplace = True) self.r1 = Residual(64, 128) self.maxpool = nn.MaxPool2d(kernel_size = 2, stride = 2) self.r4 = Residual(128, 128) self.r5 = Residual(128, self.nFeats) _hourglass, _Residual, _lin_, _tmpOut, _ll_, _tmpOut_, _reg_ = [], [], [], [], [], [], [] for i in range(self.nStack): _hourglass.append(Hourglass(4, self.nModules, self.nFeats)) for j in range(self.nModules): _Residual.append(Residual(self.nFeats, self.nFeats)) lin = nn.Sequential(nn.Conv2d(self.nFeats, self.nFeats, bias = True, kernel_size = 1, stride = 1), nn.BatchNorm2d(self.nFeats), self.relu) _lin_.append(lin) _tmpOut.append(nn.Conv2d(self.nFeats, ref.nJoints, bias = True, kernel_size = 1, stride = 1)) _ll_.append(nn.Conv2d(self.nFeats, self.nFeats, bias = True, kernel_size = 1, stride = 1)) _tmpOut_.append(nn.Conv2d(ref.nJoints, self.nFeats, bias = True, kernel_size = 1, stride = 1)) for i in range(4): for j in range(self.nRegModules): _reg_.append(Residual(self.nFeats, self.nFeats)) self.hourglass = nn.ModuleList(_hourglass) self.Residual = nn.ModuleList(_Residual) self.lin_ = nn.ModuleList(_lin_) self.tmpOut = nn.ModuleList(_tmpOut) self.ll_ = nn.ModuleList(_ll_) self.tmpOut_ = nn.ModuleList(_tmpOut_) self.reg_ = nn.ModuleList(_reg_) self.reg = nn.Linear(4 * 4 * self.nFeats, ref.nJoints) def forward(self, x): x = self.conv1_(x) x = self.bn1(x) x = self.relu(x) x = self.r1(x) x = self.maxpool(x) x = self.r4(x) x = self.r5(x) out = [] for i in range(self.nStack): hg = self.hourglass[i](x) ll = hg for j in range(self.nModules): ll = self.Residual[i * self.nModules + j](ll) ll = self.lin_[i](ll) tmpOut = self.tmpOut[i](ll) out.append(tmpOut) ll_ = self.ll_[i](ll) tmpOut_ = self.tmpOut_[i](tmpOut) x = x + ll_ + tmpOut_ for i in range(4): for j in range(self.nRegModules): x = self.reg_[i * self.nRegModules + j](x) x = self.maxpool(x) x = x.view(x.size(0), -1) reg = self.reg(x) out.append(reg) return out
32.148148
106
0.589862
from .layers.Residual import Residual import torch.nn as nn import math import ref class Hourglass(nn.Module): def __init__(self, n, nModules, nFeats): super(Hourglass, self).__init__() self.n = n self.nModules = nModules self.nFeats = nFeats _up1_, _low1_, _low2_, _low3_ = [], [], [], [] for j in range(self.nModules): _up1_.append(Residual(self.nFeats, self.nFeats)) self.low1 = nn.MaxPool2d(kernel_size = 2, stride = 2) for j in range(self.nModules): _low1_.append(Residual(self.nFeats, self.nFeats)) if self.n > 1: self.low2 = Hourglass(n - 1, self.nModules, self.nFeats) else: for j in range(self.nModules): _low2_.append(Residual(self.nFeats, self.nFeats)) self.low2_ = nn.ModuleList(_low2_) for j in range(self.nModules): _low3_.append(Residual(self.nFeats, self.nFeats)) self.up1_ = nn.ModuleList(_up1_) self.low1_ = nn.ModuleList(_low1_) self.low3_ = nn.ModuleList(_low3_) self.up2 = nn.Upsample(scale_factor = 2) def forward(self, x): up1 = x for j in range(self.nModules): up1 = self.up1_[j](up1) low1 = self.low1(x) for j in range(self.nModules): low1 = self.low1_[j](low1) if self.n > 1: low2 = self.low2(low1) else: low2 = low1 for j in range(self.nModules): low2 = self.low2_[j](low2) low3 = low2 for j in range(self.nModules): low3 = self.low3_[j](low3) up2 = self.up2(low3) return up1 + up2 class HourglassNet3D(nn.Module): def __init__(self, nStack, nModules, nFeats, nRegModules): super(HourglassNet3D, self).__init__() self.nStack = nStack self.nModules = nModules self.nFeats = nFeats self.nRegModules = nRegModules self.conv1_ = nn.Conv2d(3, 64, bias = True, kernel_size = 7, stride = 2, padding = 3) self.bn1 = nn.BatchNorm2d(64) self.relu = nn.ReLU(inplace = True) self.r1 = Residual(64, 128) self.maxpool = nn.MaxPool2d(kernel_size = 2, stride = 2) self.r4 = Residual(128, 128) self.r5 = Residual(128, self.nFeats) _hourglass, _Residual, _lin_, _tmpOut, _ll_, _tmpOut_, _reg_ = [], [], [], [], [], [], [] for i in range(self.nStack): _hourglass.append(Hourglass(4, self.nModules, self.nFeats)) for j in range(self.nModules): _Residual.append(Residual(self.nFeats, self.nFeats)) lin = nn.Sequential(nn.Conv2d(self.nFeats, self.nFeats, bias = True, kernel_size = 1, stride = 1), nn.BatchNorm2d(self.nFeats), self.relu) _lin_.append(lin) _tmpOut.append(nn.Conv2d(self.nFeats, ref.nJoints, bias = True, kernel_size = 1, stride = 1)) _ll_.append(nn.Conv2d(self.nFeats, self.nFeats, bias = True, kernel_size = 1, stride = 1)) _tmpOut_.append(nn.Conv2d(ref.nJoints, self.nFeats, bias = True, kernel_size = 1, stride = 1)) for i in range(4): for j in range(self.nRegModules): _reg_.append(Residual(self.nFeats, self.nFeats)) self.hourglass = nn.ModuleList(_hourglass) self.Residual = nn.ModuleList(_Residual) self.lin_ = nn.ModuleList(_lin_) self.tmpOut = nn.ModuleList(_tmpOut) self.ll_ = nn.ModuleList(_ll_) self.tmpOut_ = nn.ModuleList(_tmpOut_) self.reg_ = nn.ModuleList(_reg_) self.reg = nn.Linear(4 * 4 * self.nFeats, ref.nJoints) def forward(self, x): x = self.conv1_(x) x = self.bn1(x) x = self.relu(x) x = self.r1(x) x = self.maxpool(x) x = self.r4(x) x = self.r5(x) out = [] for i in range(self.nStack): hg = self.hourglass[i](x) ll = hg for j in range(self.nModules): ll = self.Residual[i * self.nModules + j](ll) ll = self.lin_[i](ll) tmpOut = self.tmpOut[i](ll) out.append(tmpOut) ll_ = self.ll_[i](ll) tmpOut_ = self.tmpOut_[i](tmpOut) x = x + ll_ + tmpOut_ for i in range(4): for j in range(self.nRegModules): x = self.reg_[i * self.nRegModules + j](x) x = self.maxpool(x) x = x.view(x.size(0), -1) reg = self.reg(x) out.append(reg) return out
true
true
1c484385a54e0922af6c19242e523fb1932ef401
698
py
Python
src/sadie/airr/airrtable/airrseries.py
jwillis0720/pybody
2d7c68650ac1ef5f3003ccb67171898eac1f63eb
[ "MIT" ]
null
null
null
src/sadie/airr/airrtable/airrseries.py
jwillis0720/pybody
2d7c68650ac1ef5f3003ccb67171898eac1f63eb
[ "MIT" ]
null
null
null
src/sadie/airr/airrtable/airrseries.py
jwillis0720/pybody
2d7c68650ac1ef5f3003ccb67171898eac1f63eb
[ "MIT" ]
null
null
null
from typing import Any import pandas as pd from sadie.airr.models import AirrSeriesModel class AirrSeries(pd.Series): _metadata = ["meta"] # add custom namespaces here def __init__(self, data: Any, copy: bool = False): super(AirrSeries, self).__init__(data=data, copy=copy) if not isinstance(data, pd.core.internals.managers.SingleBlockManager): if isinstance(data, pd.core.series.Series): self._verify() @property def _constructor(self) -> "AirrSeries": return AirrSeries # type: ignore[return-value] def _verify(self) -> None: data = AirrSeriesModel(**self).dict() # type: ignore self.update(data)
29.083333
79
0.659026
from typing import Any import pandas as pd from sadie.airr.models import AirrSeriesModel class AirrSeries(pd.Series): _metadata = ["meta"] def __init__(self, data: Any, copy: bool = False): super(AirrSeries, self).__init__(data=data, copy=copy) if not isinstance(data, pd.core.internals.managers.SingleBlockManager): if isinstance(data, pd.core.series.Series): self._verify() @property def _constructor(self) -> "AirrSeries": return AirrSeries def _verify(self) -> None: data = AirrSeriesModel(**self).dict() self.update(data)
true
true
1c48461a6c4205aa58ec966c45311939425186de
11,551
py
Python
my_app/blog/models.py
Faisal-Sey/official1
49af7a9fd60c980bd5d4ef7075a4c1f27ecc9642
[ "MIT" ]
1
2021-06-19T00:17:02.000Z
2021-06-19T00:17:02.000Z
my_app/blog/models.py
Faisal-Sey/official1
49af7a9fd60c980bd5d4ef7075a4c1f27ecc9642
[ "MIT" ]
null
null
null
my_app/blog/models.py
Faisal-Sey/official1
49af7a9fd60c980bd5d4ef7075a4c1f27ecc9642
[ "MIT" ]
null
null
null
from django.db import models from django.shortcuts import reverse, get_object_or_404 from django.conf import settings from django_countries.fields import CountryField # Create your models here. class UserDb(models.Model): Name = models.CharField(max_length=30) Email = models.EmailField(max_length=200) Message = models.CharField(max_length=500) def __str__(self): return '{}'.format(self.Message) class item(models.Model): items_name = models.CharField(max_length=300) price = models.IntegerField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) class LatestProduct(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class LatestProductOne(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class TopProduct(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class TopProductOne(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class ReviewProduct(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class ReviewProductOne(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) PRODUCT_CHOICE = { 'Jordan': 'Jordan', 'nike': 'Nike', 'man': 'Man', } class ShopMain(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.title) class Shoes(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Shoes' class Watches(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Watches' class Slippers(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Slippers' class Shorts(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Shorts' class Pants(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Pants' class Shirts(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Shirts' class OrderItem(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE, blank=True, null=True) ordered = models.BooleanField(default=False) item = models.ForeignKey(Shoes, on_delete=models.CASCADE) quantity = models.FloatField(default=1) objects = models.Manager() def __str__(self): return f"{self.quantity} of {self.item.title}" def get_total_item_price(self): return self.quantity * self.item.price class Order(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) items = models.ManyToManyField(OrderItem) start_date = models.DateTimeField(auto_now=True) ordered_date = models.DateTimeField() ordered = models.BooleanField(default=False) objects = models.Manager() billing_address = models.ForeignKey('BillingAddress', on_delete=models.SET_NULL, blank=True, null=True) def __str__(self): return self.user.username def get_total(self): total = 0 for order_item in self.items.all(): total += order_item.get_total_item_price() return total def get_sub_total(self): subtotal = 0 for order_item in self.items.all(): subtotal += order_item.item.price return subtotal class BillingAddress(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) First_name = models.CharField(max_length=100) Last_name = models.CharField(max_length=100) Email = models.EmailField() Country = CountryField(multiple=True) street_address = models.CharField(max_length=100) Apartment_address = models.CharField(max_length=100) Town_or_City = models.CharField(max_length=100) Zip = models.CharField(max_length=100) Phone = models.IntegerField() def __str__(self): return self.user.username class Post(models.Model): title = models.TextField(max_length=800) slug = models.SlugField(blank=True) class PostForm(models.Model): title = models.TextField(max_length=800) answers = models.TextField(max_length=800) slug = models.SlugField(blank=True)
27.372038
107
0.65293
from django.db import models from django.shortcuts import reverse, get_object_or_404 from django.conf import settings from django_countries.fields import CountryField class UserDb(models.Model): Name = models.CharField(max_length=30) Email = models.EmailField(max_length=200) Message = models.CharField(max_length=500) def __str__(self): return '{}'.format(self.Message) class item(models.Model): items_name = models.CharField(max_length=300) price = models.IntegerField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) class LatestProduct(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class LatestProductOne(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class TopProduct(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class TopProductOne(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class ReviewProduct(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) class ReviewProductOne(models.Model): items_name = models.CharField(max_length=300) price = models.FloatField() Description = models.TextField(max_length=600, blank=True) Image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.items_name) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) PRODUCT_CHOICE = { 'Jordan': 'Jordan', 'nike': 'Nike', 'man': 'Man', } class ShopMain(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() objects = models.Manager() def __str__(self): return '{}'.format(self.title) class Shoes(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Shoes' class Watches(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Watches' class Slippers(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Slippers' class Shorts(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Shorts' class Pants(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Pants' class Shirts(models.Model): title = models.CharField(max_length=100) price = models.FloatField() image = models.ImageField() slug = models.SlugField() description = models.TextField(max_length=3000) image1 = models.ImageField(blank=True) image2 = models.ImageField(blank=True) image3 = models.ImageField(blank=True) image4 = models.ImageField(blank=True) objects = models.Manager() def __str__(self): return '{}'.format(self.title) def get_absolute_url(self): return reverse("shop-details", kwargs={ 'slug': self.slug }) def get_add_to_cart(self): return reverse("add_to_cart", kwargs={ 'slug': self.slug }) def get_remove_from_cart(self): return reverse("remove_from_cart", kwargs={ 'slug': self.slug }) class Meta: verbose_name_plural = 'Shirts' class OrderItem(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE, blank=True, null=True) ordered = models.BooleanField(default=False) item = models.ForeignKey(Shoes, on_delete=models.CASCADE) quantity = models.FloatField(default=1) objects = models.Manager() def __str__(self): return f"{self.quantity} of {self.item.title}" def get_total_item_price(self): return self.quantity * self.item.price class Order(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) items = models.ManyToManyField(OrderItem) start_date = models.DateTimeField(auto_now=True) ordered_date = models.DateTimeField() ordered = models.BooleanField(default=False) objects = models.Manager() billing_address = models.ForeignKey('BillingAddress', on_delete=models.SET_NULL, blank=True, null=True) def __str__(self): return self.user.username def get_total(self): total = 0 for order_item in self.items.all(): total += order_item.get_total_item_price() return total def get_sub_total(self): subtotal = 0 for order_item in self.items.all(): subtotal += order_item.item.price return subtotal class BillingAddress(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) First_name = models.CharField(max_length=100) Last_name = models.CharField(max_length=100) Email = models.EmailField() Country = CountryField(multiple=True) street_address = models.CharField(max_length=100) Apartment_address = models.CharField(max_length=100) Town_or_City = models.CharField(max_length=100) Zip = models.CharField(max_length=100) Phone = models.IntegerField() def __str__(self): return self.user.username class Post(models.Model): title = models.TextField(max_length=800) slug = models.SlugField(blank=True) class PostForm(models.Model): title = models.TextField(max_length=800) answers = models.TextField(max_length=800) slug = models.SlugField(blank=True)
true
true
1c4846dac0af0ccb197b9ac571341196be451a99
619
py
Python
molecule/tang/tests/test_creation.py
stackhpc/ansible-role-luks
8c4b5f472ab0aef3d2a776d4fcd37ca17c6eac05
[ "Apache-1.1" ]
3
2020-04-14T19:57:25.000Z
2021-01-11T09:09:16.000Z
molecule/tang/tests/test_creation.py
stackhpc/ansible-role-luks
8c4b5f472ab0aef3d2a776d4fcd37ca17c6eac05
[ "Apache-1.1" ]
4
2020-08-12T10:24:25.000Z
2022-01-17T17:48:28.000Z
molecule/tang/tests/test_creation.py
stackhpc/ansible-role-luks
8c4b5f472ab0aef3d2a776d4fcd37ca17c6eac05
[ "Apache-1.1" ]
2
2021-06-17T21:57:42.000Z
2022-02-20T08:02:43.000Z
import os import pytest import testinfra.utils.ansible_runner testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner( os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all') def test_crypto_devices(host): f = host.file('/dev/mapper/cryptotest') assert f.exists def test_key_files_exist(host): f = host.file('/etc/luks-keys/dev-vdb') assert not f.exists @pytest.mark.parametrize('file, content', [ ("/etc/crypttab", "cryptotest /dev/vdb none _netdev"), ]) def test_crypttab(host, file, content): file = host.file(file) assert file.exists assert file.contains(content)
22.925926
63
0.723748
import os import pytest import testinfra.utils.ansible_runner testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner( os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all') def test_crypto_devices(host): f = host.file('/dev/mapper/cryptotest') assert f.exists def test_key_files_exist(host): f = host.file('/etc/luks-keys/dev-vdb') assert not f.exists @pytest.mark.parametrize('file, content', [ ("/etc/crypttab", "cryptotest /dev/vdb none _netdev"), ]) def test_crypttab(host, file, content): file = host.file(file) assert file.exists assert file.contains(content)
true
true
1c48470b6af2b5dacfff0ceab70a8a5b3cef97d3
23,230
py
Python
tools/imagenet-tfrecords-builder/build_dataset.py
isabella232/heldout-influence-estimation
634527bf7ca6630e6fe66867347747e2e04bc780
[ "Apache-2.0" ]
43
2020-09-11T23:40:16.000Z
2022-03-10T02:14:32.000Z
tools/imagenet-tfrecords-builder/build_dataset.py
google-research/heldout-influence-estimation
634527bf7ca6630e6fe66867347747e2e04bc780
[ "Apache-2.0" ]
1
2022-01-16T13:01:16.000Z
2022-01-16T13:01:16.000Z
tools/imagenet-tfrecords-builder/build_dataset.py
isabella232/heldout-influence-estimation
634527bf7ca6630e6fe66867347747e2e04bc780
[ "Apache-2.0" ]
5
2020-11-16T10:34:08.000Z
2022-03-20T04:42:39.000Z
# Copyright 2016 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== # Adapted from https://github.com/kmonachopoulos/ImageNet-to-TFrecord """Converts ImageNet data to TFRecords file format with Example protos. The raw ImageNet data set is expected to reside in JPEG files located in the following directory structure. data_dir/n01440764/ILSVRC2012_val_00000293.JPEG data_dir/n01440764/ILSVRC2012_val_00000543.JPEG ... where 'n01440764' is the unique synset label associated with these images. The training data set consists of 1000 sub-directories (i.e. labels) each containing 1200 JPEG images for a total of 1.2M JPEG images. The evaluation data set consists of 1000 sub-directories (i.e. labels) each containing 50 JPEG images for a total of 50K JPEG images. This TensorFlow script converts the training and evaluation data into a sharded data set consisting of 1024 and 128 TFRecord files, respectively. train_directory/train-00000-of-01024 train_directory/train-00001-of-01024 ... train_directory/train-00127-of-01024 and validation_directory/validation-00000-of-00128 validation_directory/validation-00001-of-00128 ... validation_directory/validation-00127-of-00128 Each validation TFRecord file contains ~390 records. Each training TFREcord file contains ~1250 records. Each record within the TFRecord file is a serialized Example proto. The Example proto contains the following fields: image/encoded: string containing JPEG encoded image in RGB colorspace image/height: integer, image height in pixels image/width: integer, image width in pixels image/colorspace: string, specifying the colorspace, always 'RGB' image/channels: integer, specifying the number of channels, always 3 image/format: string, specifying the format, always'JPEG' image/filename: string containing the basename of the image file e.g. 'n01440764_10026.JPEG' or 'ILSVRC2012_val_00000293.JPEG' image/class/label: integer specifying the index in a classification layer. The label ranges from [1, 1000] where 0 is not used. image/class/synset: string specifying the unique ID of the label, e.g. 'n01440764' image/class/text: string specifying the human-readable version of the label e.g. 'red fox, Vulpes vulpes' Note that the length of xmin is identical to the length of xmax, ymin and ymax for each example. Running this script using 16 threads may take around ~2.5 hours on a HP Z420. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function from datetime import datetime import os import random import sys import threading import numpy as np from six.moves import xrange # pylint: disable=redefined-builtin import tensorflow as tf import absl absl.flags.DEFINE_string('train_directory', None, 'Training data directory') absl.flags.DEFINE_string('validation_directory', None, 'Validation data directory') absl.flags.DEFINE_string('output_directory', None, 'Output data directory') absl.flags.DEFINE_integer('train_shards', 1024, 'Number of shards in training TFRecord files.') absl.flags.DEFINE_integer('validation_shards', 128, 'Number of shards in validation TFRecord files.') absl.flags.DEFINE_integer('num_threads', 1, 'Number of threads to preprocess the images.') # The labels file contains a list of valid labels are held in this file. # Assumes that the file contains entries as such: # n01440764 # n01443537 # n01484850 # where each line corresponds to a label expressed as a synset. We map # each synset contained in the file to an integer (based on the alphabetical # ordering). See below for details. absl.flags.DEFINE_string('labels_file', 'imagenet_lsvrc_2015_synsets.txt', 'Labels file') # This file containing mapping from synset to human-readable label. # Assumes each line of the file looks like: # # n02119247 black fox # n02119359 silver fox # n02119477 red fox, Vulpes fulva # # where each line corresponds to a unique mapping. Note that each line is # formatted as <synset>\t<human readable label>. absl.flags.DEFINE_string('imagenet_metadata_file', 'imagenet_metadata.txt', 'ImageNet metadata file') # ImageNet Index used in https://pluskid.github.io/influence-memorization/ absl.flags.DEFINE_string('imagenet_index_file', 'imagenet_index.npz', 'ImageNet Example Index.') FLAGS = absl.flags.FLAGS def _int64_feature(value): """Wrapper for inserting int64 features into Example proto.""" if not isinstance(value, list): value = [value] return tf.train.Feature(int64_list=tf.train.Int64List(value=value)) def _float_feature(value): """Wrapper for inserting float features into Example proto.""" if not isinstance(value, list): value = [value] return tf.train.Feature(float_list=tf.train.FloatList(value=value)) def _bytes_feature(value): """Wrapper for inserting bytes features into Example proto.""" return tf.train.Feature(bytes_list=tf.train.BytesList(value=[value])) def _convert_to_example(filename, image_buffer, label, index, synset, human, height, width): """Build an Example proto for an example. Args: filename: string, path to an image file, e.g., '/path/to/example.JPG' image_buffer: string, JPEG encoding of RGB image label: integer, identifier for the ground truth for the network index: integer, example index synset: string, unique WordNet ID specifying the label, e.g., 'n02323233' human: string, human-readable label, e.g., 'red fox, Vulpes vulpes' height: integer, image height in pixels width: integer, image width in pixels Returns: Example proto """ colorspace = b'RGB' channels = 3 image_format = b'JPEG' example = tf.train.Example(features=tf.train.Features(feature={ 'index': _int64_feature(index), 'image/height': _int64_feature(height), 'image/width': _int64_feature(width), 'image/colorspace': _bytes_feature(colorspace), 'image/channels': _int64_feature(channels), 'image/class/label': _int64_feature(label), 'image/class/synset': _bytes_feature(bytes(synset,'utf-8')), 'image/class/text': _bytes_feature(bytes(human,'utf-8')), 'image/format': _bytes_feature(image_format), 'image/filename': _bytes_feature(bytes(os.path.basename(filename),'utf-8')), 'image/encoded': _bytes_feature(image_buffer)})) return example class ImageCoder(object): """Helper class that provides TensorFlow image coding utilities.""" # def __init__(self): # # Create a single Session to run all image coding calls. # self._sess = tf.Session() # # Initializes function that converts PNG to JPEG data. # self._png_data = tf.placeholder(dtype=tf.string) # image = tf.image.decode_png(self._png_data, channels=3) # self._png_to_jpeg = tf.image.encode_jpeg(image, format='rgb', quality=100) # # Initializes function that converts CMYK JPEG data to RGB JPEG data. # self._cmyk_data = tf.placeholder(dtype=tf.string) # image = tf.image.decode_jpeg(self._cmyk_data, channels=0) # self._cmyk_to_rgb = tf.image.encode_jpeg(image, format='rgb', quality=100) # # Initializes function that decodes RGB JPEG data. # self._decode_jpeg_data = tf.placeholder(dtype=tf.string) # self._decode_jpeg = tf.image.decode_jpeg(self._decode_jpeg_data, channels=3) def png_to_jpeg(self, image_data): image = tf.image.decode_png(image_data, channels=3) return tf.image.encode_jpeg(image, format='rgb', quality=100).numpy() # return self._sess.run(self._png_to_jpeg, # feed_dict={self._png_data: image_data}) def cmyk_to_rgb(self, image_data): image = tf.image.decode_jpeg(image_data, channels=0) return tf.image.encode_jpeg(image, format='rgb', quality=100).numpy() # return self._sess.run(self._cmyk_to_rgb, # feed_dict={self._cmyk_data: image_data}) def decode_jpeg(self, image_data): image = tf.image.decode_jpeg(image_data, channels=3).numpy() # image = self._sess.run(self._decode_jpeg, # feed_dict={self._decode_jpeg_data: image_data}) assert len(image.shape) == 3 assert image.shape[2] == 3 return image def _is_png(filename): """Determine if a file contains a PNG format image. Args: filename: string, path of the image file. Returns: boolean indicating if the image is a PNG. """ # File list from: # https://groups.google.com/forum/embed/?place=forum/torch7#!topic/torch7/fOSTXHIESSU return 'n02105855_2933.JPEG' in filename def _is_cmyk(filename): """Determine if file contains a CMYK JPEG format image. Args: filename: string, path of the image file. Returns: boolean indicating if the image is a JPEG encoded with CMYK color space. """ # File list from: # https://github.com/cytsai/ilsvrc-cmyk-image-list blacklist = ['n01739381_1309.JPEG', 'n02077923_14822.JPEG', 'n02447366_23489.JPEG', 'n02492035_15739.JPEG', 'n02747177_10752.JPEG', 'n03018349_4028.JPEG', 'n03062245_4620.JPEG', 'n03347037_9675.JPEG', 'n03467068_12171.JPEG', 'n03529860_11437.JPEG', 'n03544143_17228.JPEG', 'n03633091_5218.JPEG', 'n03710637_5125.JPEG', 'n03961711_5286.JPEG', 'n04033995_2932.JPEG', 'n04258138_17003.JPEG', 'n04264628_27969.JPEG', 'n04336792_7448.JPEG', 'n04371774_5854.JPEG', 'n04596742_4225.JPEG', 'n07583066_647.JPEG', 'n13037406_4650.JPEG'] return filename.split('/')[-1] in blacklist def _process_image(filename, coder): """Process a single image file. Args: filename: string, path to an image file e.g., '/path/to/example.JPG'. coder: instance of ImageCoder to provide TensorFlow image coding utils. Returns: image_buffer: string, JPEG encoding of RGB image. height: integer, image height in pixels. width: integer, image width in pixels. """ # Read the image file. image_data = tf.io.gfile.GFile(filename, 'rb').read() # Clean the dirty data. if _is_png(filename): # 1 image is a PNG. print('Converting PNG to JPEG for %s' % filename) image_data = coder.png_to_jpeg(image_data) elif _is_cmyk(filename): # 22 JPEG images are in CMYK colorspace. print('Converting CMYK to RGB for %s' % filename) image_data = coder.cmyk_to_rgb(image_data) # Decode the RGB JPEG. image = coder.decode_jpeg(image_data) # Check that image converted to RGB assert len(image.shape) == 3 height = image.shape[0] width = image.shape[1] assert image.shape[2] == 3 return image_data, height, width def _process_image_files_batch(coder, thread_index, ranges, name, filenames, synsets, labels, idxs, humans, num_shards): """Processes and saves list of images as TFRecord in 1 thread. Args: coder: instance of ImageCoder to provide TensorFlow image coding utils. thread_index: integer, unique batch to run index is within [0, len(ranges)). ranges: list of pairs of integers specifying ranges of each batches to analyze in parallel. name: string, unique identifier specifying the data set filenames: list of strings; each string is a path to an image file synsets: list of strings; each string is a unique WordNet ID labels: list of integer; each integer identifies the ground truth idxs: list of integer; each integre identifies the example index humans: list of strings; each string is a human-readable label num_shards: integer number of shards for this data set. """ # Each thread produces N shards where N = int(num_shards / num_threads). # For instance, if num_shards = 128, and the num_threads = 2, then the first # thread would produce shards [0, 64). num_threads = len(ranges) assert not num_shards % num_threads num_shards_per_batch = int(num_shards / num_threads) shard_ranges = np.linspace(ranges[thread_index][0], ranges[thread_index][1], num_shards_per_batch + 1).astype(int) num_files_in_thread = ranges[thread_index][1] - ranges[thread_index][0] counter = 0 for s in xrange(num_shards_per_batch): # Generate a sharded version of the file name, e.g. 'train-00002-of-00010' shard = thread_index * num_shards_per_batch + s output_filename = '%s-%.5d-of-%.5d' % (name, shard, num_shards) output_file = os.path.join(FLAGS.output_directory, output_filename) writer = tf.io.TFRecordWriter(output_file) shard_counter = 0 files_in_shard = np.arange(shard_ranges[s], shard_ranges[s + 1], dtype=int) # HERE for i in files_in_shard: filename = filenames[i] label = labels[i] idx = idxs[i] synset = synsets[i] human = humans[i] image_buffer, height, width = _process_image(filename, coder) example = _convert_to_example(filename, image_buffer, label, idx, synset, human, height, width) writer.write(example.SerializeToString()) shard_counter += 1 counter += 1 if not counter % 1000: print('%s [thread %d]: Processed %d of %d images in thread batch.' % (datetime.now(), thread_index, counter, num_files_in_thread)) sys.stdout.flush() writer.close() print('%s [thread %d]: Wrote %d images to %s' % (datetime.now(), thread_index, shard_counter, output_file)) sys.stdout.flush() shard_counter = 0 print('%s [thread %d]: Wrote %d images to %d shards.' % (datetime.now(), thread_index, counter, num_files_in_thread)) sys.stdout.flush() def _process_image_files(name, filenames, synsets, labels, idxs, humans, num_shards): """Process and save list of images as TFRecord of Example protos. Args: name: string, unique identifier specifying the data set filenames: list of strings; each string is a path to an image file synsets: list of strings; each string is a unique WordNet ID labels: list of integer; each integer identifies the ground truth idxs: list of integer; each integer identifies the index of the example humans: list of strings; each string is a human-readable label num_shards: integer number of shards for this data set. """ assert len(filenames) == len(synsets) assert len(filenames) == len(labels) assert len(filenames) == len(idxs) assert len(filenames) == len(humans) # Break all images into batches with a [ranges[i][0], ranges[i][1]]. spacing = np.linspace(0, len(filenames), FLAGS.num_threads + 1).astype(np.int) ranges = [] threads = [] for i in xrange(len(spacing) - 1): ranges.append([spacing[i], spacing[i+1]]) # Launch a thread for each batch. print('Launching %d threads for spacings: %s' % (FLAGS.num_threads, ranges)) sys.stdout.flush() # Create a mechanism for monitoring when all threads are finished. coord = tf.train.Coordinator() # Create a generic TensorFlow-based utility for converting all image codings. coder = ImageCoder() threads = [] for thread_index in xrange(len(ranges)): args = (coder, thread_index, ranges, name, filenames, synsets, labels, idxs, humans, num_shards) t = threading.Thread(target=_process_image_files_batch, args=args) t.start() threads.append(t) # Wait for all the threads to terminate. coord.join(threads) print('%s: Finished writing all %d images in data set.' % (datetime.now(), len(filenames))) sys.stdout.flush() def _find_image_files(data_dir, labels_file, index_lookup): """Build a list of all images files and labels in the data set. Args: data_dir: string, path to the root directory of images. Assumes that the ImageNet data set resides in JPEG files located in the following directory structure. data_dir/n01440764/ILSVRC2012_val_00000293.JPEG data_dir/n01440764/ILSVRC2012_val_00000543.JPEG where 'n01440764' is the unique synset label associated with these images. labels_file: string, path to the labels file. The list of valid labels are held in this file. Assumes that the file contains entries as such: n01440764 n01443537 n01484850 where each line corresponds to a label expressed as a synset. We map each synset contained in the file to an integer (based on the alphabetical ordering) starting with the integer 1 corresponding to the synset contained in the first line. The reason we start the integer labels at 1 is to reserve label 0 as an unused background class. index_lookup: a dict maps from filename to (index, label) pair. Returns: filenames: list of strings; each string is a path to an image file. synsets: list of strings; each string is a unique WordNet ID. labels: list of integer; each integer identifies the ground truth. """ print('Determining list of input files and labels from %s.' % data_dir) challenge_synsets = [l.strip() for l in tf.io.gfile.GFile(labels_file, 'r').readlines()] labels = [] filenames = [] synsets = [] idxs = [] # Leave label index 0 empty as a background class. label_index = 1 # Construct the list of JPEG files and labels. for synset in challenge_synsets: jpeg_file_path = '%s/%s/*.JPEG' % (data_dir, synset) matching_files = tf.io.gfile.glob(jpeg_file_path) labels.extend([label_index] * len(matching_files)) synsets.extend([synset] * len(matching_files)) filenames.extend(matching_files) if not label_index % 100: print('Finished finding files in %d of %d classes.' % ( label_index, len(challenge_synsets))) label_index += 1 # Shuffle the ordering of all image files in order to guarantee # random ordering of the images with respect to label in the # saved TFRecord files. Make the randomization repeatable. shuffled_index = range(len(filenames)) random.seed(12345) random.shuffle(list(range(len(shuffled_index)))) filenames = [filenames[i] for i in shuffled_index] synsets = [synsets[i] for i in shuffled_index] labels = [labels[i] for i in shuffled_index] lookup_results = [index_lookup[os.path.basename(fn)] for fn in filenames] for i in range(len(lookup_results)): # +1 because the exported index file use 0-999 labels instead of 1-1001 assert labels[i] == lookup_results[i][1] + 1 idxs = [x[0] for x in lookup_results] print('Found %d JPEG files across %d labels inside %s.' % (len(filenames), len(challenge_synsets), data_dir)) return filenames, synsets, labels, idxs def _find_human_readable_labels(synsets, synset_to_human): """Build a list of human-readable labels. Args: synsets: list of strings; each string is a unique WordNet ID. synset_to_human: dict of synset to human labels, e.g., 'n02119022' --> 'red fox, Vulpes vulpes' Returns: List of human-readable strings corresponding to each synset. """ humans = [] for s in synsets: assert s in synset_to_human, ('Failed to find: %s' % s) humans.append(synset_to_human[s]) return humans def _process_dataset(name, directory, num_shards, synset_to_human, index_lookup): """Process a complete data set and save it as a TFRecord. Args: name: string, unique identifier specifying the data set. directory: string, root path to the data set. num_shards: integer number of shards for this data set. synset_to_human: dict of synset to human labels, e.g., 'n02119022' --> 'red fox, Vulpes vulpes' index_lookup: dict of filename to (index, label) pair. """ filenames, synsets, labels, idxs = _find_image_files( directory, FLAGS.labels_file, index_lookup) humans = _find_human_readable_labels(synsets, synset_to_human) _process_image_files(name, filenames, synsets, labels, idxs, humans, num_shards) def _build_synset_lookup(imagenet_metadata_file): """Build lookup for synset to human-readable label. Args: imagenet_metadata_file: string, path to file containing mapping from synset to human-readable label. Assumes each line of the file looks like: n02119247 black fox n02119359 silver fox n02119477 red fox, Vulpes fulva where each line corresponds to a unique mapping. Note that each line is formatted as <synset>\t<human readable label>. Returns: Dictionary of synset to human labels, such as: 'n02119022' --> 'red fox, Vulpes vulpes' """ lines = tf.io.gfile.GFile(imagenet_metadata_file, 'r').readlines() synset_to_human = {} for l in lines: if l: parts = l.strip().split('\t') assert len(parts) == 2 synset = parts[0] human = parts[1] synset_to_human[synset] = human return synset_to_human def _build_index_lookup(imagenet_index_file): """Build lookup table from filename to example index.""" index_data = np.load(imagenet_index_file, allow_pickle=True) lookups = {} for split, name in [('tr', 'train'), ('tt', 'validation')]: lookups[name] = {} for idx, (fn, label) in enumerate(zip(index_data['{}_filenames'.format(split)], index_data['{}_labels'.format(split)])): lookups[name][fn.decode('utf-8')] = (idx, label) return lookups def main(unused_argv): assert not FLAGS.train_shards % FLAGS.num_threads, ( 'Please make the FLAGS.num_threads commensurate with FLAGS.train_shards') assert not FLAGS.validation_shards % FLAGS.num_threads, ( 'Please make the FLAGS.num_threads commensurate with ' 'FLAGS.validation_shards') print('Saving results to %s' % FLAGS.output_directory) # Build a map from synset to human-readable label. synset_to_human = _build_synset_lookup(FLAGS.imagenet_metadata_file) index_lookups = _build_index_lookup(FLAGS.imagenet_index_file) if(FLAGS.train_directory != None): _process_dataset('train', FLAGS.train_directory, FLAGS.train_shards, synset_to_human, index_lookups['train']) if(FLAGS.validation_directory != None): _process_dataset('validation', FLAGS.validation_directory, FLAGS.validation_shards, synset_to_human, index_lookups['validation']) if __name__ == '__main__': absl.app.run(main)
39.845626
101
0.69957
from __future__ import absolute_import from __future__ import division from __future__ import print_function from datetime import datetime import os import random import sys import threading import numpy as np from six.moves import xrange import tensorflow as tf import absl absl.flags.DEFINE_string('train_directory', None, 'Training data directory') absl.flags.DEFINE_string('validation_directory', None, 'Validation data directory') absl.flags.DEFINE_string('output_directory', None, 'Output data directory') absl.flags.DEFINE_integer('train_shards', 1024, 'Number of shards in training TFRecord files.') absl.flags.DEFINE_integer('validation_shards', 128, 'Number of shards in validation TFRecord files.') absl.flags.DEFINE_integer('num_threads', 1, 'Number of threads to preprocess the images.') absl.flags.DEFINE_string('labels_file', 'imagenet_lsvrc_2015_synsets.txt', 'Labels file') absl.flags.DEFINE_string('imagenet_metadata_file', 'imagenet_metadata.txt', 'ImageNet metadata file') absl.flags.DEFINE_string('imagenet_index_file', 'imagenet_index.npz', 'ImageNet Example Index.') FLAGS = absl.flags.FLAGS def _int64_feature(value): if not isinstance(value, list): value = [value] return tf.train.Feature(int64_list=tf.train.Int64List(value=value)) def _float_feature(value): if not isinstance(value, list): value = [value] return tf.train.Feature(float_list=tf.train.FloatList(value=value)) def _bytes_feature(value): return tf.train.Feature(bytes_list=tf.train.BytesList(value=[value])) def _convert_to_example(filename, image_buffer, label, index, synset, human, height, width): colorspace = b'RGB' channels = 3 image_format = b'JPEG' example = tf.train.Example(features=tf.train.Features(feature={ 'index': _int64_feature(index), 'image/height': _int64_feature(height), 'image/width': _int64_feature(width), 'image/colorspace': _bytes_feature(colorspace), 'image/channels': _int64_feature(channels), 'image/class/label': _int64_feature(label), 'image/class/synset': _bytes_feature(bytes(synset,'utf-8')), 'image/class/text': _bytes_feature(bytes(human,'utf-8')), 'image/format': _bytes_feature(image_format), 'image/filename': _bytes_feature(bytes(os.path.basename(filename),'utf-8')), 'image/encoded': _bytes_feature(image_buffer)})) return example class ImageCoder(object): () def cmyk_to_rgb(self, image_data): image = tf.image.decode_jpeg(image_data, channels=0) return tf.image.encode_jpeg(image, format='rgb', quality=100).numpy() def decode_jpeg(self, image_data): image = tf.image.decode_jpeg(image_data, channels=3).numpy() assert len(image.shape) == 3 assert image.shape[2] == 3 return image def _is_png(filename): JPEG' in filename def _is_cmyk(filename): blacklist = ['n01739381_1309.JPEG', 'n02077923_14822.JPEG', 'n02447366_23489.JPEG', 'n02492035_15739.JPEG', 'n02747177_10752.JPEG', 'n03018349_4028.JPEG', 'n03062245_4620.JPEG', 'n03347037_9675.JPEG', 'n03467068_12171.JPEG', 'n03529860_11437.JPEG', 'n03544143_17228.JPEG', 'n03633091_5218.JPEG', 'n03710637_5125.JPEG', 'n03961711_5286.JPEG', 'n04033995_2932.JPEG', 'n04258138_17003.JPEG', 'n04264628_27969.JPEG', 'n04336792_7448.JPEG', 'n04371774_5854.JPEG', 'n04596742_4225.JPEG', 'n07583066_647.JPEG', 'n13037406_4650.JPEG'] return filename.split('/')[-1] in blacklist def _process_image(filename, coder): image_data = tf.io.gfile.GFile(filename, 'rb').read() if _is_png(filename): print('Converting PNG to JPEG for %s' % filename) image_data = coder.png_to_jpeg(image_data) elif _is_cmyk(filename): print('Converting CMYK to RGB for %s' % filename) image_data = coder.cmyk_to_rgb(image_data) image = coder.decode_jpeg(image_data) assert len(image.shape) == 3 height = image.shape[0] width = image.shape[1] assert image.shape[2] == 3 return image_data, height, width def _process_image_files_batch(coder, thread_index, ranges, name, filenames, synsets, labels, idxs, humans, num_shards): num_threads = len(ranges) assert not num_shards % num_threads num_shards_per_batch = int(num_shards / num_threads) shard_ranges = np.linspace(ranges[thread_index][0], ranges[thread_index][1], num_shards_per_batch + 1).astype(int) num_files_in_thread = ranges[thread_index][1] - ranges[thread_index][0] counter = 0 for s in xrange(num_shards_per_batch): shard = thread_index * num_shards_per_batch + s output_filename = '%s-%.5d-of-%.5d' % (name, shard, num_shards) output_file = os.path.join(FLAGS.output_directory, output_filename) writer = tf.io.TFRecordWriter(output_file) shard_counter = 0 files_in_shard = np.arange(shard_ranges[s], shard_ranges[s + 1], dtype=int) for i in files_in_shard: filename = filenames[i] label = labels[i] idx = idxs[i] synset = synsets[i] human = humans[i] image_buffer, height, width = _process_image(filename, coder) example = _convert_to_example(filename, image_buffer, label, idx, synset, human, height, width) writer.write(example.SerializeToString()) shard_counter += 1 counter += 1 if not counter % 1000: print('%s [thread %d]: Processed %d of %d images in thread batch.' % (datetime.now(), thread_index, counter, num_files_in_thread)) sys.stdout.flush() writer.close() print('%s [thread %d]: Wrote %d images to %s' % (datetime.now(), thread_index, shard_counter, output_file)) sys.stdout.flush() shard_counter = 0 print('%s [thread %d]: Wrote %d images to %d shards.' % (datetime.now(), thread_index, counter, num_files_in_thread)) sys.stdout.flush() def _process_image_files(name, filenames, synsets, labels, idxs, humans, num_shards): assert len(filenames) == len(synsets) assert len(filenames) == len(labels) assert len(filenames) == len(idxs) assert len(filenames) == len(humans) spacing = np.linspace(0, len(filenames), FLAGS.num_threads + 1).astype(np.int) ranges = [] threads = [] for i in xrange(len(spacing) - 1): ranges.append([spacing[i], spacing[i+1]]) print('Launching %d threads for spacings: %s' % (FLAGS.num_threads, ranges)) sys.stdout.flush() coord = tf.train.Coordinator() coder = ImageCoder() threads = [] for thread_index in xrange(len(ranges)): args = (coder, thread_index, ranges, name, filenames, synsets, labels, idxs, humans, num_shards) t = threading.Thread(target=_process_image_files_batch, args=args) t.start() threads.append(t) coord.join(threads) print('%s: Finished writing all %d images in data set.' % (datetime.now(), len(filenames))) sys.stdout.flush() def _find_image_files(data_dir, labels_file, index_lookup): print('Determining list of input files and labels from %s.' % data_dir) challenge_synsets = [l.strip() for l in tf.io.gfile.GFile(labels_file, 'r').readlines()] labels = [] filenames = [] synsets = [] idxs = [] label_index = 1 for synset in challenge_synsets: jpeg_file_path = '%s/%s/*.JPEG' % (data_dir, synset) matching_files = tf.io.gfile.glob(jpeg_file_path) labels.extend([label_index] * len(matching_files)) synsets.extend([synset] * len(matching_files)) filenames.extend(matching_files) if not label_index % 100: print('Finished finding files in %d of %d classes.' % ( label_index, len(challenge_synsets))) label_index += 1 shuffled_index = range(len(filenames)) random.seed(12345) random.shuffle(list(range(len(shuffled_index)))) filenames = [filenames[i] for i in shuffled_index] synsets = [synsets[i] for i in shuffled_index] labels = [labels[i] for i in shuffled_index] lookup_results = [index_lookup[os.path.basename(fn)] for fn in filenames] for i in range(len(lookup_results)): assert labels[i] == lookup_results[i][1] + 1 idxs = [x[0] for x in lookup_results] print('Found %d JPEG files across %d labels inside %s.' % (len(filenames), len(challenge_synsets), data_dir)) return filenames, synsets, labels, idxs def _find_human_readable_labels(synsets, synset_to_human): humans = [] for s in synsets: assert s in synset_to_human, ('Failed to find: %s' % s) humans.append(synset_to_human[s]) return humans def _process_dataset(name, directory, num_shards, synset_to_human, index_lookup): filenames, synsets, labels, idxs = _find_image_files( directory, FLAGS.labels_file, index_lookup) humans = _find_human_readable_labels(synsets, synset_to_human) _process_image_files(name, filenames, synsets, labels, idxs, humans, num_shards) def _build_synset_lookup(imagenet_metadata_file): lines = tf.io.gfile.GFile(imagenet_metadata_file, 'r').readlines() synset_to_human = {} for l in lines: if l: parts = l.strip().split('\t') assert len(parts) == 2 synset = parts[0] human = parts[1] synset_to_human[synset] = human return synset_to_human def _build_index_lookup(imagenet_index_file): index_data = np.load(imagenet_index_file, allow_pickle=True) lookups = {} for split, name in [('tr', 'train'), ('tt', 'validation')]: lookups[name] = {} for idx, (fn, label) in enumerate(zip(index_data['{}_filenames'.format(split)], index_data['{}_labels'.format(split)])): lookups[name][fn.decode('utf-8')] = (idx, label) return lookups def main(unused_argv): assert not FLAGS.train_shards % FLAGS.num_threads, ( 'Please make the FLAGS.num_threads commensurate with FLAGS.train_shards') assert not FLAGS.validation_shards % FLAGS.num_threads, ( 'Please make the FLAGS.num_threads commensurate with ' 'FLAGS.validation_shards') print('Saving results to %s' % FLAGS.output_directory) synset_to_human = _build_synset_lookup(FLAGS.imagenet_metadata_file) index_lookups = _build_index_lookup(FLAGS.imagenet_index_file) if(FLAGS.train_directory != None): _process_dataset('train', FLAGS.train_directory, FLAGS.train_shards, synset_to_human, index_lookups['train']) if(FLAGS.validation_directory != None): _process_dataset('validation', FLAGS.validation_directory, FLAGS.validation_shards, synset_to_human, index_lookups['validation']) if __name__ == '__main__': absl.app.run(main)
true
true
1c484798429c47a0c48e129b17f74d46202aa650
625
py
Python
ssr-top/ssr_monitor.py
BooAA/SSR
6f976dc30a975544cd111806ed6ffc5a760d2836
[ "BSD-3-Clause" ]
1
2021-10-03T11:56:32.000Z
2021-10-03T11:56:32.000Z
ssr-top/ssr_monitor.py
BooAA/SSR
6f976dc30a975544cd111806ed6ffc5a760d2836
[ "BSD-3-Clause" ]
null
null
null
ssr-top/ssr_monitor.py
BooAA/SSR
6f976dc30a975544cd111806ed6ffc5a760d2836
[ "BSD-3-Clause" ]
null
null
null
import os class ssr_monitor: ssr_path = "" def __init__(self, path = '/sys/kernel/rdma_rxe') -> None: self.ssr_path = path def get_qp_list(self) -> list: return os.listdir(self.ssr_path) def get_qp_counters(self, qpn) -> dict: qp_dir_path = os.path.join(self.ssr_path, str(qpn)) ret = {} try: for counter in os.listdir(qp_dir_path): with open(os.path.join(qp_dir_path, counter)) as file: for line in file: ret[counter] = int(line) except: pass return ret
25
70
0.5344
import os class ssr_monitor: ssr_path = "" def __init__(self, path = '/sys/kernel/rdma_rxe') -> None: self.ssr_path = path def get_qp_list(self) -> list: return os.listdir(self.ssr_path) def get_qp_counters(self, qpn) -> dict: qp_dir_path = os.path.join(self.ssr_path, str(qpn)) ret = {} try: for counter in os.listdir(qp_dir_path): with open(os.path.join(qp_dir_path, counter)) as file: for line in file: ret[counter] = int(line) except: pass return ret
true
true
1c484836cd45e3a21df1132cbfa29ac7fe759213
16,530
py
Python
intersight/model/appliance_data_export_policy.py
CiscoDevNet/intersight-python
04b721f37c3044646a91c185c7259edfb991557a
[ "Apache-2.0" ]
5
2021-12-16T15:13:32.000Z
2022-03-29T16:09:54.000Z
intersight/model/appliance_data_export_policy.py
CiscoDevNet/intersight-python
04b721f37c3044646a91c185c7259edfb991557a
[ "Apache-2.0" ]
4
2022-01-25T19:05:51.000Z
2022-03-29T20:18:37.000Z
intersight/model/appliance_data_export_policy.py
CiscoDevNet/intersight-python
04b721f37c3044646a91c185c7259edfb991557a
[ "Apache-2.0" ]
2
2020-07-07T15:01:08.000Z
2022-01-31T04:27:35.000Z
""" Cisco Intersight Cisco Intersight is a management platform delivered as a service with embedded analytics for your Cisco and 3rd party IT infrastructure. This platform offers an intelligent level of management that enables IT organizations to analyze, simplify, and automate their environments in more advanced ways than the prior generations of tools. Cisco Intersight provides an integrated and intuitive management experience for resources in the traditional data center as well as at the edge. With flexible deployment options to address complex security needs, getting started with Intersight is quick and easy. Cisco Intersight has deep integration with Cisco UCS and HyperFlex systems allowing for remote deployment, configuration, and ongoing maintenance. The model-based deployment works for a single system in a remote location or hundreds of systems in a data center and enables rapid, standardized configuration and deployment. It also streamlines maintaining those systems whether you are working with small or very large configurations. The Intersight OpenAPI document defines the complete set of properties that are returned in the HTTP response. From that perspective, a client can expect that no additional properties are returned, unless these properties are explicitly defined in the OpenAPI document. However, when a client uses an older version of the Intersight OpenAPI document, the server may send additional properties because the software is more recent than the client. In that case, the client may receive properties that it does not know about. Some generated SDKs perform a strict validation of the HTTP response body against the OpenAPI document. # noqa: E501 The version of the OpenAPI document: 1.0.9-4950 Contact: intersight@cisco.com Generated by: https://openapi-generator.tech """ import re # noqa: F401 import sys # noqa: F401 from intersight.model_utils import ( # noqa: F401 ApiTypeError, ModelComposed, ModelNormal, ModelSimple, cached_property, change_keys_js_to_python, convert_js_args_to_python_args, date, datetime, file_type, none_type, validate_get_composed_info, ) def lazy_import(): from intersight.model.appliance_data_export_policy_all_of import ApplianceDataExportPolicyAllOf from intersight.model.appliance_data_export_policy_relationship import ApplianceDataExportPolicyRelationship from intersight.model.display_names import DisplayNames from intersight.model.iam_account_relationship import IamAccountRelationship from intersight.model.mo_base_mo import MoBaseMo from intersight.model.mo_base_mo_relationship import MoBaseMoRelationship from intersight.model.mo_tag import MoTag from intersight.model.mo_version_context import MoVersionContext globals()['ApplianceDataExportPolicyAllOf'] = ApplianceDataExportPolicyAllOf globals()['ApplianceDataExportPolicyRelationship'] = ApplianceDataExportPolicyRelationship globals()['DisplayNames'] = DisplayNames globals()['IamAccountRelationship'] = IamAccountRelationship globals()['MoBaseMo'] = MoBaseMo globals()['MoBaseMoRelationship'] = MoBaseMoRelationship globals()['MoTag'] = MoTag globals()['MoVersionContext'] = MoVersionContext class ApplianceDataExportPolicy(ModelComposed): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. Attributes: allowed_values (dict): The key is the tuple path to the attribute and the for var_name this is (var_name,). The value is a dict with a capitalized key describing the allowed value and an allowed value. These dicts store the allowed enum values. attribute_map (dict): The key is attribute name and the value is json key in definition. discriminator_value_class_map (dict): A dict to go from the discriminator variable value to the discriminator class name. validations (dict): The key is the tuple path to the attribute and the for var_name this is (var_name,). The value is a dict that stores validations for max_length, min_length, max_items, min_items, exclusive_maximum, inclusive_maximum, exclusive_minimum, inclusive_minimum, and regex. additional_properties_type (tuple): A tuple of classes accepted as additional properties values. """ allowed_values = { ('class_id',): { 'APPLIANCE.DATAEXPORTPOLICY': "appliance.DataExportPolicy", }, ('object_type',): { 'APPLIANCE.DATAEXPORTPOLICY': "appliance.DataExportPolicy", }, } validations = { } @cached_property def additional_properties_type(): """ This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @cached_property def openapi_types(): """ This must be a method because a model may have properties that are of type self, this must run after the class is loaded Returns openapi_types (dict): The key is attribute name and the value is attribute type. """ lazy_import() return { 'class_id': (str,), # noqa: E501 'object_type': (str,), # noqa: E501 'enable': (bool,), # noqa: E501 'name': (str,), # noqa: E501 'account': (IamAccountRelationship,), # noqa: E501 'parent_config': (ApplianceDataExportPolicyRelationship,), # noqa: E501 'sub_configs': ([ApplianceDataExportPolicyRelationship], none_type,), # noqa: E501 'account_moid': (str,), # noqa: E501 'create_time': (datetime,), # noqa: E501 'domain_group_moid': (str,), # noqa: E501 'mod_time': (datetime,), # noqa: E501 'moid': (str,), # noqa: E501 'owners': ([str], none_type,), # noqa: E501 'shared_scope': (str,), # noqa: E501 'tags': ([MoTag], none_type,), # noqa: E501 'version_context': (MoVersionContext,), # noqa: E501 'ancestors': ([MoBaseMoRelationship], none_type,), # noqa: E501 'parent': (MoBaseMoRelationship,), # noqa: E501 'permission_resources': ([MoBaseMoRelationship], none_type,), # noqa: E501 'display_names': (DisplayNames,), # noqa: E501 } @cached_property def discriminator(): val = { } if not val: return None return {'class_id': val} attribute_map = { 'class_id': 'ClassId', # noqa: E501 'object_type': 'ObjectType', # noqa: E501 'enable': 'Enable', # noqa: E501 'name': 'Name', # noqa: E501 'account': 'Account', # noqa: E501 'parent_config': 'ParentConfig', # noqa: E501 'sub_configs': 'SubConfigs', # noqa: E501 'account_moid': 'AccountMoid', # noqa: E501 'create_time': 'CreateTime', # noqa: E501 'domain_group_moid': 'DomainGroupMoid', # noqa: E501 'mod_time': 'ModTime', # noqa: E501 'moid': 'Moid', # noqa: E501 'owners': 'Owners', # noqa: E501 'shared_scope': 'SharedScope', # noqa: E501 'tags': 'Tags', # noqa: E501 'version_context': 'VersionContext', # noqa: E501 'ancestors': 'Ancestors', # noqa: E501 'parent': 'Parent', # noqa: E501 'permission_resources': 'PermissionResources', # noqa: E501 'display_names': 'DisplayNames', # noqa: E501 } required_properties = set([ '_data_store', '_check_type', '_spec_property_naming', '_path_to_item', '_configuration', '_visited_composed_classes', '_composed_instances', '_var_name_to_model_instances', '_additional_properties_model_instances', ]) @convert_js_args_to_python_args def __init__(self, *args, **kwargs): # noqa: E501 """ApplianceDataExportPolicy - a model defined in OpenAPI Args: Keyword Args: class_id (str): The fully-qualified name of the instantiated, concrete type. This property is used as a discriminator to identify the type of the payload when marshaling and unmarshaling data.. defaults to "appliance.DataExportPolicy", must be one of ["appliance.DataExportPolicy", ] # noqa: E501 object_type (str): The fully-qualified name of the instantiated, concrete type. The value should be the same as the 'ClassId' property.. defaults to "appliance.DataExportPolicy", must be one of ["appliance.DataExportPolicy", ] # noqa: E501 _check_type (bool): if True, values for parameters in openapi_types will be type checked and a TypeError will be raised if the wrong type is input. Defaults to True _path_to_item (tuple/list): This is a list of keys or values to drill down to the model in received_data when deserializing a response _spec_property_naming (bool): True if the variable names in the input data are serialized names, as specified in the OpenAPI document. False if the variable names in the input data are pythonic names, e.g. snake case (default) _configuration (Configuration): the instance to use when deserializing a file_type parameter. If passed, type conversion is attempted If omitted no type conversion is done. _visited_composed_classes (tuple): This stores a tuple of classes that we have traveled through so that if we see that class again we will not use its discriminator again. When traveling through a discriminator, the composed schema that is is traveled through is added to this set. For example if Animal has a discriminator petType and we pass in "Dog", and the class Dog allOf includes Animal, we move through Animal once using the discriminator, and pick Dog. Then in Dog, we will make an instance of the Animal class but this time we won't travel through its discriminator because we passed in _visited_composed_classes = (Animal,) enable (bool): Status of the data collection mode. If the value is 'true', then data collection is enabled.. [optional] # noqa: E501 name (str): Name of the Data Export Policy.. [optional] # noqa: E501 account (IamAccountRelationship): [optional] # noqa: E501 parent_config (ApplianceDataExportPolicyRelationship): [optional] # noqa: E501 sub_configs ([ApplianceDataExportPolicyRelationship], none_type): An array of relationships to applianceDataExportPolicy resources.. [optional] # noqa: E501 account_moid (str): The Account ID for this managed object.. [optional] # noqa: E501 create_time (datetime): The time when this managed object was created.. [optional] # noqa: E501 domain_group_moid (str): The DomainGroup ID for this managed object.. [optional] # noqa: E501 mod_time (datetime): The time when this managed object was last modified.. [optional] # noqa: E501 moid (str): The unique identifier of this Managed Object instance.. [optional] # noqa: E501 owners ([str], none_type): [optional] # noqa: E501 shared_scope (str): Intersight provides pre-built workflows, tasks and policies to end users through global catalogs. Objects that are made available through global catalogs are said to have a 'shared' ownership. Shared objects are either made globally available to all end users or restricted to end users based on their license entitlement. Users can use this property to differentiate the scope (global or a specific license tier) to which a shared MO belongs.. [optional] # noqa: E501 tags ([MoTag], none_type): [optional] # noqa: E501 version_context (MoVersionContext): [optional] # noqa: E501 ancestors ([MoBaseMoRelationship], none_type): An array of relationships to moBaseMo resources.. [optional] # noqa: E501 parent (MoBaseMoRelationship): [optional] # noqa: E501 permission_resources ([MoBaseMoRelationship], none_type): An array of relationships to moBaseMo resources.. [optional] # noqa: E501 display_names (DisplayNames): [optional] # noqa: E501 """ class_id = kwargs.get('class_id', "appliance.DataExportPolicy") object_type = kwargs.get('object_type', "appliance.DataExportPolicy") _check_type = kwargs.pop('_check_type', True) _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: raise ApiTypeError( "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( args, self.__class__.__name__, ), path_to_item=_path_to_item, valid_classes=(self.__class__,), ) self._data_store = {} self._check_type = _check_type self._spec_property_naming = _spec_property_naming self._path_to_item = _path_to_item self._configuration = _configuration self._visited_composed_classes = _visited_composed_classes + (self.__class__,) constant_args = { '_check_type': _check_type, '_path_to_item': _path_to_item, '_spec_property_naming': _spec_property_naming, '_configuration': _configuration, '_visited_composed_classes': self._visited_composed_classes, } required_args = { 'class_id': class_id, 'object_type': object_type, } model_args = {} model_args.update(required_args) model_args.update(kwargs) composed_info = validate_get_composed_info( constant_args, model_args, self) self._composed_instances = composed_info[0] self._var_name_to_model_instances = composed_info[1] self._additional_properties_model_instances = composed_info[2] unused_args = composed_info[3] for var_name, var_value in required_args.items(): setattr(self, var_name, var_value) for var_name, var_value in kwargs.items(): if var_name in unused_args and \ self._configuration is not None and \ self._configuration.discard_unknown_keys and \ not self._additional_properties_model_instances: # discard variable. continue setattr(self, var_name, var_value) @cached_property def _composed_schemas(): # we need this here to make our import statements work # we must store _composed_schemas in here so the code is only run # when we invoke this method. If we kept this at the class # level we would get an error beause the class level # code would be run when this module is imported, and these composed # classes don't exist yet because their module has not finished # loading lazy_import() return { 'anyOf': [ ], 'allOf': [ ApplianceDataExportPolicyAllOf, MoBaseMo, ], 'oneOf': [ ], }
53.495146
1,678
0.642105
import re import sys from intersight.model_utils import ( ApiTypeError, ModelComposed, ModelNormal, ModelSimple, cached_property, change_keys_js_to_python, convert_js_args_to_python_args, date, datetime, file_type, none_type, validate_get_composed_info, ) def lazy_import(): from intersight.model.appliance_data_export_policy_all_of import ApplianceDataExportPolicyAllOf from intersight.model.appliance_data_export_policy_relationship import ApplianceDataExportPolicyRelationship from intersight.model.display_names import DisplayNames from intersight.model.iam_account_relationship import IamAccountRelationship from intersight.model.mo_base_mo import MoBaseMo from intersight.model.mo_base_mo_relationship import MoBaseMoRelationship from intersight.model.mo_tag import MoTag from intersight.model.mo_version_context import MoVersionContext globals()['ApplianceDataExportPolicyAllOf'] = ApplianceDataExportPolicyAllOf globals()['ApplianceDataExportPolicyRelationship'] = ApplianceDataExportPolicyRelationship globals()['DisplayNames'] = DisplayNames globals()['IamAccountRelationship'] = IamAccountRelationship globals()['MoBaseMo'] = MoBaseMo globals()['MoBaseMoRelationship'] = MoBaseMoRelationship globals()['MoTag'] = MoTag globals()['MoVersionContext'] = MoVersionContext class ApplianceDataExportPolicy(ModelComposed): allowed_values = { ('class_id',): { 'APPLIANCE.DATAEXPORTPOLICY': "appliance.DataExportPolicy", }, ('object_type',): { 'APPLIANCE.DATAEXPORTPOLICY': "appliance.DataExportPolicy", }, } validations = { } @cached_property def additional_properties_type(): lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) _nullable = False @cached_property def openapi_types(): lazy_import() return { 'class_id': (str,), 'object_type': (str,), 'enable': (bool,), 'name': (str,), 'account': (IamAccountRelationship,), 'parent_config': (ApplianceDataExportPolicyRelationship,), 'sub_configs': ([ApplianceDataExportPolicyRelationship], none_type,), 'account_moid': (str,), 'create_time': (datetime,), 'domain_group_moid': (str,), 'mod_time': (datetime,), 'moid': (str,), 'owners': ([str], none_type,), 'shared_scope': (str,), 'tags': ([MoTag], none_type,), 'version_context': (MoVersionContext,), 'ancestors': ([MoBaseMoRelationship], none_type,), 'parent': (MoBaseMoRelationship,), 'permission_resources': ([MoBaseMoRelationship], none_type,), 'display_names': (DisplayNames,), } @cached_property def discriminator(): val = { } if not val: return None return {'class_id': val} attribute_map = { 'class_id': 'ClassId', 'object_type': 'ObjectType', 'enable': 'Enable', 'name': 'Name', 'account': 'Account', 'parent_config': 'ParentConfig', 'sub_configs': 'SubConfigs', 'account_moid': 'AccountMoid', 'create_time': 'CreateTime', 'domain_group_moid': 'DomainGroupMoid', 'mod_time': 'ModTime', 'moid': 'Moid', 'owners': 'Owners', 'shared_scope': 'SharedScope', 'tags': 'Tags', 'version_context': 'VersionContext', 'ancestors': 'Ancestors', 'parent': 'Parent', 'permission_resources': 'PermissionResources', 'display_names': 'DisplayNames', } required_properties = set([ '_data_store', '_check_type', '_spec_property_naming', '_path_to_item', '_configuration', '_visited_composed_classes', '_composed_instances', '_var_name_to_model_instances', '_additional_properties_model_instances', ]) @convert_js_args_to_python_args def __init__(self, *args, **kwargs): class_id = kwargs.get('class_id', "appliance.DataExportPolicy") object_type = kwargs.get('object_type', "appliance.DataExportPolicy") _check_type = kwargs.pop('_check_type', True) _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: raise ApiTypeError( "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( args, self.__class__.__name__, ), path_to_item=_path_to_item, valid_classes=(self.__class__,), ) self._data_store = {} self._check_type = _check_type self._spec_property_naming = _spec_property_naming self._path_to_item = _path_to_item self._configuration = _configuration self._visited_composed_classes = _visited_composed_classes + (self.__class__,) constant_args = { '_check_type': _check_type, '_path_to_item': _path_to_item, '_spec_property_naming': _spec_property_naming, '_configuration': _configuration, '_visited_composed_classes': self._visited_composed_classes, } required_args = { 'class_id': class_id, 'object_type': object_type, } model_args = {} model_args.update(required_args) model_args.update(kwargs) composed_info = validate_get_composed_info( constant_args, model_args, self) self._composed_instances = composed_info[0] self._var_name_to_model_instances = composed_info[1] self._additional_properties_model_instances = composed_info[2] unused_args = composed_info[3] for var_name, var_value in required_args.items(): setattr(self, var_name, var_value) for var_name, var_value in kwargs.items(): if var_name in unused_args and \ self._configuration is not None and \ self._configuration.discard_unknown_keys and \ not self._additional_properties_model_instances: continue setattr(self, var_name, var_value) @cached_property def _composed_schemas(): # loading lazy_import() return { 'anyOf': [ ], 'allOf': [ ApplianceDataExportPolicyAllOf, MoBaseMo, ], 'oneOf': [ ], }
true
true
1c48483b923cfbba2965d9e0ab0864ff4b50c87f
15,564
py
Python
LINETCR/LineApi.py
Ricco13/SelfSiri
96b2692d3a471bafe80d7d23693c495e273480f1
[ "MIT" ]
null
null
null
LINETCR/LineApi.py
Ricco13/SelfSiri
96b2692d3a471bafe80d7d23693c495e273480f1
[ "MIT" ]
null
null
null
LINETCR/LineApi.py
Ricco13/SelfSiri
96b2692d3a471bafe80d7d23693c495e273480f1
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- from Api import Poll, Talk, channel from lib.curve.ttypes import * import requests import shutil import json from random import randint def def_callback(str): print(str) class LINE: mid = None authToken = None cert = None channel_access_token = None token = None obs_token = None refresh_token = None def __init__(self): self.Talk = Talk() self._session = requests.session() self._headers = {'X-Line-Application': 'CHROMEOS\t1.4.17\tChrome_OS\t1'} def login(self, mail=None, passwd=None, cert=None, token=None, qr=False, callback=None): if callback is None: callback = def_callback resp = self.__validate(mail,passwd,cert,token,qr) if resp == 1: self.Talk.login(mail, passwd, callback=callback) elif resp == 2: self.Talk.login(mail,passwd,cert, callback=callback) elif resp == 3: self.Talk.TokenLogin(token) elif resp == 4: self.Talk.qrLogin(callback) else: raise Exception("invalid arguments") self.authToken = self.Talk.authToken self.cert = self.Talk.cert self._headers = { 'X-Line-Application': 'CHROMEOS\t1.4.17\tChrome_OS\t1', 'X-Line-Access': self.authToken, 'User-Agent': 'Line/1.4.17' } self.Poll = Poll(self.authToken) self.channel = channel.Channel(self.authToken) self.channel.login() self.mid = self.channel.mid self.channel_access_token = self.channel.channel_access_token self.token = self.channel.token self.obs_token = self.channel.obs_token self.refresh_token = self.channel.refresh_token """User""" def getProfile(self): return self.Talk.client.getProfile() def getSettings(self): return self.Talk.client.getSettings() def getUserTicket(self): return self.Talk.client.getUserTicket() def updateProfile(self, profileObject): return self.Talk.client.updateProfile(0, profileObject) def updateSettings(self, settingObject): return self.Talk.client.updateSettings(0, settingObject) def cloneContactProfile(self, mid): contact = self.getContact(mid) profile = self.getProfile() profile.displayName = contact.displayName profile.statusMessage = contact.statusMessage profile.pictureStatus = contact.pictureStatus self.updateDisplayPicture(profile.pictureStatus) return self.updateProfile(profile) def updateDisplayPicture(self, hash_id): return self.Talk.client.updateProfileAttribute(0, 8, hash_id) """Operation""" def fetchOperation(self, revision, count): return self.Poll.client.fetchOperations(revision, count) def fetchOps(self, rev, count): return self.Poll.client.fetchOps(rev, count, 0, 0) def getLastOpRevision(self): return self.Talk.client.getLastOpRevision() def stream(self): return self.Poll.stream() """Message""" def kedapkedip(self, tomid, text): M = Message() M.to = tomid t1 = "\xf4\x80\xb0\x82\xf4\x80\xb0\x82\xf4\x80\xb0\x82\xf4\x80\xb0\x82" t2 = "\xf4\x80\x82\xb3\xf4\x8f\xbf\xbf" rst = t1 + text + t2 M.text = rst.replace("\n", " ") return self.Talk.client.sendMessage(0, M) def removeAllMessages(self, lastMessageId): return self.Talk.client.removeAllMessages(0,lastMessageId) def sendMessage(self, messageObject): return self.Talk.client.sendMessage(0,messageObject) def sendText(self, Tomid, text): msg = Message() msg.to = Tomid msg.text = text return self.Talk.client.sendMessage(0, msg) def post_content(self, url, data=None, files=None): return self._session.post(url, headers=self._headers, data=data, files=files) def sendImage(self, to_, path): M = Message(to=to_, text=None, contentType = 1) M.contentMetadata = None M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'image', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) print r if r.status_code != 201: raise Exception('Upload image failure.') return True def sendImageWithURL(self, to_, url): """Send a image with given image url :param url: image url to send """ path = 'pythonLine.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download image failure.') try: self.sendImage(to_, path) except Exception as e: raise e def sendAudioWithURL(self, to_, url): path = 'pythonLiness.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download Audio failure.') try: self.sendAudio(to_, path) except Exception as e: raise e def sendAudio(self, to_, path): M = Message(to=to_,contentType = 3) M.contentMetadata = None M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'audio', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) if r.status_code != 201: raise Exception('Upload image failure.') return True def sendVideo(self, to_, path): M = Message(to=to_,contentType = 2) M.contentMetadata = { 'VIDLEN' : '0', 'DURATION' : '0' } M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'video', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) if r.status_code != 201: raise Exception('Upload image failure.') return True def sendVideoWithURL(self, to_, url): path = 'pythonLines.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download Audio failure.') try: self.sendVideo(to_, path) except Exception as e: raise e def sendGif(self, to_, path): M = Message(to=to_,contentType = 1) M.contentMetadata = { 'VIDLEN' : '0', 'DURATION' : '0' } M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'image', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) if r.status_code != 201: raise Exception('Upload Gif failure.') return True def sendGifWithURL(self, to_, url): path = 'pythonLiness.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download Gif failure.') try: self.sendGif(to_, path) except Exception as e: raise e def sendEvent(self, messageObject): return self.Talk.client.sendEvent(0, messageObject) def sendChatChecked(self, mid, lastMessageId): return self.Talk.client.sendChatChecked(0, mid, lastMessageId) def getMessageBoxCompactWrapUp(self, mid): return self.Talk.client.getMessageBoxCompactWrapUp(mid) def getMessageBoxCompactWrapUpList(self, start, messageBox): return self.Talk.client.getMessageBoxCompactWrapUpList(start, messageBox) def getRecentMessages(self, messageBox, count): return self.Talk.client.getRecentMessages(messageBox.id, count) def getMessageBox(self, channelId, messageboxId, lastMessagesCount): return self.Talk.client.getMessageBox(channelId, messageboxId, lastMessagesCount) def getMessageBoxList(self, channelId, lastMessagesCount): return self.Talk.client.getMessageBoxList(channelId, lastMessagesCount) def getMessageBoxListByStatus(self, channelId, lastMessagesCount, status): return self.Talk.client.getMessageBoxListByStatus(channelId, lastMessagesCount, status) def getMessageBoxWrapUp(self, mid): return self.Talk.client.getMessageBoxWrapUp(mid) def getMessageBoxWrapUpList(self, start, messageBoxCount): return self.Talk.client.getMessageBoxWrapUpList(start, messageBoxCount) def getCover(self,mid): h = self.getHome(mid) objId = h["result"]["homeInfo"]["objectId"] return "http://dl.profile.line-cdn.net/myhome/c/download.nhn?userid=" + mid+ "&oid=" + objId """Contact""" def blockContact(self, mid): return self.Talk.client.blockContact(0, mid) def unblockContact(self, mid): return self.Talk.client.unblockContact(0, mid) def findAndAddContactsByMid(self, mid): return self.Talk.client.findAndAddContactsByMid(0, mid) def findAndAddContactsByMids(self, midlist): for i in midlist: self.Talk.client.findAndAddContactsByMid(0, i) def findAndAddContactsByUserid(self, userid): return self.Talk.client.findAndAddContactsByUserid(0, userid) def findContactsByUserid(self, userid): return self.Talk.client.findContactByUserid(userid) def findContactByTicket(self, ticketId): return self.Talk.client.findContactByUserTicket(ticketId) def getAllContactIds(self): return self.Talk.client.getAllContactIds() def getBlockedContactIds(self): return self.Talk.client.getBlockedContactIds() def getContact(self, mid): return self.Talk.client.getContact(mid) def getContacts(self, midlist): return self.Talk.client.getContacts(midlist) def getFavoriteMids(self): return self.Talk.client.getFavoriteMids() def getHiddenContactMids(self): return self.Talk.client.getHiddenContactMids() def CloneContactProfile(self, mid): contact = self.getContact(mid) profile = self.getProfile() profile.displayName = contact.displayName profile.statusMessage = contact.statusMessage profile.pictureStatus = contact.pictureStatus self.updateDisplayPicture(profile.pictureStatus) return self.updateProfile(profile) def updateDisplayPicture(self, hash_id): return self.Talk.client.updateProfileAttribute(0, 8, hash_id) """Group""" def findGroupByTicket(self, ticketId): return self.Talk.client.findGroupByTicket(ticketId) def acceptGroupInvitation(self, groupId): return self.Talk.client.acceptGroupInvitation(0, groupId) def acceptGroupInvitationByTicket(self, groupId, ticketId): return self.Talk.client.acceptGroupInvitationByTicket(0, groupId, ticketId) def cancelGroupInvitation(self, groupId, contactIds): return self.Talk.client.cancelGroupInvitation(0, groupId, contactIds) def createGroup(self, name, midlist): return self.Talk.client.createGroup(0, name, midlist) def getGroup(self, groupId): return self.Talk.client.getGroup(groupId) def getGroups(self, groupIds): return self.Talk.client.getGroups(groupIds) def getGroupIdsInvited(self): return self.Talk.client.getGroupIdsInvited() def getGroupIdsJoined(self): return self.Talk.client.getGroupIdsJoined() def inviteIntoGroup(self, groupId, midlist): return self.Talk.client.inviteIntoGroup(0, groupId, midlist) def kickoutFromGroup(self, groupId, midlist): return self.Talk.client.kickoutFromGroup(0, groupId, midlist) def leaveGroup(self, groupId): return self.Talk.client.leaveGroup(0, groupId) def rejectGroupInvitation(self, groupId): return self.Talk.client.rejectGroupInvitation(0, groupId) def reissueGroupTicket(self, groupId): return self.Talk.client.reissueGroupTicket(groupId) def updateGroup(self, groupObject): return self.Talk.client.updateGroup(0, groupObject) def findGroupByTicket(self,ticketId): return self.Talk.client.findGroupByTicket(0,ticketId) """Room""" def createRoom(self, midlist): return self.Talk.client.createRoom(0, midlist) def getRoom(self, roomId): return self.Talk.client.getRoom(roomId) def inviteIntoRoom(self, roomId, midlist): return self.Talk.client.inviteIntoRoom(0, roomId, midlist) def leaveRoom(self, roomId): return self.Talk.client.leaveRoom(0, roomId) """TIMELINE""" def new_post(self, text): return self.channel.new_post(text) def like(self, mid, postid, likeType=1001): return self.channel.like(mid, postid, likeType) def comment(self, mid, postid, text): return self.channel.comment(mid, postid, text) def activity(self, limit=20): return self.channel.activity(limit) def getAlbum(self, gid): return self.channel.getAlbum(gid) def changeAlbumName(self, gid, name, albumId): return self.channel.changeAlbumName(gid, name, albumId) def deleteAlbum(self, gid, albumId): return self.channel.deleteAlbum(gid,albumId) def getNote(self,gid, commentLimit, likeLimit): return self.channel.getNote(gid, commentLimit, likeLimit) def getDetail(self,mid): return self.channel.getDetail(mid) def getHome(self,mid): return self.channel.getHome(mid) def createAlbum(self, gid, name): return self.channel.createAlbum(gid,name) def createAlbum2(self, gid, name, path): return self.channel.createAlbum(gid, name, path, oid) def __validate(self, mail, passwd, cert, token, qr): if mail is not None and passwd is not None and cert is None: return 1 elif mail is not None and passwd is not None and cert is not None: return 2 elif token is not None: return 3 elif qr is True: return 4 else: return 5 def loginResult(self, callback=True): if callback is True: callback = def_callback prof = self.getProfile() print("===============[zz7_Combat]================") print(" Thanks to GOD and my friend") print("===============[© By_Ricco]================") print("mid -> " + prof.mid) print("name -> " + prof.displayName) print("authToken -> " + self.authToken) print("cert -> " + self.cert if self.cert is not None else "")
30.457926
108
0.640838
from Api import Poll, Talk, channel from lib.curve.ttypes import * import requests import shutil import json from random import randint def def_callback(str): print(str) class LINE: mid = None authToken = None cert = None channel_access_token = None token = None obs_token = None refresh_token = None def __init__(self): self.Talk = Talk() self._session = requests.session() self._headers = {'X-Line-Application': 'CHROMEOS\t1.4.17\tChrome_OS\t1'} def login(self, mail=None, passwd=None, cert=None, token=None, qr=False, callback=None): if callback is None: callback = def_callback resp = self.__validate(mail,passwd,cert,token,qr) if resp == 1: self.Talk.login(mail, passwd, callback=callback) elif resp == 2: self.Talk.login(mail,passwd,cert, callback=callback) elif resp == 3: self.Talk.TokenLogin(token) elif resp == 4: self.Talk.qrLogin(callback) else: raise Exception("invalid arguments") self.authToken = self.Talk.authToken self.cert = self.Talk.cert self._headers = { 'X-Line-Application': 'CHROMEOS\t1.4.17\tChrome_OS\t1', 'X-Line-Access': self.authToken, 'User-Agent': 'Line/1.4.17' } self.Poll = Poll(self.authToken) self.channel = channel.Channel(self.authToken) self.channel.login() self.mid = self.channel.mid self.channel_access_token = self.channel.channel_access_token self.token = self.channel.token self.obs_token = self.channel.obs_token self.refresh_token = self.channel.refresh_token """User""" def getProfile(self): return self.Talk.client.getProfile() def getSettings(self): return self.Talk.client.getSettings() def getUserTicket(self): return self.Talk.client.getUserTicket() def updateProfile(self, profileObject): return self.Talk.client.updateProfile(0, profileObject) def updateSettings(self, settingObject): return self.Talk.client.updateSettings(0, settingObject) def cloneContactProfile(self, mid): contact = self.getContact(mid) profile = self.getProfile() profile.displayName = contact.displayName profile.statusMessage = contact.statusMessage profile.pictureStatus = contact.pictureStatus self.updateDisplayPicture(profile.pictureStatus) return self.updateProfile(profile) def updateDisplayPicture(self, hash_id): return self.Talk.client.updateProfileAttribute(0, 8, hash_id) """Operation""" def fetchOperation(self, revision, count): return self.Poll.client.fetchOperations(revision, count) def fetchOps(self, rev, count): return self.Poll.client.fetchOps(rev, count, 0, 0) def getLastOpRevision(self): return self.Talk.client.getLastOpRevision() def stream(self): return self.Poll.stream() """Message""" def kedapkedip(self, tomid, text): M = Message() M.to = tomid t1 = "\xf4\x80\xb0\x82\xf4\x80\xb0\x82\xf4\x80\xb0\x82\xf4\x80\xb0\x82" t2 = "\xf4\x80\x82\xb3\xf4\x8f\xbf\xbf" rst = t1 + text + t2 M.text = rst.replace("\n", " ") return self.Talk.client.sendMessage(0, M) def removeAllMessages(self, lastMessageId): return self.Talk.client.removeAllMessages(0,lastMessageId) def sendMessage(self, messageObject): return self.Talk.client.sendMessage(0,messageObject) def sendText(self, Tomid, text): msg = Message() msg.to = Tomid msg.text = text return self.Talk.client.sendMessage(0, msg) def post_content(self, url, data=None, files=None): return self._session.post(url, headers=self._headers, data=data, files=files) def sendImage(self, to_, path): M = Message(to=to_, text=None, contentType = 1) M.contentMetadata = None M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'image', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) print r if r.status_code != 201: raise Exception('Upload image failure.') return True def sendImageWithURL(self, to_, url): """Send a image with given image url :param url: image url to send """ path = 'pythonLine.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download image failure.') try: self.sendImage(to_, path) except Exception as e: raise e def sendAudioWithURL(self, to_, url): path = 'pythonLiness.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download Audio failure.') try: self.sendAudio(to_, path) except Exception as e: raise e def sendAudio(self, to_, path): M = Message(to=to_,contentType = 3) M.contentMetadata = None M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'audio', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) if r.status_code != 201: raise Exception('Upload image failure.') return True def sendVideo(self, to_, path): M = Message(to=to_,contentType = 2) M.contentMetadata = { 'VIDLEN' : '0', 'DURATION' : '0' } M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'video', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) if r.status_code != 201: raise Exception('Upload image failure.') return True def sendVideoWithURL(self, to_, url): path = 'pythonLines.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download Audio failure.') try: self.sendVideo(to_, path) except Exception as e: raise e def sendGif(self, to_, path): M = Message(to=to_,contentType = 1) M.contentMetadata = { 'VIDLEN' : '0', 'DURATION' : '0' } M.contentPreview = None M_id = self.Talk.client.sendMessage(0,M).id files = { 'file': open(path, 'rb'), } params = { 'name': 'media', 'oid': M_id, 'size': len(open(path, 'rb').read()), 'type': 'image', 'ver': '1.0', } data = { 'params': json.dumps(params) } r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files) if r.status_code != 201: raise Exception('Upload Gif failure.') return True def sendGifWithURL(self, to_, url): path = 'pythonLiness.data' r = requests.get(url, stream=True) if r.status_code == 200: with open(path, 'w') as f: shutil.copyfileobj(r.raw, f) else: raise Exception('Download Gif failure.') try: self.sendGif(to_, path) except Exception as e: raise e def sendEvent(self, messageObject): return self.Talk.client.sendEvent(0, messageObject) def sendChatChecked(self, mid, lastMessageId): return self.Talk.client.sendChatChecked(0, mid, lastMessageId) def getMessageBoxCompactWrapUp(self, mid): return self.Talk.client.getMessageBoxCompactWrapUp(mid) def getMessageBoxCompactWrapUpList(self, start, messageBox): return self.Talk.client.getMessageBoxCompactWrapUpList(start, messageBox) def getRecentMessages(self, messageBox, count): return self.Talk.client.getRecentMessages(messageBox.id, count) def getMessageBox(self, channelId, messageboxId, lastMessagesCount): return self.Talk.client.getMessageBox(channelId, messageboxId, lastMessagesCount) def getMessageBoxList(self, channelId, lastMessagesCount): return self.Talk.client.getMessageBoxList(channelId, lastMessagesCount) def getMessageBoxListByStatus(self, channelId, lastMessagesCount, status): return self.Talk.client.getMessageBoxListByStatus(channelId, lastMessagesCount, status) def getMessageBoxWrapUp(self, mid): return self.Talk.client.getMessageBoxWrapUp(mid) def getMessageBoxWrapUpList(self, start, messageBoxCount): return self.Talk.client.getMessageBoxWrapUpList(start, messageBoxCount) def getCover(self,mid): h = self.getHome(mid) objId = h["result"]["homeInfo"]["objectId"] return "http://dl.profile.line-cdn.net/myhome/c/download.nhn?userid=" + mid+ "&oid=" + objId """Contact""" def blockContact(self, mid): return self.Talk.client.blockContact(0, mid) def unblockContact(self, mid): return self.Talk.client.unblockContact(0, mid) def findAndAddContactsByMid(self, mid): return self.Talk.client.findAndAddContactsByMid(0, mid) def findAndAddContactsByMids(self, midlist): for i in midlist: self.Talk.client.findAndAddContactsByMid(0, i) def findAndAddContactsByUserid(self, userid): return self.Talk.client.findAndAddContactsByUserid(0, userid) def findContactsByUserid(self, userid): return self.Talk.client.findContactByUserid(userid) def findContactByTicket(self, ticketId): return self.Talk.client.findContactByUserTicket(ticketId) def getAllContactIds(self): return self.Talk.client.getAllContactIds() def getBlockedContactIds(self): return self.Talk.client.getBlockedContactIds() def getContact(self, mid): return self.Talk.client.getContact(mid) def getContacts(self, midlist): return self.Talk.client.getContacts(midlist) def getFavoriteMids(self): return self.Talk.client.getFavoriteMids() def getHiddenContactMids(self): return self.Talk.client.getHiddenContactMids() def CloneContactProfile(self, mid): contact = self.getContact(mid) profile = self.getProfile() profile.displayName = contact.displayName profile.statusMessage = contact.statusMessage profile.pictureStatus = contact.pictureStatus self.updateDisplayPicture(profile.pictureStatus) return self.updateProfile(profile) def updateDisplayPicture(self, hash_id): return self.Talk.client.updateProfileAttribute(0, 8, hash_id) """Group""" def findGroupByTicket(self, ticketId): return self.Talk.client.findGroupByTicket(ticketId) def acceptGroupInvitation(self, groupId): return self.Talk.client.acceptGroupInvitation(0, groupId) def acceptGroupInvitationByTicket(self, groupId, ticketId): return self.Talk.client.acceptGroupInvitationByTicket(0, groupId, ticketId) def cancelGroupInvitation(self, groupId, contactIds): return self.Talk.client.cancelGroupInvitation(0, groupId, contactIds) def createGroup(self, name, midlist): return self.Talk.client.createGroup(0, name, midlist) def getGroup(self, groupId): return self.Talk.client.getGroup(groupId) def getGroups(self, groupIds): return self.Talk.client.getGroups(groupIds) def getGroupIdsInvited(self): return self.Talk.client.getGroupIdsInvited() def getGroupIdsJoined(self): return self.Talk.client.getGroupIdsJoined() def inviteIntoGroup(self, groupId, midlist): return self.Talk.client.inviteIntoGroup(0, groupId, midlist) def kickoutFromGroup(self, groupId, midlist): return self.Talk.client.kickoutFromGroup(0, groupId, midlist) def leaveGroup(self, groupId): return self.Talk.client.leaveGroup(0, groupId) def rejectGroupInvitation(self, groupId): return self.Talk.client.rejectGroupInvitation(0, groupId) def reissueGroupTicket(self, groupId): return self.Talk.client.reissueGroupTicket(groupId) def updateGroup(self, groupObject): return self.Talk.client.updateGroup(0, groupObject) def findGroupByTicket(self,ticketId): return self.Talk.client.findGroupByTicket(0,ticketId) """Room""" def createRoom(self, midlist): return self.Talk.client.createRoom(0, midlist) def getRoom(self, roomId): return self.Talk.client.getRoom(roomId) def inviteIntoRoom(self, roomId, midlist): return self.Talk.client.inviteIntoRoom(0, roomId, midlist) def leaveRoom(self, roomId): return self.Talk.client.leaveRoom(0, roomId) """TIMELINE""" def new_post(self, text): return self.channel.new_post(text) def like(self, mid, postid, likeType=1001): return self.channel.like(mid, postid, likeType) def comment(self, mid, postid, text): return self.channel.comment(mid, postid, text) def activity(self, limit=20): return self.channel.activity(limit) def getAlbum(self, gid): return self.channel.getAlbum(gid) def changeAlbumName(self, gid, name, albumId): return self.channel.changeAlbumName(gid, name, albumId) def deleteAlbum(self, gid, albumId): return self.channel.deleteAlbum(gid,albumId) def getNote(self,gid, commentLimit, likeLimit): return self.channel.getNote(gid, commentLimit, likeLimit) def getDetail(self,mid): return self.channel.getDetail(mid) def getHome(self,mid): return self.channel.getHome(mid) def createAlbum(self, gid, name): return self.channel.createAlbum(gid,name) def createAlbum2(self, gid, name, path): return self.channel.createAlbum(gid, name, path, oid) def __validate(self, mail, passwd, cert, token, qr): if mail is not None and passwd is not None and cert is None: return 1 elif mail is not None and passwd is not None and cert is not None: return 2 elif token is not None: return 3 elif qr is True: return 4 else: return 5 def loginResult(self, callback=True): if callback is True: callback = def_callback prof = self.getProfile() print("===============[zz7_Combat]================") print(" Thanks to GOD and my friend") print("===============[© By_Ricco]================") print("mid -> " + prof.mid) print("name -> " + prof.displayName) print("authToken -> " + self.authToken) print("cert -> " + self.cert if self.cert is not None else "")
false
true
1c4848c1f29b4b4d5d33da873e06fe8c0aa82152
2,834
py
Python
src/python/pants/backend/codegen/wire/java/java_wire_library.py
lahosken/pants
1b0340987c9b2eab9411416803c75b80736716e4
[ "Apache-2.0" ]
1
2021-11-11T14:04:24.000Z
2021-11-11T14:04:24.000Z
src/python/pants/backend/codegen/wire/java/java_wire_library.py
lahosken/pants
1b0340987c9b2eab9411416803c75b80736716e4
[ "Apache-2.0" ]
null
null
null
src/python/pants/backend/codegen/wire/java/java_wire_library.py
lahosken/pants
1b0340987c9b2eab9411416803c75b80736716e4
[ "Apache-2.0" ]
1
2021-11-11T14:04:12.000Z
2021-11-11T14:04:12.000Z
# coding=utf-8 # Copyright 2014 Pants project contributors (see CONTRIBUTORS.md). # Licensed under the Apache License, Version 2.0 (see LICENSE). from __future__ import (absolute_import, division, generators, nested_scopes, print_function, unicode_literals, with_statement) import logging from pants.backend.jvm.targets.exportable_jvm_library import ExportableJvmLibrary from pants.base.exceptions import TargetDefinitionException from pants.base.payload import Payload from pants.base.payload_field import PrimitiveField from pants.base.validation import assert_list logger = logging.getLogger(__name__) class JavaWireLibrary(ExportableJvmLibrary): """A Java library generated from Wire IDL files. Supports Wire 1.x only. For an example Wire 2.x interface that generates service stubs see: https://github.com/ericzundel/mvn2pants/tree/master/src/python/squarepants/plugins/sake_wire_codegen But note this requires you to write a custom wire code generator with a command line interface. :API: public """ def __init__(self, payload=None, service_writer=None, service_writer_options=None, roots=None, registry_class=None, enum_options=None, no_options=None, **kwargs): """ :param string service_writer: the name of the class to pass as the --service_writer option to the Wire compiler (For wire 1.0 only) :param list service_writer_options: A list of options to pass to the service writer (For wire 1.x only) :param list roots: passed through to the --roots option of the Wire compiler :param string registry_class: fully qualified class name of RegistryClass to create. If in doubt, specify com.squareup.wire.SimpleServiceWriter :param list enum_options: list of enums to pass to as the --enum-enum_options option, # optional :param boolean no_options: boolean that determines if --no_options flag is passed """ if not service_writer and service_writer_options: raise TargetDefinitionException(self, 'service_writer_options requires setting service_writer') payload = payload or Payload() payload.add_fields({ 'service_writer': PrimitiveField(service_writer or None), 'service_writer_options': PrimitiveField( assert_list(service_writer_options, key_arg='service_writer_options', raise_type=TargetDefinitionException)), 'roots': PrimitiveField(roots or []), 'registry_class': PrimitiveField(registry_class or None), 'enum_options': PrimitiveField(enum_options or []), 'no_options': PrimitiveField(no_options or False), }) super(JavaWireLibrary, self).__init__(payload=payload, **kwargs)
39.915493
102
0.717713
from __future__ import (absolute_import, division, generators, nested_scopes, print_function, unicode_literals, with_statement) import logging from pants.backend.jvm.targets.exportable_jvm_library import ExportableJvmLibrary from pants.base.exceptions import TargetDefinitionException from pants.base.payload import Payload from pants.base.payload_field import PrimitiveField from pants.base.validation import assert_list logger = logging.getLogger(__name__) class JavaWireLibrary(ExportableJvmLibrary): def __init__(self, payload=None, service_writer=None, service_writer_options=None, roots=None, registry_class=None, enum_options=None, no_options=None, **kwargs): if not service_writer and service_writer_options: raise TargetDefinitionException(self, 'service_writer_options requires setting service_writer') payload = payload or Payload() payload.add_fields({ 'service_writer': PrimitiveField(service_writer or None), 'service_writer_options': PrimitiveField( assert_list(service_writer_options, key_arg='service_writer_options', raise_type=TargetDefinitionException)), 'roots': PrimitiveField(roots or []), 'registry_class': PrimitiveField(registry_class or None), 'enum_options': PrimitiveField(enum_options or []), 'no_options': PrimitiveField(no_options or False), }) super(JavaWireLibrary, self).__init__(payload=payload, **kwargs)
true
true
1c4848ec3d5e9913f64fb0a43e2ba992978ad91b
5,150
py
Python
tools/error.py
Alex-Au1/Haku_Bot
87cc1813546797eec0f4760fafa76ce65a387a1d
[ "MIT" ]
null
null
null
tools/error.py
Alex-Au1/Haku_Bot
87cc1813546797eec0f4760fafa76ce65a387a1d
[ "MIT" ]
null
null
null
tools/error.py
Alex-Au1/Haku_Bot
87cc1813546797eec0f4760fafa76ce65a387a1d
[ "MIT" ]
null
null
null
import discord import copy from tools.embed import Embed, EmbededMessage from tools.string import StringTools import tools.members as Members from typing import Optional # Err_Embed: A template for embeding errors and warnings class Err_Embed(): def __init__(self, title, description): self.title = title self.description = description errors = {1: Err_Embed("Unable to Find Error", "{bot_nickname} is unable to find the error by the code `{err_code}`"), 2: Err_Embed("Unable to Warning", "{bot_nickname} is unable to find the warning by the code `{warn_code}`"), 3: Err_Embed("Unable to Find Selected Guild and Channel", "{bot_nickname} is unable to find the guild by the {guild_search_type} `{search_guild}` and the channel by the {channel_search_type} `{search_channel}`"), 4: Err_Embed("Unable to Find Selected Guild", "{bot_nickname} is unable to find the guild by the {guild_search_type} `{search_guild}`"), 5: Err_Embed("Unable to Find Selected Channel", "{bot_nickname} is unable to find the channel by the {channel_search_type} `{search_channel}`"), 6: Err_Embed("Please Enter {type_article} {correct_type} Parameter", "Please enter {type_article} **{correct_type}** for the parameter `{parameter}`"), 7: Err_Embed("Please Enter {type_article} {correct_type} Greater or Equal to {value}", "Please enter {type_article} **{correct_type} greater or equal to {value}** for the parameter `{parameter}`"), 8: Err_Embed("Please Enter {type_article} {correct_type} Lesser or Equal to {value}", "Please enter {type_article} **{correct_type} lesser or equal to {value}** for the parameter `{parameter}`"), 9: Err_Embed("Cannot Perform Action in the Channel: {channel}", "Cannot {action} to the channel, `{channel}`, in the guild, `{guild}`"), 10: Err_Embed("{element} is not part of {group}", "The input `{element}` for the parameter `{parameter}` is not an element of `{group}`"), 11: Err_Embed("Please Enter {type_article} {correct_type} Greater than {value}", "Please enter {type_article} **{correct_type} greater than {value}** for the parameter `{parameter}`"), 12: Err_Embed("Please Enter {type_article} {correct_type} Lesser than {value}", "Please enter {type_article} **{correct_type} lesser than {value}** for the parameter `{parameter}`"), 13: Err_Embed("Please Enter a Valid Subcommand", "Please enter a valid subcommand for the command `{command}`"), 14: Err_Embed("Please Enter a Valid Url for {type_article} {correct_type}", "Please enter a valid url for **{type_article} {correct_type}** in the parameter `{parameter}`"), 15: Err_Embed("Please Enter a Valid {correct_type}", "Please enter a valid **{correct_type}** for the parameter `{parameter}`"), 18: Err_Embed("Unable to Find Selected {member}", "{bot_nickname} is unable to find the {member} by the {member_search_type} `{search_member}`"), 19: Err_Embed("{action} Failed", "{bot_nickname} is unable to {action}"), 20: Err_Embed("Please Enter {type_article} {correct_type} {scope} in between {left} and {right}", "Please enter {type_article} **{correct_type} {scope} in between __{left}__ and __{right}__** for the parameter `{parameter}`")} warnings = {} # display_error(client, code, type, choice, **kwargs) Displays an error/ warning embeded message # depending on 'code' # requires: 0 <= code # 0 <= choice # 'type' is either "error" or "warning" def display_error(client: discord.Client, code: int, type: str = "error", choice: int = 0, **kwargs) -> Optional[EmbededMessage]: embed = Embed(client) description = "" title = "" if (type == "error"): title = "ERROR " elif (type == "warning"): title = "Warning " kw_keys = list(kwargs.keys()) for k in kw_keys: new_key = "{" + k + "}" kwargs[new_key] = kwargs.pop(k) if (type == "error"): if (code <= len(errors) and code > 0): err_title = errors[code].title err_description = errors[code].description else: err_title = errors[1].title err_description = errors[1].description kwargs = {"{err_code}": f"{code}"} code = 1 elif (type == "warning"): if (code <= len(warnings) and code > 0): err_title = warnings[code].title err_description = warnings[code].description else: err_title = warnings[2].title err_description = warnings[2].description kwargs = {"{warn_code}": f"{code}"} code = 2 kwargs.update({"{bot_nickname}": Members.DEFAULT_BOT_NAME}) title += f"{code}: {err_title}" title = StringTools.word_replace(title, kwargs, capitalize = True) description = StringTools.word_replace(err_description, kwargs) if (type == "error"): embeded_message = embed.error_embed(description, title) elif (type == "warning"): embeded_message = embed.warning_embed(description, title, choice = choice) return embeded_message
57.222222
236
0.65767
import discord import copy from tools.embed import Embed, EmbededMessage from tools.string import StringTools import tools.members as Members from typing import Optional class Err_Embed(): def __init__(self, title, description): self.title = title self.description = description errors = {1: Err_Embed("Unable to Find Error", "{bot_nickname} is unable to find the error by the code `{err_code}`"), 2: Err_Embed("Unable to Warning", "{bot_nickname} is unable to find the warning by the code `{warn_code}`"), 3: Err_Embed("Unable to Find Selected Guild and Channel", "{bot_nickname} is unable to find the guild by the {guild_search_type} `{search_guild}` and the channel by the {channel_search_type} `{search_channel}`"), 4: Err_Embed("Unable to Find Selected Guild", "{bot_nickname} is unable to find the guild by the {guild_search_type} `{search_guild}`"), 5: Err_Embed("Unable to Find Selected Channel", "{bot_nickname} is unable to find the channel by the {channel_search_type} `{search_channel}`"), 6: Err_Embed("Please Enter {type_article} {correct_type} Parameter", "Please enter {type_article} **{correct_type}** for the parameter `{parameter}`"), 7: Err_Embed("Please Enter {type_article} {correct_type} Greater or Equal to {value}", "Please enter {type_article} **{correct_type} greater or equal to {value}** for the parameter `{parameter}`"), 8: Err_Embed("Please Enter {type_article} {correct_type} Lesser or Equal to {value}", "Please enter {type_article} **{correct_type} lesser or equal to {value}** for the parameter `{parameter}`"), 9: Err_Embed("Cannot Perform Action in the Channel: {channel}", "Cannot {action} to the channel, `{channel}`, in the guild, `{guild}`"), 10: Err_Embed("{element} is not part of {group}", "The input `{element}` for the parameter `{parameter}` is not an element of `{group}`"), 11: Err_Embed("Please Enter {type_article} {correct_type} Greater than {value}", "Please enter {type_article} **{correct_type} greater than {value}** for the parameter `{parameter}`"), 12: Err_Embed("Please Enter {type_article} {correct_type} Lesser than {value}", "Please enter {type_article} **{correct_type} lesser than {value}** for the parameter `{parameter}`"), 13: Err_Embed("Please Enter a Valid Subcommand", "Please enter a valid subcommand for the command `{command}`"), 14: Err_Embed("Please Enter a Valid Url for {type_article} {correct_type}", "Please enter a valid url for **{type_article} {correct_type}** in the parameter `{parameter}`"), 15: Err_Embed("Please Enter a Valid {correct_type}", "Please enter a valid **{correct_type}** for the parameter `{parameter}`"), 18: Err_Embed("Unable to Find Selected {member}", "{bot_nickname} is unable to find the {member} by the {member_search_type} `{search_member}`"), 19: Err_Embed("{action} Failed", "{bot_nickname} is unable to {action}"), 20: Err_Embed("Please Enter {type_article} {correct_type} {scope} in between {left} and {right}", "Please enter {type_article} **{correct_type} {scope} in between __{left}__ and __{right}__** for the parameter `{parameter}`")} warnings = {} def display_error(client: discord.Client, code: int, type: str = "error", choice: int = 0, **kwargs) -> Optional[EmbededMessage]: embed = Embed(client) description = "" title = "" if (type == "error"): title = "ERROR " elif (type == "warning"): title = "Warning " kw_keys = list(kwargs.keys()) for k in kw_keys: new_key = "{" + k + "}" kwargs[new_key] = kwargs.pop(k) if (type == "error"): if (code <= len(errors) and code > 0): err_title = errors[code].title err_description = errors[code].description else: err_title = errors[1].title err_description = errors[1].description kwargs = {"{err_code}": f"{code}"} code = 1 elif (type == "warning"): if (code <= len(warnings) and code > 0): err_title = warnings[code].title err_description = warnings[code].description else: err_title = warnings[2].title err_description = warnings[2].description kwargs = {"{warn_code}": f"{code}"} code = 2 kwargs.update({"{bot_nickname}": Members.DEFAULT_BOT_NAME}) title += f"{code}: {err_title}" title = StringTools.word_replace(title, kwargs, capitalize = True) description = StringTools.word_replace(err_description, kwargs) if (type == "error"): embeded_message = embed.error_embed(description, title) elif (type == "warning"): embeded_message = embed.warning_embed(description, title, choice = choice) return embeded_message
true
true
1c484939ae60c601405bced057c4dedd90dff5c0
772
py
Python
lessons/best-practices/boulder_dem.py
csdms/ivy
862fc8bafa665864ceae25c4ead9e376ffe175cb
[ "CC-BY-4.0" ]
null
null
null
lessons/best-practices/boulder_dem.py
csdms/ivy
862fc8bafa665864ceae25c4ead9e376ffe175cb
[ "CC-BY-4.0" ]
1
2022-03-30T18:18:50.000Z
2022-03-30T18:18:50.000Z
lessons/best-practices/boulder_dem.py
csdms/ivy
862fc8bafa665864ceae25c4ead9e376ffe175cb
[ "CC-BY-4.0" ]
null
null
null
"""An example of reading topographical data from a file and displaying it.""" import pandas as pd import matplotlib.pyplot as plt topo_file = "../../data/topo.asc" def read(): try: topo = pd.read_csv(topo_file, header=None) except IOError: print("IOError: file '{}' cannot be read".format(topo_file)) else: return topo def display(data, show=False, outfile="boulder_dem.png"): fig, ax = plt.subplots() elev = ax.imshow(data, cmap="jet") fig.colorbar(elev, label="Elevation (m)") plt.title("Boulder Topography") if show is True: plt.show() else: plt.savefig(outfile, dpi=96) plt.close() if __name__ == "__main__": topo = read() if topo is not None: display(topo)
21.444444
77
0.619171
import pandas as pd import matplotlib.pyplot as plt topo_file = "../../data/topo.asc" def read(): try: topo = pd.read_csv(topo_file, header=None) except IOError: print("IOError: file '{}' cannot be read".format(topo_file)) else: return topo def display(data, show=False, outfile="boulder_dem.png"): fig, ax = plt.subplots() elev = ax.imshow(data, cmap="jet") fig.colorbar(elev, label="Elevation (m)") plt.title("Boulder Topography") if show is True: plt.show() else: plt.savefig(outfile, dpi=96) plt.close() if __name__ == "__main__": topo = read() if topo is not None: display(topo)
true
true
1c48496a2fcea3d1c774c8ee9daba45438f2e15a
166
py
Python
logging_middleware/checks.py
fearsd/django-logging-middleware
6eb95774c1bcb1829aa1a94223d9e2c39217d8f9
[ "MIT" ]
4
2021-04-08T14:14:04.000Z
2021-09-08T07:57:38.000Z
logging_middleware/checks.py
fearsd/django-logging-middleware
6eb95774c1bcb1829aa1a94223d9e2c39217d8f9
[ "MIT" ]
null
null
null
logging_middleware/checks.py
fearsd/django-logging-middleware
6eb95774c1bcb1829aa1a94223d9e2c39217d8f9
[ "MIT" ]
null
null
null
# from django.conf import settings from django.core import checks @checks.register def check_settings(app_configs, **kwargs): # temporary solution return []
20.75
42
0.753012
from django.core import checks @checks.register def check_settings(app_configs, **kwargs): return []
true
true
1c484a7d6852dd81d7c8ee92a960b16a3012a4e2
1,341
py
Python
backend/src/ml_models/context.py
lukemiloszewski/ml-models
826ab6c0adebe851e73b9e883af8abccfaebdacb
[ "MIT" ]
null
null
null
backend/src/ml_models/context.py
lukemiloszewski/ml-models
826ab6c0adebe851e73b9e883af8abccfaebdacb
[ "MIT" ]
16
2022-02-21T19:27:42.000Z
2022-03-31T01:47:33.000Z
backend/src/ml_models/context.py
lukemiloszewski/ml-models
826ab6c0adebe851e73b9e883af8abccfaebdacb
[ "MIT" ]
null
null
null
from __future__ import annotations from pathlib import Path from typing import Any, Dict, Optional from ml_models.clients.mnist_client import MNISTClient _CONTEXT: Optional[Context] = None class Attributes: def __init__(self, attributes_dict: Dict[str, Any]) -> None: self._attributes_dict = attributes_dict def get(self, attribute_id: str) -> Any: rv = self._get_attribute(attribute_id=attribute_id) return rv def _get_attribute(self, attribute_id: str) -> Any: attribute = self._attributes_dict.get(attribute_id, None) if attribute is None: err_msg = f"Invalid attribute: {attribute_id}, available attributes: {list(self._attributes_dict.keys())}" raise AttributeError(err_msg) return attribute class Context: def __init__(self, clients: Attributes) -> None: self.clients = clients def configure_context(root_path: Path, mnist_onnx_path: Path): global _CONTEXT client_dict = { "mnist": MNISTClient(str(root_path / mnist_onnx_path)), } client_attributes = Attributes(attributes_dict=client_dict) context = Context(clients=client_attributes) _CONTEXT = context def get_context() -> Context: if _CONTEXT is None: raise ValueError("Context has not been initialised") return _CONTEXT
27.367347
118
0.706189
from __future__ import annotations from pathlib import Path from typing import Any, Dict, Optional from ml_models.clients.mnist_client import MNISTClient _CONTEXT: Optional[Context] = None class Attributes: def __init__(self, attributes_dict: Dict[str, Any]) -> None: self._attributes_dict = attributes_dict def get(self, attribute_id: str) -> Any: rv = self._get_attribute(attribute_id=attribute_id) return rv def _get_attribute(self, attribute_id: str) -> Any: attribute = self._attributes_dict.get(attribute_id, None) if attribute is None: err_msg = f"Invalid attribute: {attribute_id}, available attributes: {list(self._attributes_dict.keys())}" raise AttributeError(err_msg) return attribute class Context: def __init__(self, clients: Attributes) -> None: self.clients = clients def configure_context(root_path: Path, mnist_onnx_path: Path): global _CONTEXT client_dict = { "mnist": MNISTClient(str(root_path / mnist_onnx_path)), } client_attributes = Attributes(attributes_dict=client_dict) context = Context(clients=client_attributes) _CONTEXT = context def get_context() -> Context: if _CONTEXT is None: raise ValueError("Context has not been initialised") return _CONTEXT
true
true
1c484c826f55f41d4e0d2c5b6336352e83d80519
2,571
py
Python
preprocessors/rcv1v2_data.py
laddie132/LW-PT
28b469ba68a5d4fba68b992cff5372e63ec2ed42
[ "MIT" ]
9
2020-08-20T18:15:43.000Z
2022-02-10T02:54:30.000Z
preprocessors/rcv1v2_data.py
laddie132/LW-PT
28b469ba68a5d4fba68b992cff5372e63ec2ed42
[ "MIT" ]
1
2021-11-19T01:29:47.000Z
2021-11-19T09:58:38.000Z
preprocessors/rcv1v2_data.py
laddie132/LW-PT
28b469ba68a5d4fba68b992cff5372e63ec2ed42
[ "MIT" ]
3
2021-05-29T02:11:34.000Z
2021-12-14T15:43:22.000Z
#!/usr/bin/env python # -*- coding: utf-8 -*- __author__ = "Han" __email__ = "liuhan132@foxmail.com" import os import logging from .base import BaseDataset logger = logging.getLogger(__name__) class RCV1V2(BaseDataset): """ RCV1-V2 dataset """ def __init__(self, data_path, random_seed): super(RCV1V2, self).__init__(h5_path='data/rcv1v2.h5', save_data_path='data/rcv1v2.pkl', save_meta_data_path='data/rcv1v2.meta.json', w2v_path='data/rcv1v2_word2vec.model', load_emb=False, emb_dim=256, max_vocab_size=None, max_sent_num=15, max_sent_len=50, max_doc_len=500, hier=False, random_seed=random_seed) self.data_path = 'data/rcv1-v2/sgm' if data_path is '' else data_path def load_all_data(self): train_text_labels = self.load_data(os.path.join(self.data_path, 'train.src.id'), os.path.join(self.data_path, 'train.tgt.id')) val_text_labels = self.load_data(os.path.join(self.data_path, 'valid.src.id'), os.path.join(self.data_path, 'valid.tgt.id')) test_text_labels = self.load_data(os.path.join(self.data_path, 'test.src.id'), os.path.join(self.data_path, 'test.tgt.id')) return train_text_labels, val_text_labels, test_text_labels def load_data(self, text_path, label_path): texts_labels = [] with open(text_path, 'r') as tf, open(label_path, 'r') as lf: for text, label in zip(tf, lf): if text != '' and label != '': text = list(map(lambda x: int(x), text.strip().split())) label = list(map(lambda x: int(x), label.strip().split())) label = label[1:-1] # text = text.strip().split() # label = label.strip().split() self.texts_labels_sum += len(label) self.texts_words_sum += len(text) self.all_texts.append(text) self.all_labels.extend(label) texts_labels.append((text, label)) return texts_labels
38.373134
88
0.492804
__author__ = "Han" __email__ = "liuhan132@foxmail.com" import os import logging from .base import BaseDataset logger = logging.getLogger(__name__) class RCV1V2(BaseDataset): def __init__(self, data_path, random_seed): super(RCV1V2, self).__init__(h5_path='data/rcv1v2.h5', save_data_path='data/rcv1v2.pkl', save_meta_data_path='data/rcv1v2.meta.json', w2v_path='data/rcv1v2_word2vec.model', load_emb=False, emb_dim=256, max_vocab_size=None, max_sent_num=15, max_sent_len=50, max_doc_len=500, hier=False, random_seed=random_seed) self.data_path = 'data/rcv1-v2/sgm' if data_path is '' else data_path def load_all_data(self): train_text_labels = self.load_data(os.path.join(self.data_path, 'train.src.id'), os.path.join(self.data_path, 'train.tgt.id')) val_text_labels = self.load_data(os.path.join(self.data_path, 'valid.src.id'), os.path.join(self.data_path, 'valid.tgt.id')) test_text_labels = self.load_data(os.path.join(self.data_path, 'test.src.id'), os.path.join(self.data_path, 'test.tgt.id')) return train_text_labels, val_text_labels, test_text_labels def load_data(self, text_path, label_path): texts_labels = [] with open(text_path, 'r') as tf, open(label_path, 'r') as lf: for text, label in zip(tf, lf): if text != '' and label != '': text = list(map(lambda x: int(x), text.strip().split())) label = list(map(lambda x: int(x), label.strip().split())) label = label[1:-1] self.texts_labels_sum += len(label) self.texts_words_sum += len(text) self.all_texts.append(text) self.all_labels.extend(label) texts_labels.append((text, label)) return texts_labels
true
true
1c484d2322cae0bb96e69cc698013819ab7ee299
3,346
py
Python
PyWidget3/shape/__init__.py
galaxyjim/PyWidget3
eb3d269e4e7d8a68ca957d32bc704e31eca20015
[ "BSD-3-Clause" ]
null
null
null
PyWidget3/shape/__init__.py
galaxyjim/PyWidget3
eb3d269e4e7d8a68ca957d32bc704e31eca20015
[ "BSD-3-Clause" ]
23
2015-03-14T00:03:11.000Z
2015-04-10T23:24:21.000Z
PyWidget3/shape/__init__.py
galaxyjim/PyWidget3
eb3d269e4e7d8a68ca957d32bc704e31eca20015
[ "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python # -*- coding: utf-8 -*- # ------------------------------------------------------------------------------ # Copyright (c) 2009 Nicolas Rougier # Copyright (c) 2015 James Gaston # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in # the documentation and/or other materials provided with the # distribution. # * Neither the name of pyglet nor the names of its # contributors may be used to endorse or promote products # derived from this software without specific prior written # permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS # FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE # COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, # INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, # BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT # LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN # ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE # POSSIBILITY OF SUCH DAMAGE. # ----------------------------------------------------------------------------- '''Defines a set of basic 2D shapes. All shapes have: ---------------- - a position in 2D space - a dimension in 2D space - a x alignment ('left', 'center' or 'right') - a y alignment ('top', 'center' or 'bottom') - background color(s) - background texture - foreground color(s) (for the one pixel border) Display Model: -------------- Any shape is defined by the x, y, width and height attributes. Borders are drawn on the inside of the shape as a single pixel line in the specified border color(s). Foreground or background color can be specified as a single tuple of 4 floats for uniform color, 2 tuples of 4 floats for radial color patterns (going from inner to outer) or 4 tuples of 4 floats for an interpolated pattern between the four corners. Note that the radial pattern does not work for triangle or rectangle. Available shapes: ----------------- - Rectangle (with round corners or not) - Ellipse (circle if width == height) - Triangle - Cross (with any number of branches) - Star (with any number of branches) Example usage: -------------- rectangle = Rectangle(x=100,y=100,width=100,height=100,radius=10) @window.event def on_draw(): window.clear() rectangle.draw() @window.event def on_mouse_press(x,y,button,modifiers): if rectangle.hit_test(x,y): print 'Hit' :requires: pyglet 1.1 ''' __docformat__ = 'restructuredtext' __version__ = '1.0' from .rectangle import Rectangle from .triangle import Triangle from .ellipse import Ellipse from .cross import Cross from .star import Star
34.854167
80
0.698745
__docformat__ = 'restructuredtext' __version__ = '1.0' from .rectangle import Rectangle from .triangle import Triangle from .ellipse import Ellipse from .cross import Cross from .star import Star
true
true
1c484dccfdcbf952d46374d4d53d3daed255caa8
48
py
Python
samcli/__init__.py
rawhideron/mav_0122
3f8b92347087f94ec76667dbb2f647937725660d
[ "BSD-2-Clause", "Apache-2.0" ]
1
2021-07-10T14:19:00.000Z
2021-07-10T14:19:00.000Z
samcli/__init__.py
QPC-database/aws-sam-cli
59c85768356089edb265c2ea7f53bce2412f9e19
[ "BSD-2-Clause", "Apache-2.0" ]
null
null
null
samcli/__init__.py
QPC-database/aws-sam-cli
59c85768356089edb265c2ea7f53bce2412f9e19
[ "BSD-2-Clause", "Apache-2.0" ]
null
null
null
""" SAM CLI version """ __version__ = "1.26.0"
8
22
0.583333
__version__ = "1.26.0"
true
true
1c484f70ffaacf4cb7ed13a8f996d67a217a2f85
1,558
py
Python
user_main.py
s-jun/OSS_Term_Project
47747a92944f7f94f1393c9072f7ee9034de090a
[ "MIT" ]
null
null
null
user_main.py
s-jun/OSS_Term_Project
47747a92944f7f94f1393c9072f7ee9034de090a
[ "MIT" ]
null
null
null
user_main.py
s-jun/OSS_Term_Project
47747a92944f7f94f1393c9072f7ee9034de090a
[ "MIT" ]
null
null
null
import sys from PyQt5.QtWidgets import * from PyQt5 import uic import matplotlib.pyplot as plt from matplotlib.backends.backend_qt5agg import FigureCanvasQTAgg as FigureCanvas import chart from data_manager import read_predict form_class = uic.loadUiType("user.ui")[0] class WindowClass(QMainWindow, form_class): def __init__(self): super().__init__() self.setupUi(self) self.initUI() self.comboBox.activated[str].connect(self.clicked) def initUI(self): self.fig = plt.Figure() combo = self.comboBox.currentText() self.canvas = FigureCanvas(self.fig) self.chart.addWidget(self.canvas) chart.draw_chart(self, combo) self.canvas.draw() self.prediction.append(read_predict(combo)) def clicked(self, text): self.clear() self.prediction.append(read_predict(text)) self.fig = plt.Figure() self.canvas = FigureCanvas(self.fig) self.chart.addWidget(self.canvas) chart.draw_chart(self, text) self.canvas.draw() def clear(self): self.prediction.clear() tmp = self.chart if tmp is not None: while tmp.count(): item = tmp.takeAt(0) widget = item.widget() if widget is not None: widget.deleteLater() else: self.clearvbox(item.layout()) if __name__== "__main__" : app = QApplication(sys.argv) myWindow = WindowClass() myWindow.show() app.exec_()
27.333333
80
0.616175
import sys from PyQt5.QtWidgets import * from PyQt5 import uic import matplotlib.pyplot as plt from matplotlib.backends.backend_qt5agg import FigureCanvasQTAgg as FigureCanvas import chart from data_manager import read_predict form_class = uic.loadUiType("user.ui")[0] class WindowClass(QMainWindow, form_class): def __init__(self): super().__init__() self.setupUi(self) self.initUI() self.comboBox.activated[str].connect(self.clicked) def initUI(self): self.fig = plt.Figure() combo = self.comboBox.currentText() self.canvas = FigureCanvas(self.fig) self.chart.addWidget(self.canvas) chart.draw_chart(self, combo) self.canvas.draw() self.prediction.append(read_predict(combo)) def clicked(self, text): self.clear() self.prediction.append(read_predict(text)) self.fig = plt.Figure() self.canvas = FigureCanvas(self.fig) self.chart.addWidget(self.canvas) chart.draw_chart(self, text) self.canvas.draw() def clear(self): self.prediction.clear() tmp = self.chart if tmp is not None: while tmp.count(): item = tmp.takeAt(0) widget = item.widget() if widget is not None: widget.deleteLater() else: self.clearvbox(item.layout()) if __name__== "__main__" : app = QApplication(sys.argv) myWindow = WindowClass() myWindow.show() app.exec_()
true
true
1c484f734e415132256c335e128e6abcf0544a59
13,032
py
Python
tests/generator/test_compression.py
randywessels/tad-blockchain
08a5f9565aa27f211350717d5e8cda14b46359e4
[ "Apache-2.0" ]
null
null
null
tests/generator/test_compression.py
randywessels/tad-blockchain
08a5f9565aa27f211350717d5e8cda14b46359e4
[ "Apache-2.0" ]
null
null
null
tests/generator/test_compression.py
randywessels/tad-blockchain
08a5f9565aa27f211350717d5e8cda14b46359e4
[ "Apache-2.0" ]
null
null
null
# flake8: noqa: F501 from dataclasses import dataclass from typing import List, Any from unittest import TestCase from tad.full_node.bundle_tools import ( bundle_suitable_for_compression, compressed_coin_solution_entry_list, compressed_spend_bundle_solution, match_standard_transaction_at_any_index, simple_solution_generator, spend_bundle_to_serialized_coin_solution_entry_list, ) from tad.full_node.generator import run_generator, create_generator_args from tad.types.blockchain_format.program import Program, SerializedProgram, INFINITE_COST from tad.types.generator_types import BlockGenerator, CompressorArg, GeneratorArg from tad.types.spend_bundle import SpendBundle from tad.util.byte_types import hexstr_to_bytes from tad.util.ints import uint32 from tad.wallet.puzzles.load_clvm import load_clvm from tests.core.make_block_generator import make_spend_bundle from clvm import SExp import io from clvm.serialize import sexp_from_stream from clvm_tools import binutils TEST_GEN_DESERIALIZE = load_clvm("test_generator_deserialize.clvm", package_or_requirement="tad.wallet.puzzles") DESERIALIZE_MOD = load_clvm("chialisp_deserialisation.clvm", package_or_requirement="tad.wallet.puzzles") DECOMPRESS_PUZZLE = load_clvm("decompress_puzzle.clvm", package_or_requirement="tad.wallet.puzzles") DECOMPRESS_CSE = load_clvm("decompress_coin_solution_entry.clvm", package_or_requirement="tad.wallet.puzzles") DECOMPRESS_CSE_WITH_PREFIX = load_clvm( "decompress_coin_solution_entry_with_prefix.clvm", package_or_requirement="tad.wallet.puzzles" ) DECOMPRESS_BLOCK = load_clvm("block_program_zero.clvm", package_or_requirement="tad.wallet.puzzles") TEST_MULTIPLE = load_clvm("test_multiple_generator_input_arguments.clvm", package_or_requirement="tad.wallet.puzzles") Nil = Program.from_bytes(b"\x80") original_generator = hexstr_to_bytes( "ff01ffffffa00000000000000000000000000000000000000000000000000000000000000000ff830186a080ffffff02ffff01ff02ffff01ff02ffff03ff0bffff01ff02ffff03ffff09ff05ffff1dff0bffff1effff0bff0bffff02ff06ffff04ff02ffff04ff17ff8080808080808080ffff01ff02ff17ff2f80ffff01ff088080ff0180ffff01ff04ffff04ff04ffff04ff05ffff04ffff02ff06ffff04ff02ffff04ff17ff80808080ff80808080ffff02ff17ff2f808080ff0180ffff04ffff01ff32ff02ffff03ffff07ff0580ffff01ff0bffff0102ffff02ff06ffff04ff02ffff04ff09ff80808080ffff02ff06ffff04ff02ffff04ff0dff8080808080ffff01ff0bffff0101ff058080ff0180ff018080ffff04ffff01b081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3ff018080ffff80ffff01ffff33ffa06b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9ff830186a08080ff8080808080" ) # noqa gen1 = b"aaaaaaaaaa" + original_generator gen2 = b"bb" + original_generator FAKE_BLOCK_HEIGHT1 = uint32(100) FAKE_BLOCK_HEIGHT2 = uint32(200) @dataclass(frozen=True) class MultipleCompressorArg: arg: List[CompressorArg] split_offset: int def create_multiple_ref_generator(args: MultipleCompressorArg, spend_bundle: SpendBundle) -> BlockGenerator: """ Decompress a transaction by referencing bytes from multiple input generator references """ compressed_cse_list = compressed_coin_solution_entry_list(spend_bundle) program = TEST_MULTIPLE.curry( DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, args.arg[0].start, args.arg[0].end - args.split_offset, args.arg[1].end - args.split_offset, args.arg[1].end, compressed_cse_list, ) # TODO aqk: Improve ergonomics of CompressorArg -> GeneratorArg conversion generator_args = [ GeneratorArg(FAKE_BLOCK_HEIGHT1, args.arg[0].generator), GeneratorArg(FAKE_BLOCK_HEIGHT2, args.arg[1].generator), ] return BlockGenerator(program, generator_args) def spend_bundle_to_coin_solution_entry_list(bundle: SpendBundle) -> List[Any]: r = [] for coin_solution in bundle.coin_solutions: entry = [ coin_solution.coin.parent_coin_info, sexp_from_stream(io.BytesIO(bytes(coin_solution.puzzle_reveal)), SExp.to), coin_solution.coin.amount, sexp_from_stream(io.BytesIO(bytes(coin_solution.solution)), SExp.to), ] r.append(entry) return r class TestCompression(TestCase): def test_spend_bundle_suitable(self): sb: SpendBundle = make_spend_bundle(1) assert bundle_suitable_for_compression(sb) def test_compress_spend_bundle(self): pass def test_multiple_input_gen_refs(self): start1, end1 = match_standard_transaction_at_any_index(gen1) start2, end2 = match_standard_transaction_at_any_index(gen2) ca1 = CompressorArg(FAKE_BLOCK_HEIGHT1, SerializedProgram.from_bytes(gen1), start1, end1) ca2 = CompressorArg(FAKE_BLOCK_HEIGHT2, SerializedProgram.from_bytes(gen2), start2, end2) prefix_len1 = end1 - start1 prefix_len2 = end2 - start2 assert prefix_len1 == prefix_len2 prefix_len = prefix_len1 results = [] for split_offset in range(prefix_len): gen_args = MultipleCompressorArg([ca1, ca2], split_offset) spend_bundle: SpendBundle = make_spend_bundle(1) multi_gen = create_multiple_ref_generator(gen_args, spend_bundle) cost, result = run_generator(multi_gen, INFINITE_COST) results.append(result) assert result is not None assert cost > 0 assert all(r == results[0] for r in results) def test_compressed_block_results(self): sb: SpendBundle = make_spend_bundle(1) start, end = match_standard_transaction_at_any_index(original_generator) ca = CompressorArg(uint32(0), SerializedProgram.from_bytes(original_generator), start, end) c = compressed_spend_bundle_solution(ca, sb) s = simple_solution_generator(sb) assert c != s cost_c, result_c = run_generator(c, INFINITE_COST) cost_s, result_s = run_generator(s, INFINITE_COST) print(result_c) assert result_c is not None assert result_s is not None assert result_c == result_s def test_spend_byndle_coin_solution(self): for i in range(0, 10): sb: SpendBundle = make_spend_bundle(i) cs1 = SExp.to(spend_bundle_to_coin_solution_entry_list(sb)).as_bin() cs2 = spend_bundle_to_serialized_coin_solution_entry_list(sb) assert cs1 == cs2 class TestDecompression(TestCase): def __init__(self, *args, **kwargs): super(TestDecompression, self).__init__(*args, **kwargs) self.maxDiff = None def test_deserialization(self): self.maxDiff = None cost, out = DESERIALIZE_MOD.run_with_cost(INFINITE_COST, [bytes(Program.to("hello"))]) assert out == Program.to("hello") def test_deserialization_as_argument(self): self.maxDiff = None cost, out = TEST_GEN_DESERIALIZE.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, Nil, bytes(Program.to("hello"))] ) print(bytes(Program.to("hello"))) print() print(out) assert out == Program.to("hello") def test_decompress_puzzle(self): cost, out = DECOMPRESS_PUZZLE.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, b"\xff", bytes(Program.to("pubkey")), b"\x80"] ) print() print(out) # An empty CSE is invalid. (An empty CSE list may be okay) # def test_decompress_empty_cse(self): # cse0 = binutils.assemble("()") # cost, out = DECOMPRESS_CSE.run_with_cost(INFINITE_COST, [DESERIALIZE_MOD, DECOMPRESS_PUZZLE, b"\xff", b"\x80", cse0]) # print() # print(out) def test_decompress_cse(self): """Decompress a single CSE / CoinSolutionEntry""" cse0 = binutils.assemble( "((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())))" ) # noqa cost, out = DECOMPRESS_CSE.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, DECOMPRESS_PUZZLE, b"\xff", b"\x80", cse0] ) print() print(out) def test_decompress_cse_with_prefix(self): cse0 = binutils.assemble( "((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())))" ) # noqa start = 2 + 44 end = start + 238 prefix = original_generator[start:end] # (deserialize decompress_puzzle puzzle_prefix cse) cost, out = DECOMPRESS_CSE_WITH_PREFIX.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, DECOMPRESS_PUZZLE, prefix, cse0] ) print() print(out) def test_block_program_zero(self): "Decompress a list of CSEs" self.maxDiff = None cse1 = binutils.assemble( "(((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ()))))" ) # noqa cse2 = binutils.assemble( """ ( ((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())) ) ((0x0000000000000000000000000000000000000000000000000000000000000001 0x0186a0) (0xb0a6207f5173ec41491d9f2c1b8fff5579e13703077e0eaca8fe587669dcccf51e9209a6b65576845ece5f7c2f3229e7e3 (() (q (51 0x24254a3efc3ebfac9979bbe0d615e2eda043aa329905f65b63846fa24149e2b6 0x0186a0)) ()))) ) """ ) # noqa start = 2 + 44 end = start + 238 # (mod (decompress_puzzle decompress_coin_solution_entry start end compressed_cses deserialize generator_list reserved_arg) # cost, out = DECOMPRESS_BLOCK.run_with_cost(INFINITE_COST, [DECOMPRESS_PUZZLE, DECOMPRESS_CSE, start, Program.to(end), cse0, DESERIALIZE_MOD, bytes(original_generator)]) cost, out = DECOMPRESS_BLOCK.run_with_cost( INFINITE_COST, [ DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, start, Program.to(end), cse2, DESERIALIZE_MOD, [bytes(original_generator)], ], ) print() print(out) def test_block_program_zero_with_curry(self): self.maxDiff = None cse1 = binutils.assemble( "(((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ()))))" ) # noqa cse2 = binutils.assemble( """ ( ((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())) ) ((0x0000000000000000000000000000000000000000000000000000000000000001 0x0186a0) (0xb0a6207f5173ec41491d9f2c1b8fff5579e13703077e0eaca8fe587669dcccf51e9209a6b65576845ece5f7c2f3229e7e3 (() (q (51 0x24254a3efc3ebfac9979bbe0d615e2eda043aa329905f65b63846fa24149e2b6 0x0186a0)) ()))) ) """ ) # noqa start = 2 + 44 end = start + 238 # (mod (decompress_puzzle decompress_coin_solution_entry start end compressed_cses deserialize generator_list reserved_arg) # cost, out = DECOMPRESS_BLOCK.run_with_cost(INFINITE_COST, [DECOMPRESS_PUZZLE, DECOMPRESS_CSE, start, Program.to(end), cse0, DESERIALIZE_MOD, bytes(original_generator)]) p = DECOMPRESS_BLOCK.curry(DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, start, Program.to(end)) cost, out = p.run_with_cost(INFINITE_COST, [cse2, DESERIALIZE_MOD, [bytes(original_generator)]]) print() print(p) print(out) p_with_cses = DECOMPRESS_BLOCK.curry( DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, start, Program.to(end), cse2, DESERIALIZE_MOD ) generator_args = create_generator_args([SerializedProgram.from_bytes(original_generator)]) cost, out = p_with_cses.run_with_cost(INFINITE_COST, generator_args) print() print(p_with_cses) print(out)
44.176271
792
0.73872
from dataclasses import dataclass from typing import List, Any from unittest import TestCase from tad.full_node.bundle_tools import ( bundle_suitable_for_compression, compressed_coin_solution_entry_list, compressed_spend_bundle_solution, match_standard_transaction_at_any_index, simple_solution_generator, spend_bundle_to_serialized_coin_solution_entry_list, ) from tad.full_node.generator import run_generator, create_generator_args from tad.types.blockchain_format.program import Program, SerializedProgram, INFINITE_COST from tad.types.generator_types import BlockGenerator, CompressorArg, GeneratorArg from tad.types.spend_bundle import SpendBundle from tad.util.byte_types import hexstr_to_bytes from tad.util.ints import uint32 from tad.wallet.puzzles.load_clvm import load_clvm from tests.core.make_block_generator import make_spend_bundle from clvm import SExp import io from clvm.serialize import sexp_from_stream from clvm_tools import binutils TEST_GEN_DESERIALIZE = load_clvm("test_generator_deserialize.clvm", package_or_requirement="tad.wallet.puzzles") DESERIALIZE_MOD = load_clvm("chialisp_deserialisation.clvm", package_or_requirement="tad.wallet.puzzles") DECOMPRESS_PUZZLE = load_clvm("decompress_puzzle.clvm", package_or_requirement="tad.wallet.puzzles") DECOMPRESS_CSE = load_clvm("decompress_coin_solution_entry.clvm", package_or_requirement="tad.wallet.puzzles") DECOMPRESS_CSE_WITH_PREFIX = load_clvm( "decompress_coin_solution_entry_with_prefix.clvm", package_or_requirement="tad.wallet.puzzles" ) DECOMPRESS_BLOCK = load_clvm("block_program_zero.clvm", package_or_requirement="tad.wallet.puzzles") TEST_MULTIPLE = load_clvm("test_multiple_generator_input_arguments.clvm", package_or_requirement="tad.wallet.puzzles") Nil = Program.from_bytes(b"\x80") original_generator = hexstr_to_bytes( "ff01ffffffa00000000000000000000000000000000000000000000000000000000000000000ff830186a080ffffff02ffff01ff02ffff01ff02ffff03ff0bffff01ff02ffff03ffff09ff05ffff1dff0bffff1effff0bff0bffff02ff06ffff04ff02ffff04ff17ff8080808080808080ffff01ff02ff17ff2f80ffff01ff088080ff0180ffff01ff04ffff04ff04ffff04ff05ffff04ffff02ff06ffff04ff02ffff04ff17ff80808080ff80808080ffff02ff17ff2f808080ff0180ffff04ffff01ff32ff02ffff03ffff07ff0580ffff01ff0bffff0102ffff02ff06ffff04ff02ffff04ff09ff80808080ffff02ff06ffff04ff02ffff04ff0dff8080808080ffff01ff0bffff0101ff058080ff0180ff018080ffff04ffff01b081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3ff018080ffff80ffff01ffff33ffa06b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9ff830186a08080ff8080808080" ) gen1 = b"aaaaaaaaaa" + original_generator gen2 = b"bb" + original_generator FAKE_BLOCK_HEIGHT1 = uint32(100) FAKE_BLOCK_HEIGHT2 = uint32(200) @dataclass(frozen=True) class MultipleCompressorArg: arg: List[CompressorArg] split_offset: int def create_multiple_ref_generator(args: MultipleCompressorArg, spend_bundle: SpendBundle) -> BlockGenerator: compressed_cse_list = compressed_coin_solution_entry_list(spend_bundle) program = TEST_MULTIPLE.curry( DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, args.arg[0].start, args.arg[0].end - args.split_offset, args.arg[1].end - args.split_offset, args.arg[1].end, compressed_cse_list, ) generator_args = [ GeneratorArg(FAKE_BLOCK_HEIGHT1, args.arg[0].generator), GeneratorArg(FAKE_BLOCK_HEIGHT2, args.arg[1].generator), ] return BlockGenerator(program, generator_args) def spend_bundle_to_coin_solution_entry_list(bundle: SpendBundle) -> List[Any]: r = [] for coin_solution in bundle.coin_solutions: entry = [ coin_solution.coin.parent_coin_info, sexp_from_stream(io.BytesIO(bytes(coin_solution.puzzle_reveal)), SExp.to), coin_solution.coin.amount, sexp_from_stream(io.BytesIO(bytes(coin_solution.solution)), SExp.to), ] r.append(entry) return r class TestCompression(TestCase): def test_spend_bundle_suitable(self): sb: SpendBundle = make_spend_bundle(1) assert bundle_suitable_for_compression(sb) def test_compress_spend_bundle(self): pass def test_multiple_input_gen_refs(self): start1, end1 = match_standard_transaction_at_any_index(gen1) start2, end2 = match_standard_transaction_at_any_index(gen2) ca1 = CompressorArg(FAKE_BLOCK_HEIGHT1, SerializedProgram.from_bytes(gen1), start1, end1) ca2 = CompressorArg(FAKE_BLOCK_HEIGHT2, SerializedProgram.from_bytes(gen2), start2, end2) prefix_len1 = end1 - start1 prefix_len2 = end2 - start2 assert prefix_len1 == prefix_len2 prefix_len = prefix_len1 results = [] for split_offset in range(prefix_len): gen_args = MultipleCompressorArg([ca1, ca2], split_offset) spend_bundle: SpendBundle = make_spend_bundle(1) multi_gen = create_multiple_ref_generator(gen_args, spend_bundle) cost, result = run_generator(multi_gen, INFINITE_COST) results.append(result) assert result is not None assert cost > 0 assert all(r == results[0] for r in results) def test_compressed_block_results(self): sb: SpendBundle = make_spend_bundle(1) start, end = match_standard_transaction_at_any_index(original_generator) ca = CompressorArg(uint32(0), SerializedProgram.from_bytes(original_generator), start, end) c = compressed_spend_bundle_solution(ca, sb) s = simple_solution_generator(sb) assert c != s cost_c, result_c = run_generator(c, INFINITE_COST) cost_s, result_s = run_generator(s, INFINITE_COST) print(result_c) assert result_c is not None assert result_s is not None assert result_c == result_s def test_spend_byndle_coin_solution(self): for i in range(0, 10): sb: SpendBundle = make_spend_bundle(i) cs1 = SExp.to(spend_bundle_to_coin_solution_entry_list(sb)).as_bin() cs2 = spend_bundle_to_serialized_coin_solution_entry_list(sb) assert cs1 == cs2 class TestDecompression(TestCase): def __init__(self, *args, **kwargs): super(TestDecompression, self).__init__(*args, **kwargs) self.maxDiff = None def test_deserialization(self): self.maxDiff = None cost, out = DESERIALIZE_MOD.run_with_cost(INFINITE_COST, [bytes(Program.to("hello"))]) assert out == Program.to("hello") def test_deserialization_as_argument(self): self.maxDiff = None cost, out = TEST_GEN_DESERIALIZE.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, Nil, bytes(Program.to("hello"))] ) print(bytes(Program.to("hello"))) print() print(out) assert out == Program.to("hello") def test_decompress_puzzle(self): cost, out = DECOMPRESS_PUZZLE.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, b"\xff", bytes(Program.to("pubkey")), b"\x80"] ) print() print(out) def test_decompress_cse(self): cse0 = binutils.assemble( "((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())))" ) cost, out = DECOMPRESS_CSE.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, DECOMPRESS_PUZZLE, b"\xff", b"\x80", cse0] ) print() print(out) def test_decompress_cse_with_prefix(self): cse0 = binutils.assemble( "((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())))" ) start = 2 + 44 end = start + 238 prefix = original_generator[start:end] cost, out = DECOMPRESS_CSE_WITH_PREFIX.run_with_cost( INFINITE_COST, [DESERIALIZE_MOD, DECOMPRESS_PUZZLE, prefix, cse0] ) print() print(out) def test_block_program_zero(self): self.maxDiff = None cse1 = binutils.assemble( "(((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ()))))" ) cse2 = binutils.assemble( """ ( ((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())) ) ((0x0000000000000000000000000000000000000000000000000000000000000001 0x0186a0) (0xb0a6207f5173ec41491d9f2c1b8fff5579e13703077e0eaca8fe587669dcccf51e9209a6b65576845ece5f7c2f3229e7e3 (() (q (51 0x24254a3efc3ebfac9979bbe0d615e2eda043aa329905f65b63846fa24149e2b6 0x0186a0)) ()))) ) """ ) start = 2 + 44 end = start + 238 cost, out = DECOMPRESS_BLOCK.run_with_cost( INFINITE_COST, [ DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, start, Program.to(end), cse2, DESERIALIZE_MOD, [bytes(original_generator)], ], ) print() print(out) def test_block_program_zero_with_curry(self): self.maxDiff = None cse1 = binutils.assemble( "(((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ()))))" ) cse2 = binutils.assemble( """ ( ((0x0000000000000000000000000000000000000000000000000000000000000000 0x0186a0) (0xb081963921826355dcb6c355ccf9c2637c18adf7d38ee44d803ea9ca41587e48c913d8d46896eb830aeadfc13144a8eac3 (() (q (51 0x6b7a83babea1eec790c947db4464ab657dbe9b887fe9acc247062847b8c2a8a9 0x0186a0)) ())) ) ((0x0000000000000000000000000000000000000000000000000000000000000001 0x0186a0) (0xb0a6207f5173ec41491d9f2c1b8fff5579e13703077e0eaca8fe587669dcccf51e9209a6b65576845ece5f7c2f3229e7e3 (() (q (51 0x24254a3efc3ebfac9979bbe0d615e2eda043aa329905f65b63846fa24149e2b6 0x0186a0)) ()))) ) """ ) start = 2 + 44 end = start + 238 p = DECOMPRESS_BLOCK.curry(DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, start, Program.to(end)) cost, out = p.run_with_cost(INFINITE_COST, [cse2, DESERIALIZE_MOD, [bytes(original_generator)]]) print() print(p) print(out) p_with_cses = DECOMPRESS_BLOCK.curry( DECOMPRESS_PUZZLE, DECOMPRESS_CSE_WITH_PREFIX, start, Program.to(end), cse2, DESERIALIZE_MOD ) generator_args = create_generator_args([SerializedProgram.from_bytes(original_generator)]) cost, out = p_with_cses.run_with_cost(INFINITE_COST, generator_args) print() print(p_with_cses) print(out)
true
true
1c484fa08fc49e2469b08d94e3f89720e8e00a3f
4,022
py
Python
utils/extra/common.py
ekojs/mitra
9c2458b7bf83af4a7e56b0e227f07454d99298e1
[ "MIT" ]
null
null
null
utils/extra/common.py
ekojs/mitra
9c2458b7bf83af4a7e56b0e227f07454d99298e1
[ "MIT" ]
null
null
null
utils/extra/common.py
ekojs/mitra
9c2458b7bf83af4a7e56b0e227f07454d99298e1
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # common functions # Ange Albertini 2020 import random import re from string import punctuation, digits, ascii_letters def randblock(l): return bytes([random.randrange(255) for i in range(l)]) # Cosmetic functions ########################################################### ASCII = (punctuation + digits + ascii_letters + " ").encode() def hexii(c): #replace 00 by empty char if c == b"\0": return b" " #replace printable char by .<char> if c in ASCII: return b" " + bytes([c]) if c == 0x0a: return b"\n" if c == b"\r": return b"\\r" #otherwise, return hex return b"%02X" % c def hexiis(s): return repr(b" ".join([hexii(c) for c in s]))[2:-1] def showsplit(d, i): WIDTH = 8 return "%s | %s" % (hexiis(d[i-WIDTH:i]), hexiis(d[i:i+WIDTH])) # 'GCM' functions ############################################################## def cut3(data, a): # skip 0:a[0] -- not needed ? return data[a[0]:a[1]], data[a[1]:a[2]], data[a[2]:] def mixfiles(d1, d2, cuts): """mixing data with exclusive parts of each data""" assert len(d1) == len(d2) d = b"" start = 0 keep = d1 skip = d2 for end in cuts: d += keep[start:end] start = end keep, skip = skip, keep d += keep[start:] return d def splitfile(data, cuts): p1 = b"" p2 = b"" start = 0 count = 0 for end in cuts: count += 1 p1 += data[start:end] p2 += randblock(end-start) start = end p1, p2 = p2, p1 p1 += data[end:] p2 += randblock(len(data)-end) assert len(p1) == len(p2) if count % 2 == 1: p1, p2 = p2, p1 return p1, p2 # PDF functions ################################################################ def EnclosedStringS(d, starts, ends): off = d.find(starts) return d[off:d.find(ends, off + len(starts))] def EnclosedString(d, starts, ends): off = d.find(starts) + len(starts) return d[off:d.find(ends, off)] def getCount(d): s = EnclosedString(d, b"/Count ", b"/") count = int(s) return count def getObjDecl(d, s): val = EnclosedString(d, s, b"0 R") val = val.strip() if val.decode().isnumeric(): return b"%s %s 0 R" % (s, val) else: return b"" def getValDecl(d, s): """locates declaration such as '/PageMode /UseOutlines' """ off = d.find(s) + len(s) if off == -1: return b"" match = re.match(b" *\/[A-Za-z0-9]*", d[off:]) if match is None: return b"" else: return b"%s %s" % (s, match[0]) def adjustToC(toc): """increasing page numbers of each ToC entry""" for entry in toc: d = entry[3] if d["kind"] == 1: d["page"] += 1 entry[2] += 1 return toc def adjustPDF(contents): startSig = contents.find(b"%PDF") # relative to file start startXREF = contents.find(b"\nxref\n0 ") + 1 endXREF = contents.find(b" \n\n", startXREF) + 1 origXref = contents[startXREF:endXREF] objCount = int(origXref.splitlines()[1].split(b" ")[1]) xrefLines = [ b"xref", b"0 %i" % objCount, # mutool declare its first xref like this b"0000000000 00001 f " ] i = 1 while i < objCount: # only very standard object declarations off = contents.find(b"\n%i 0 obj\n" % i) + 1 xrefLines.append(b"%010i 00000 n " % (off - startSig)) i += 1 xref = b"\n".join(xrefLines) # XREF length should be unchanged try: assert len(xref) == len(origXref) except AssertionError: print("<:", repr(origXref)) print(">:", repr(xref)) contents = contents[:startXREF] + xref + contents[endXREF:] startStartXref = contents.find(b"\nstartxref\n", endXREF) + len(b"\nstartxref\n") endStartXref = contents.find(b"\n%%EOF", startStartXref) contents = contents[:startStartXref] + b"%08i" % (startXREF - startSig) + contents[endStartXref:] return contents template = b"""%%PDF-1.3 %%\xC2\xB5\xC2\xB6 1 0 obj <</Length 2 0 R>> stream %(payload)s endstream endobj 2 0 obj %(payload_l)i endobj 3 0 obj << /Type /Catalog /Pages 4 0 R /Payload 1 0 R %% to prevent garbage collection %(extra)s %% optional: Names + OpenAction + Outlines + PageMode >> endobj 4 0 obj <</Type/Pages/Count %(count)i/Kids[%(kids)s]>> endobj """
19.812808
98
0.598707
import random import re from string import punctuation, digits, ascii_letters def randblock(l): return bytes([random.randrange(255) for i in range(l)])
true
true
1c484fc0efb0612c2fdd4194aa295cc755d295d9
6,922
py
Python
pyrsss/signal/spectrum.py
grawe/pyrsss
31fd88734b00f814e7aaa5829c4ac49c7bf53563
[ "MIT" ]
null
null
null
pyrsss/signal/spectrum.py
grawe/pyrsss
31fd88734b00f814e7aaa5829c4ac49c7bf53563
[ "MIT" ]
null
null
null
pyrsss/signal/spectrum.py
grawe/pyrsss
31fd88734b00f814e7aaa5829c4ac49c7bf53563
[ "MIT" ]
null
null
null
from __future__ import division import math import numpy as NP import scipy.signal def rect(t, a): """ Return a vector of the same length as $t$ that is equal to 1 for absolute values of $t$ less than $a$/2 and 0 otherwise. """ x = NP.zeros_like(t) x[NP.abs(t) < a/2] = 1 x[NP.abs(t) == a/2] = 1/2 return x def nextpow2(N): """ Return the power of 2 greater than or equal to *N*. """ return 2**int(math.ceil(math.log(N, 2))) def spectrum(x, n0=0, T_s=1, oversample=1, only_positive=True): """ Return the spectrum for the signal *x* calculated via FFT and the associated frequencies as a tuple. The *n0* parameter gives the index in *x* for time index 0 (*n0* = 0 means that `x[0]` is at time 0). The number of spectral samples returned is the next power of 2 greater than the length of *x* multiplied by *oversample*. If *only_positive*, return the spectrum only for positive frequencies (assuming *x* is real). """ assert oversample >= 1 and isinstance(oversample, int) N = nextpow2(len(x)) * 2**(oversample - 1) if only_positive: X = NP.fft.rfft(x, n=N) * T_s f = NP.fft.rfftfreq(N, d=T_s) else: X = NP.fft.fft(x, n=N) * T_s f = NP.fft.fftfreq(N, d=T_s) X = NP.fft.fftshift(X) f = NP.fft.fftshift(f) if n0 != 0: X *= NP.exp(-1j * 2 * math.pi * NP.arange(N) * n0 / N) return f, X def blackman_tukey(x, M, L, y=None, window='boxcar', d=1, full=False): """ Compute the Blackman-Tukey cross power spectral density (PSD) estimate between the time-domain signals *x* and *y* (must be the same length as *x*). If *y* is not given, compute the power spectral density estimate of *x*. Use the spectral window with identifier *window* (see the options in :func:scipy.`signal.get_window`, e.g., a tuple can be used to pass arguments to the window function) and length *M* (i.e., the maximum auto-correlation lag to include in the estimate). Compute the estimate at *L* uniformly spaced frequency samples where *d* is the time domain sample interval. If not *full*, return the tuple containing the length *L* PSD estimate and length *L* corresponding frequencies. If *full*, also return the estimated cross correlation and window function (i.e., a tuple with four elements). """ N = len(x) assert M <= N if y is None: y = x else: assert len(y) == N Rxy = scipy.signal.correlate(x, y) / N Rxy_window = Rxy[(N - 1) - M:(N - 1) + M + 1] window = scipy.signal.get_window(window, 2*M + 1, fftbins=False) k_range = NP.arange(0, L) shift = NP.exp(2j * NP.pi * k_range * M / L) Sxy = NP.fft.fft(window * Rxy_window, n=L) * shift f = NP.fft.fftfreq(L, d=d) if full: return (Sxy, f, Rxy, window) else: return (Sxy, f) def periodogram(x, L, y=None, d=1, full=False): """ Compute the periodogram of the cross power spectral density of *x* and *y*. The implementation is based on :func:`blackman-tukey`, following the same input and output conventions. """ return blackman_tukey(x, len(x) - 1, L, y=y, d=d, full=full) def etfe(x, y, M, L, d=1, window='parzen'): """ Compute the empirical transfer function estimate (ETFE) relating the input time series *x* to the output time series *y*. Compute the response at *L* equally spaced frequency samples (where the sampling period is *D*). Limit the correlations to a lag of *M* (and *M* <= len(*x*) - 1) and use the window function *window* (see :func:`scipy.signal.get_window`). Return the tuple containing the ETFE and the frequency sample points. See Section 6.3 of Ljung, System Identification Theory for the User, 2nd Edition. """ Phi_yu, f = blackman_tukey(y, M, L, y=x, d=d, window=window) Phi_u, _ = blackman_tukey(x, M, L, d=d, window=window) return Phi_yu / Phi_u, f if __name__ == '__main__': import pylab as PL # Reproduction of Oppenheim and Schafer (O&S), 3rd edition, # Example 10.4. The example considers the effects of windowing and # frequency sampling of a two sinusoid superposition example. N = 64 # number of samples (window length) fs = 10e3 # sampling frequency (Hz) T = 1 / fs # sample period (s) def W_r(w, L): """ DTFT of rectangular window of length *L* evaluated an angular frequencies *w*. See O&S (10.11). """ return NP.exp(-1j * w * (L - 1) / 2) * NP.sin(w * L / 2) / NP.sin(w / 2) K = 2048 # number of frequency samples w = NP.linspace(-math.pi, math.pi, K) W = W_r(w, N) n = NP.arange(N) v = NP.cos(2*math.pi/14 * n) + 0.75 * NP.cos(4*math.pi/15 * n) def V(w, A0, w0, A1, w1, L): """ DTFT of the superposition of two sinusoids with amplitudes *A0* and *A1*, angular frequencies *w0* and *w1*, and a recangular window length *L* and angular frequencies *w*. """ V1 = A0 / 2 * W_r(w - w0, L) V2 = A0 / 2 * W_r(w + w0, L) V3 = A1 / 2 * W_r(w - w1, L) V4 = A1 / 2 * W_r(w + w1, L) return V1 + V2 + V3 + V4 A0 = 1 A1 = 0.75 f2 = NP.linspace(-fs/2, fs/2, K) w2 = f2 / fs * 2 * math.pi V_dtft = V(w2, A0, 2*math.pi/14, A1, 4*math.pi/15, N) V_spectrum, f_spectrum = spectrum(v, T_s=T, only_positive=False) V_spectrum2, f_spectrum2 = spectrum(v, T_s=T, oversample=2, only_positive=False) V_spectrum3, f_spectrum3 = spectrum(v, T_s=T) PL.figure(figsize=(10, 4)) PL.plot(f2, NP.abs(V_dtft) * T, c='C0', label='DTFT (scaled)') PL.scatter(f_spectrum, NP.abs(V_spectrum), c='C1', s=10, label='spectrum') PL.scatter(f_spectrum2, NP.abs(V_spectrum2), c='C2', zorder=-1, s=20, label='spectrum (oversample=2)') PL.scatter(f_spectrum3, NP.abs(V_spectrum3), c='C3', zorder=-2, s=40, label='spectrum (positive-only)') PL.legend() PL.xlabel('Frequency (Hz)') PL.ylabel('Amplitude') PL.title('Comparison of pyrsss spectrum and scaled DTFT') PL.show()
31.463636
80
0.545652
from __future__ import division import math import numpy as NP import scipy.signal def rect(t, a): x = NP.zeros_like(t) x[NP.abs(t) < a/2] = 1 x[NP.abs(t) == a/2] = 1/2 return x def nextpow2(N): return 2**int(math.ceil(math.log(N, 2))) def spectrum(x, n0=0, T_s=1, oversample=1, only_positive=True): assert oversample >= 1 and isinstance(oversample, int) N = nextpow2(len(x)) * 2**(oversample - 1) if only_positive: X = NP.fft.rfft(x, n=N) * T_s f = NP.fft.rfftfreq(N, d=T_s) else: X = NP.fft.fft(x, n=N) * T_s f = NP.fft.fftfreq(N, d=T_s) X = NP.fft.fftshift(X) f = NP.fft.fftshift(f) if n0 != 0: X *= NP.exp(-1j * 2 * math.pi * NP.arange(N) * n0 / N) return f, X def blackman_tukey(x, M, L, y=None, window='boxcar', d=1, full=False): N = len(x) assert M <= N if y is None: y = x else: assert len(y) == N Rxy = scipy.signal.correlate(x, y) / N Rxy_window = Rxy[(N - 1) - M:(N - 1) + M + 1] window = scipy.signal.get_window(window, 2*M + 1, fftbins=False) k_range = NP.arange(0, L) shift = NP.exp(2j * NP.pi * k_range * M / L) Sxy = NP.fft.fft(window * Rxy_window, n=L) * shift f = NP.fft.fftfreq(L, d=d) if full: return (Sxy, f, Rxy, window) else: return (Sxy, f) def periodogram(x, L, y=None, d=1, full=False): return blackman_tukey(x, len(x) - 1, L, y=y, d=d, full=full) def etfe(x, y, M, L, d=1, window='parzen'): Phi_yu, f = blackman_tukey(y, M, L, y=x, d=d, window=window) Phi_u, _ = blackman_tukey(x, M, L, d=d, window=window) return Phi_yu / Phi_u, f if __name__ == '__main__': import pylab as PL N = 64 fs = 10e3 T = 1 / fs def W_r(w, L): return NP.exp(-1j * w * (L - 1) / 2) * NP.sin(w * L / 2) / NP.sin(w / 2) K = 2048 w = NP.linspace(-math.pi, math.pi, K) W = W_r(w, N) n = NP.arange(N) v = NP.cos(2*math.pi/14 * n) + 0.75 * NP.cos(4*math.pi/15 * n) def V(w, A0, w0, A1, w1, L): V1 = A0 / 2 * W_r(w - w0, L) V2 = A0 / 2 * W_r(w + w0, L) V3 = A1 / 2 * W_r(w - w1, L) V4 = A1 / 2 * W_r(w + w1, L) return V1 + V2 + V3 + V4 A0 = 1 A1 = 0.75 f2 = NP.linspace(-fs/2, fs/2, K) w2 = f2 / fs * 2 * math.pi V_dtft = V(w2, A0, 2*math.pi/14, A1, 4*math.pi/15, N) V_spectrum, f_spectrum = spectrum(v, T_s=T, only_positive=False) V_spectrum2, f_spectrum2 = spectrum(v, T_s=T, oversample=2, only_positive=False) V_spectrum3, f_spectrum3 = spectrum(v, T_s=T) PL.figure(figsize=(10, 4)) PL.plot(f2, NP.abs(V_dtft) * T, c='C0', label='DTFT (scaled)') PL.scatter(f_spectrum, NP.abs(V_spectrum), c='C1', s=10, label='spectrum') PL.scatter(f_spectrum2, NP.abs(V_spectrum2), c='C2', zorder=-1, s=20, label='spectrum (oversample=2)') PL.scatter(f_spectrum3, NP.abs(V_spectrum3), c='C3', zorder=-2, s=40, label='spectrum (positive-only)') PL.legend() PL.xlabel('Frequency (Hz)') PL.ylabel('Amplitude') PL.title('Comparison of pyrsss spectrum and scaled DTFT') PL.show()
true
true
1c485046ce096457e354bad1db4cbc7a66d76bb5
2,862
py
Python
awardsApp/api.py
MutuaFranklin/App-Awards
020c85db144156ec02f12815cd675245d4ad9db3
[ "MIT" ]
null
null
null
awardsApp/api.py
MutuaFranklin/App-Awards
020c85db144156ec02f12815cd675245d4ad9db3
[ "MIT" ]
null
null
null
awardsApp/api.py
MutuaFranklin/App-Awards
020c85db144156ec02f12815cd675245d4ad9db3
[ "MIT" ]
null
null
null
from django.http import JsonResponse from django.http import HttpResponse, Http404,HttpResponseRedirect from .permissions import IsAdminOrReadOnly from rest_framework.permissions import IsAuthenticated from rest_framework.response import Response from rest_framework.views import APIView from profiles.models import Profile from .models import Project from .serializer import ProfileSerializer, ProjectSerializer from rest_framework import viewsets from rest_framework import status # class ProfileViewSet(viewsets.ModelViewSet): # queryset = Profile.objects.all() # serializer_class = ProfileSerializer # class ProjectViewSet(viewsets.ModelViewSet): # queryset = Project.objects.all() # serializer_class = ProjectSerializer #LMS class ProfileList(APIView): permission_classes = (IsAdminOrReadOnly,) def get(self, request, format=None): profiles = Profile.objects.all() serializers = ProfileSerializer(profiles, many=True) return Response(serializers.data) def post(self, request, format=None): serializers = Profile(data=request.data) if serializers.is_valid(): serializers.save() return Response(serializers.data, status=status.HTTP_201_CREATED) return Response(serializers.errors, status=status.HTTP_400_BAD_REQUEST) class ProjectList(APIView): permission_classes = (IsAuthenticated,) def get(self, request, format=None): projects = Project.objects.all() serializers = ProjectSerializer(projects, many=True) return Response(serializers.data) def post(self, request, format=None): serializers = Project(data=request.data) if serializers.is_valid(): serializers.save() return Response(serializers.data, status=status.HTTP_201_CREATED) return Response(serializers.errors, status=status.HTTP_400_BAD_REQUEST) class ProjectDescription(APIView): permission_classes = (IsAuthenticated,) def get_project(self, pk): try: return Project.objects.get(pk=pk) except Project.DoesNotExist: return Http404 def get(self, request, pk, format=None): project = self.get_project(pk) serializers = ProjectSerializer(project) return Response(serializers.data) def put(self, request, pk, format=None): project = self.get_project()(pk) serializers = ProjectSerializer(project, request.data) if serializers.is_valid(): serializers.save() return Response(serializers.data) else: return Response(serializers.errors, status=status.HTTP_400_BAD_REQUEST) def delete(self, request, pk, format=None): project = self.get_project(pk) project.delete() return Response(status=status.HTTP_204_NO_CONTENT)
30.774194
83
0.710342
from django.http import JsonResponse from django.http import HttpResponse, Http404,HttpResponseRedirect from .permissions import IsAdminOrReadOnly from rest_framework.permissions import IsAuthenticated from rest_framework.response import Response from rest_framework.views import APIView from profiles.models import Profile from .models import Project from .serializer import ProfileSerializer, ProjectSerializer from rest_framework import viewsets from rest_framework import status class ProfileList(APIView): permission_classes = (IsAdminOrReadOnly,) def get(self, request, format=None): profiles = Profile.objects.all() serializers = ProfileSerializer(profiles, many=True) return Response(serializers.data) def post(self, request, format=None): serializers = Profile(data=request.data) if serializers.is_valid(): serializers.save() return Response(serializers.data, status=status.HTTP_201_CREATED) return Response(serializers.errors, status=status.HTTP_400_BAD_REQUEST) class ProjectList(APIView): permission_classes = (IsAuthenticated,) def get(self, request, format=None): projects = Project.objects.all() serializers = ProjectSerializer(projects, many=True) return Response(serializers.data) def post(self, request, format=None): serializers = Project(data=request.data) if serializers.is_valid(): serializers.save() return Response(serializers.data, status=status.HTTP_201_CREATED) return Response(serializers.errors, status=status.HTTP_400_BAD_REQUEST) class ProjectDescription(APIView): permission_classes = (IsAuthenticated,) def get_project(self, pk): try: return Project.objects.get(pk=pk) except Project.DoesNotExist: return Http404 def get(self, request, pk, format=None): project = self.get_project(pk) serializers = ProjectSerializer(project) return Response(serializers.data) def put(self, request, pk, format=None): project = self.get_project()(pk) serializers = ProjectSerializer(project, request.data) if serializers.is_valid(): serializers.save() return Response(serializers.data) else: return Response(serializers.errors, status=status.HTTP_400_BAD_REQUEST) def delete(self, request, pk, format=None): project = self.get_project(pk) project.delete() return Response(status=status.HTTP_204_NO_CONTENT)
true
true
1c48504c2a6e00bf9546d691cb3602dd96353db6
193
py
Python
language/equivalence/formation.py
jedhsu/language
3772a4a0ff287e1fc5ebefc716b8d91928d04c72
[ "MIT" ]
null
null
null
language/equivalence/formation.py
jedhsu/language
3772a4a0ff287e1fc5ebefc716b8d91928d04c72
[ "MIT" ]
null
null
null
language/equivalence/formation.py
jedhsu/language
3772a4a0ff287e1fc5ebefc716b8d91928d04c72
[ "MIT" ]
null
null
null
""" *Product Formation* A, B: Type _proves_ A x B: Type "The prevalence of products of types concides with the prevalence of cartesian products in categories." (Corfield 40) """
21.444444
64
0.689119
true
true
1c48511bdf2bd0df09a51e286757ce2441bb1185
441
py
Python
pkgs/ops-pkg/src/genie/libs/ops/ospf/iosxe/yang/ospf.py
miott/genielibs
6464642cdd67aa2367bdbb12561af4bb060e5e62
[ "Apache-2.0" ]
94
2018-04-30T20:29:15.000Z
2022-03-29T13:40:31.000Z
pkgs/ops-pkg/src/genie/libs/ops/ospf/iosxe/yang/ospf.py
miott/genielibs
6464642cdd67aa2367bdbb12561af4bb060e5e62
[ "Apache-2.0" ]
67
2018-12-06T21:08:09.000Z
2022-03-29T18:00:46.000Z
pkgs/ops-pkg/src/genie/libs/ops/ospf/iosxe/yang/ospf.py
miott/genielibs
6464642cdd67aa2367bdbb12561af4bb060e5e62
[ "Apache-2.0" ]
49
2018-06-29T18:59:03.000Z
2022-03-10T02:07:59.000Z
from genie.ops.base import Context from genie.libs.ops.ospf.iosxe.ospf import Ospf as b_ospf from genie.libs.parser.iosxe import show_ospf class Ospf(b_ospf): '''Ospf Ops Object''' # To keep short names def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.context_manager[show_ospf.ShowIpOspf] = Context.yang # Rest use cli as their info cannot be retrieve via yang at the moment
33.923077
78
0.705215
from genie.ops.base import Context from genie.libs.ops.ospf.iosxe.ospf import Ospf as b_ospf from genie.libs.parser.iosxe import show_ospf class Ospf(b_ospf): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.context_manager[show_ospf.ShowIpOspf] = Context.yang
true
true
1c48511d9b3288b699faa35bf674bb9a5336cf68
1,995
py
Python
Code/test-PSO.py
gitFloyd/AAI-Project-2
c6bb4d389248c3385e58a0c399343322a6dd887f
[ "MIT" ]
null
null
null
Code/test-PSO.py
gitFloyd/AAI-Project-2
c6bb4d389248c3385e58a0c399343322a6dd887f
[ "MIT" ]
null
null
null
Code/test-PSO.py
gitFloyd/AAI-Project-2
c6bb4d389248c3385e58a0c399343322a6dd887f
[ "MIT" ]
null
null
null
import random #random.seed(111) import PSO import Model import Dataset as DS from Log import Log Dataset = DS.Dataset Pistachio = DS.Pistachio Layer = Model.Layer DenseLayer = Model.DenseLayer SparseLayer = Model.SparseLayer FMLayer = Model.FuzzyMembershipLayer InputLayer = Model.InputLayer # particles, weights # X = [1,2,3,4,5] # y = [1,0,0] pistachios = Pistachio(Dataset.LINUX_NL) pistachios.Load() pistachios.Shuffle() offset = 0 data = pistachios.Fetch('Area', 'Solidity', 'Roundness', 'Compactness', 'Shapefactor_1', 'Class', limit = 100, offset = offset) X = [row[0:-1] for row in data] y = [row[-1] for row in data] val_data = pistachios.Fetch('Area', 'Solidity', 'Roundness', 'Compactness', 'Shapefactor_1', 'Class', limit = 10, offset = offset + 100) val_X = [row[0:-1] for row in val_data] val_y = [row[-1] for row in val_data] #randConnections = SparseLayer.RandConnections(len(X[0]), len(X[0])*2, len(X[0])) myModel = Model.Model([ InputLayer(len(X[0])), #SparseLayer(randConnections, Layer.RELU), FMLayer(len(X[0]), len(X[0]) * 2, Layer.RELU), DenseLayer(128, Layer.RELU), DenseLayer(16, Layer.RELU), DenseLayer(2, Layer.SOFTMAX), ]) numInputs = 20 numParticles = 20 numWeights = myModel.NumWeights() psoTest = PSO.PSO(myModel, (numParticles, myModel.NumWeights())) records = psoTest.ExecuteMany(X[0:numInputs], y[0:numInputs], iterations=20) # bestWeights = records[-1][1] bestWeights = records.pop()[1] for i in range(10): # print(records[i][1][0:10]) pass # Log.Print() for i in range(len(val_X)): #print(val_X[i]) print('Predict: {}; Actual: {}; val_X: {}'.format(myModel.Execute(val_X[i], bestWeights), val_y[i], val_X[i])) print('---------------------') for i in range(20): #print(val_X[i]) print('Predict: {}'.format(myModel.Execute([random.randint(-1,0) + random.random() for _ in range(5)], bestWeights))) Log.Add('bestWeights', '{}'.format(bestWeights)) Log.SaveAll()
27.328767
137
0.663659
import random import PSO import Model import Dataset as DS from Log import Log Dataset = DS.Dataset Pistachio = DS.Pistachio Layer = Model.Layer DenseLayer = Model.DenseLayer SparseLayer = Model.SparseLayer FMLayer = Model.FuzzyMembershipLayer InputLayer = Model.InputLayer pistachios = Pistachio(Dataset.LINUX_NL) pistachios.Load() pistachios.Shuffle() offset = 0 data = pistachios.Fetch('Area', 'Solidity', 'Roundness', 'Compactness', 'Shapefactor_1', 'Class', limit = 100, offset = offset) X = [row[0:-1] for row in data] y = [row[-1] for row in data] val_data = pistachios.Fetch('Area', 'Solidity', 'Roundness', 'Compactness', 'Shapefactor_1', 'Class', limit = 10, offset = offset + 100) val_X = [row[0:-1] for row in val_data] val_y = [row[-1] for row in val_data] myModel = Model.Model([ InputLayer(len(X[0])), FMLayer(len(X[0]), len(X[0]) * 2, Layer.RELU), DenseLayer(128, Layer.RELU), DenseLayer(16, Layer.RELU), DenseLayer(2, Layer.SOFTMAX), ]) numInputs = 20 numParticles = 20 numWeights = myModel.NumWeights() psoTest = PSO.PSO(myModel, (numParticles, myModel.NumWeights())) records = psoTest.ExecuteMany(X[0:numInputs], y[0:numInputs], iterations=20) bestWeights = records.pop()[1] for i in range(10): pass for i in range(len(val_X)): print('Predict: {}; Actual: {}; val_X: {}'.format(myModel.Execute(val_X[i], bestWeights), val_y[i], val_X[i])) print('---------------------') for i in range(20): print('Predict: {}'.format(myModel.Execute([random.randint(-1,0) + random.random() for _ in range(5)], bestWeights))) Log.Add('bestWeights', '{}'.format(bestWeights)) Log.SaveAll()
true
true
1c4852d27622febf8afb7626b16291be56d91b72
2,894
py
Python
vimcc.py
joas77/vim-winccoa
46c84244e5bf6679e7ef00aaf814bcfdedb596b6
[ "MIT" ]
2
2021-03-12T04:48:48.000Z
2021-09-27T15:09:33.000Z
vimcc.py
joas77/vim-winccoa
46c84244e5bf6679e7ef00aaf814bcfdedb596b6
[ "MIT" ]
null
null
null
vimcc.py
joas77/vim-winccoa
46c84244e5bf6679e7ef00aaf814bcfdedb596b6
[ "MIT" ]
1
2021-04-15T18:13:50.000Z
2021-04-15T18:13:50.000Z
#!/usr/bin/env python3 import sys import os import configparser import shutil import subprocess def find_project_config_file(directory, max_levels = 5): if max_levels == 0 or directory == "/" or directory == "": return None #Config file not found fileName = os.path.join(directory, "config/config") if os.path.isfile(fileName): return fileName else: return find_project_config_file(os.path.dirname(directory), max_levels-1) #Returns true: success, false: no success def copy_file_from_subprojects(file_name): abs_file_name = os.path.abspath(file_name) #Find project config file cfgFile = find_project_config_file(os.path.dirname(abs_file_name)) if not cfgFile: print("Could not find the config file of the WinCC project for '{}'!" "".format(file_name)) return False config = configparser.RawConfigParser(dict_type=MultiOrderedDict, strict=False) if not config.read(cfgFile): print("Could not parse config file '{}'!".format(cfgFile)) return False projDirs = config.get("general", "proj_path", fallback=None) if not projDirs: print("Could not parse general:proj_path from '{}'!".format(cfgFile)) return False projDirs = [ directory.replace('"', '') for directory in projDirs] #Main and sub project directories mainDir = projDirs.pop(-1) subDirs = projDirs rel_file_name = abs_file_name.replace(mainDir+"/", "") for subDir in subDirs[::-1]: #subDirs are listed from least to highest priority new_file_name = os.path.join(subDir, rel_file_name) if os.path.isfile(new_file_name): destDir = os.path.dirname(file_name) if destDir != "": os.makedirs(os.path.dirname(file_name), exist_ok=True) #Create dir if needed shutil.copy2(new_file_name, file_name) print("Copied '{}' to '{}'!".format(new_file_name, file_name)) return True print("File '{}' not found in subprojects.".format(file_name)) return False #Necessary to load duplicate keys/options from config file from collections import OrderedDict class MultiOrderedDict(OrderedDict): def __setitem__(self, key, value): if isinstance(value, list) and key in self: self[key].extend(value) else: super(OrderedDict, self).__setitem__(key, value) def keys(self): return super(OrderedDict, self).keys() if __name__ == "__main__": allFiles = sys.argv[1:] #All provided files #Existing and missing files in current directory existingFiles = [] missingFiles = [] for file in allFiles: if os.path.isfile(file) or copy_file_from_subprojects(file): existingFiles.append(file) else: missingFiles.append(file) subprocess.run(["vim"]+existingFiles+missingFiles)
32.516854
93
0.666551
import sys import os import configparser import shutil import subprocess def find_project_config_file(directory, max_levels = 5): if max_levels == 0 or directory == "/" or directory == "": return None fileName = os.path.join(directory, "config/config") if os.path.isfile(fileName): return fileName else: return find_project_config_file(os.path.dirname(directory), max_levels-1) def copy_file_from_subprojects(file_name): abs_file_name = os.path.abspath(file_name) cfgFile = find_project_config_file(os.path.dirname(abs_file_name)) if not cfgFile: print("Could not find the config file of the WinCC project for '{}'!" "".format(file_name)) return False config = configparser.RawConfigParser(dict_type=MultiOrderedDict, strict=False) if not config.read(cfgFile): print("Could not parse config file '{}'!".format(cfgFile)) return False projDirs = config.get("general", "proj_path", fallback=None) if not projDirs: print("Could not parse general:proj_path from '{}'!".format(cfgFile)) return False projDirs = [ directory.replace('"', '') for directory in projDirs] #Main and sub project directories mainDir = projDirs.pop(-1) subDirs = projDirs rel_file_name = abs_file_name.replace(mainDir+"/", "") for subDir in subDirs[::-1]: #subDirs are listed from least to highest priority new_file_name = os.path.join(subDir, rel_file_name) if os.path.isfile(new_file_name): destDir = os.path.dirname(file_name) if destDir != "": os.makedirs(os.path.dirname(file_name), exist_ok=True) #Create dir if needed shutil.copy2(new_file_name, file_name) print("Copied '{}' to '{}'!".format(new_file_name, file_name)) return True print("File '{}' not found in subprojects.".format(file_name)) return False #Necessary to load duplicate keys/options from config file from collections import OrderedDict class MultiOrderedDict(OrderedDict): def __setitem__(self, key, value): if isinstance(value, list) and key in self: self[key].extend(value) else: super(OrderedDict, self).__setitem__(key, value) def keys(self): return super(OrderedDict, self).keys() if __name__ == "__main__": allFiles = sys.argv[1:] #All provided files #Existing and missing files in current directory existingFiles = [] missingFiles = [] for file in allFiles: if os.path.isfile(file) or copy_file_from_subprojects(file): existingFiles.append(file) else: missingFiles.append(file) subprocess.run(["vim"]+existingFiles+missingFiles)
true
true
1c48532d9b777c913fbc3fc8cf4210092c8650ef
101
py
Python
other/exawizards2019_b.py
ryosuke0825/atcoder_python
185cdbe7db44ecca1aaf357858d16d31ce515ddb
[ "MIT" ]
null
null
null
other/exawizards2019_b.py
ryosuke0825/atcoder_python
185cdbe7db44ecca1aaf357858d16d31ce515ddb
[ "MIT" ]
null
null
null
other/exawizards2019_b.py
ryosuke0825/atcoder_python
185cdbe7db44ecca1aaf357858d16d31ce515ddb
[ "MIT" ]
null
null
null
n = int(input()) s = input() if s.count('R') > s.count('B'): print('Yes') else: print('No')
12.625
31
0.49505
n = int(input()) s = input() if s.count('R') > s.count('B'): print('Yes') else: print('No')
true
true
1c48542a1c9ecd55c6e00af6037038147a68539f
59,704
py
Python
sklearn/decomposition/_dict_learning.py
emarkou/scikit-learn
d73822f84f2832dcc25f0ff58769f60871a78025
[ "BSD-3-Clause" ]
1
2021-05-25T18:06:44.000Z
2021-05-25T18:06:44.000Z
sklearn/decomposition/_dict_learning.py
emarkou/scikit-learn
d73822f84f2832dcc25f0ff58769f60871a78025
[ "BSD-3-Clause" ]
null
null
null
sklearn/decomposition/_dict_learning.py
emarkou/scikit-learn
d73822f84f2832dcc25f0ff58769f60871a78025
[ "BSD-3-Clause" ]
null
null
null
""" Dictionary learning. """ # Author: Vlad Niculae, Gael Varoquaux, Alexandre Gramfort # License: BSD 3 clause import time import sys import itertools import warnings from math import ceil import numpy as np from scipy import linalg from joblib import Parallel, effective_n_jobs from ..base import BaseEstimator, TransformerMixin from ..utils import deprecated from ..utils import (check_array, check_random_state, gen_even_slices, gen_batches) from ..utils.extmath import randomized_svd, row_norms, svd_flip from ..utils.validation import check_is_fitted from ..utils.fixes import delayed from ..linear_model import Lasso, orthogonal_mp_gram, LassoLars, Lars def _check_positive_coding(method, positive): if positive and method in ["omp", "lars"]: raise ValueError( "Positive constraint not supported for '{}' " "coding method.".format(method) ) def _sparse_encode(X, dictionary, gram, cov=None, algorithm='lasso_lars', regularization=None, copy_cov=True, init=None, max_iter=1000, check_input=True, verbose=0, positive=False): """Generic sparse coding. Each column of the result is the solution to a Lasso problem. Parameters ---------- X : ndarray of shape (n_samples, n_features) Data matrix. dictionary : ndarray of shape (n_components, n_features) The dictionary matrix against which to solve the sparse coding of the data. Some of the algorithms assume normalized rows. gram : ndarray of shape (n_components, n_components) or None Precomputed Gram matrix, `dictionary * dictionary'` gram can be `None` if method is 'threshold'. cov : ndarray of shape (n_components, n_samples), default=None Precomputed covariance, `dictionary * X'`. algorithm : {'lasso_lars', 'lasso_cd', 'lars', 'omp', 'threshold'}, \ default='lasso_lars' The algorithm used: * `'lars'`: uses the least angle regression method (`linear_model.lars_path`); * `'lasso_lars'`: uses Lars to compute the Lasso solution; * `'lasso_cd'`: uses the coordinate descent method to compute the Lasso solution (`linear_model.Lasso`). lasso_lars will be faster if the estimated components are sparse; * `'omp'`: uses orthogonal matching pursuit to estimate the sparse solution; * `'threshold'`: squashes to zero all coefficients less than regularization from the projection `dictionary * data'`. regularization : int or float, default=None The regularization parameter. It corresponds to alpha when algorithm is `'lasso_lars'`, `'lasso_cd'` or `'threshold'`. Otherwise it corresponds to `n_nonzero_coefs`. init : ndarray of shape (n_samples, n_components), default=None Initialization value of the sparse code. Only used if `algorithm='lasso_cd'`. max_iter : int, default=1000 Maximum number of iterations to perform if `algorithm='lasso_cd'` or `'lasso_lars'`. copy_cov : bool, default=True Whether to copy the precomputed covariance matrix; if `False`, it may be overwritten. check_input : bool, default=True If `False`, the input arrays `X` and dictionary will not be checked. verbose : int, default=0 Controls the verbosity; the higher, the more messages. positive: bool, default=False Whether to enforce a positivity constraint on the sparse code. .. versionadded:: 0.20 Returns ------- code : ndarray of shape (n_components, n_features) The sparse codes. See Also -------- sklearn.linear_model.lars_path sklearn.linear_model.orthogonal_mp sklearn.linear_model.Lasso SparseCoder """ if X.ndim == 1: X = X[:, np.newaxis] n_samples, n_features = X.shape n_components = dictionary.shape[0] if dictionary.shape[1] != X.shape[1]: raise ValueError("Dictionary and X have different numbers of features:" "dictionary.shape: {} X.shape{}".format( dictionary.shape, X.shape)) if cov is None and algorithm != 'lasso_cd': # overwriting cov is safe copy_cov = False cov = np.dot(dictionary, X.T) _check_positive_coding(algorithm, positive) if algorithm == 'lasso_lars': alpha = float(regularization) / n_features # account for scaling try: err_mgt = np.seterr(all='ignore') # Not passing in verbose=max(0, verbose-1) because Lars.fit already # corrects the verbosity level. lasso_lars = LassoLars(alpha=alpha, fit_intercept=False, verbose=verbose, normalize=False, precompute=gram, fit_path=False, positive=positive, max_iter=max_iter) lasso_lars.fit(dictionary.T, X.T, Xy=cov) new_code = lasso_lars.coef_ finally: np.seterr(**err_mgt) elif algorithm == 'lasso_cd': alpha = float(regularization) / n_features # account for scaling # TODO: Make verbosity argument for Lasso? # sklearn.linear_model.coordinate_descent.enet_path has a verbosity # argument that we could pass in from Lasso. clf = Lasso(alpha=alpha, fit_intercept=False, normalize=False, precompute=gram, max_iter=max_iter, warm_start=True, positive=positive) if init is not None: clf.coef_ = init clf.fit(dictionary.T, X.T, check_input=check_input) new_code = clf.coef_ elif algorithm == 'lars': try: err_mgt = np.seterr(all='ignore') # Not passing in verbose=max(0, verbose-1) because Lars.fit already # corrects the verbosity level. lars = Lars(fit_intercept=False, verbose=verbose, normalize=False, precompute=gram, n_nonzero_coefs=int(regularization), fit_path=False) lars.fit(dictionary.T, X.T, Xy=cov) new_code = lars.coef_ finally: np.seterr(**err_mgt) elif algorithm == 'threshold': new_code = ((np.sign(cov) * np.maximum(np.abs(cov) - regularization, 0)).T) if positive: np.clip(new_code, 0, None, out=new_code) elif algorithm == 'omp': new_code = orthogonal_mp_gram( Gram=gram, Xy=cov, n_nonzero_coefs=int(regularization), tol=None, norms_squared=row_norms(X, squared=True), copy_Xy=copy_cov).T else: raise ValueError('Sparse coding method must be "lasso_lars" ' '"lasso_cd", "lasso", "threshold" or "omp", got %s.' % algorithm) if new_code.ndim != 2: return new_code.reshape(n_samples, n_components) return new_code # XXX : could be moved to the linear_model module def sparse_encode(X, dictionary, *, gram=None, cov=None, algorithm='lasso_lars', n_nonzero_coefs=None, alpha=None, copy_cov=True, init=None, max_iter=1000, n_jobs=None, check_input=True, verbose=0, positive=False): """Sparse coding Each row of the result is the solution to a sparse coding problem. The goal is to find a sparse array `code` such that:: X ~= code * dictionary Read more in the :ref:`User Guide <SparseCoder>`. Parameters ---------- X : ndarray of shape (n_samples, n_features) Data matrix. dictionary : ndarray of shape (n_components, n_features) The dictionary matrix against which to solve the sparse coding of the data. Some of the algorithms assume normalized rows for meaningful output. gram : ndarray of shape (n_components, n_components), default=None Precomputed Gram matrix, `dictionary * dictionary'`. cov : ndarray of shape (n_components, n_samples), default=None Precomputed covariance, `dictionary' * X`. algorithm : {'lasso_lars', 'lasso_cd', 'lars', 'omp', 'threshold'}, \ default='lasso_lars' The algorithm used: * `'lars'`: uses the least angle regression method (`linear_model.lars_path`); * `'lasso_lars'`: uses Lars to compute the Lasso solution; * `'lasso_cd'`: uses the coordinate descent method to compute the Lasso solution (`linear_model.Lasso`). lasso_lars will be faster if the estimated components are sparse; * `'omp'`: uses orthogonal matching pursuit to estimate the sparse solution; * `'threshold'`: squashes to zero all coefficients less than regularization from the projection `dictionary * data'`. n_nonzero_coefs : int, default=None Number of nonzero coefficients to target in each column of the solution. This is only used by `algorithm='lars'` and `algorithm='omp'` and is overridden by `alpha` in the `omp` case. If `None`, then `n_nonzero_coefs=int(n_features / 10)`. alpha : float, default=None If `algorithm='lasso_lars'` or `algorithm='lasso_cd'`, `alpha` is the penalty applied to the L1 norm. If `algorithm='threshold'`, `alpha` is the absolute value of the threshold below which coefficients will be squashed to zero. If `algorithm='omp'`, `alpha` is the tolerance parameter: the value of the reconstruction error targeted. In this case, it overrides `n_nonzero_coefs`. If `None`, default to 1. copy_cov : bool, default=True Whether to copy the precomputed covariance matrix; if `False`, it may be overwritten. init : ndarray of shape (n_samples, n_components), default=None Initialization value of the sparse codes. Only used if `algorithm='lasso_cd'`. max_iter : int, default=1000 Maximum number of iterations to perform if `algorithm='lasso_cd'` or `'lasso_lars'`. n_jobs : int, default=None Number of parallel jobs to run. ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. ``-1`` means using all processors. See :term:`Glossary <n_jobs>` for more details. check_input : bool, default=True If `False`, the input arrays X and dictionary will not be checked. verbose : int, default=0 Controls the verbosity; the higher, the more messages. positive : bool, default=False Whether to enforce positivity when finding the encoding. .. versionadded:: 0.20 Returns ------- code : ndarray of shape (n_samples, n_components) The sparse codes See Also -------- sklearn.linear_model.lars_path sklearn.linear_model.orthogonal_mp sklearn.linear_model.Lasso SparseCoder """ if check_input: if algorithm == 'lasso_cd': dictionary = check_array(dictionary, order='C', dtype='float64') X = check_array(X, order='C', dtype='float64') else: dictionary = check_array(dictionary) X = check_array(X) n_samples, n_features = X.shape n_components = dictionary.shape[0] if gram is None and algorithm != 'threshold': gram = np.dot(dictionary, dictionary.T) if cov is None and algorithm != 'lasso_cd': copy_cov = False cov = np.dot(dictionary, X.T) if algorithm in ('lars', 'omp'): regularization = n_nonzero_coefs if regularization is None: regularization = min(max(n_features / 10, 1), n_components) else: regularization = alpha if regularization is None: regularization = 1. if effective_n_jobs(n_jobs) == 1 or algorithm == 'threshold': code = _sparse_encode(X, dictionary, gram, cov=cov, algorithm=algorithm, regularization=regularization, copy_cov=copy_cov, init=init, max_iter=max_iter, check_input=False, verbose=verbose, positive=positive) return code # Enter parallel code block code = np.empty((n_samples, n_components)) slices = list(gen_even_slices(n_samples, effective_n_jobs(n_jobs))) code_views = Parallel(n_jobs=n_jobs, verbose=verbose)( delayed(_sparse_encode)( X[this_slice], dictionary, gram, cov[:, this_slice] if cov is not None else None, algorithm, regularization=regularization, copy_cov=copy_cov, init=init[this_slice] if init is not None else None, max_iter=max_iter, check_input=False, verbose=verbose, positive=positive) for this_slice in slices) for this_slice, this_view in zip(slices, code_views): code[this_slice] = this_view return code def _update_dict(dictionary, Y, code, A=None, B=None, verbose=False, random_state=None, positive=False): """Update the dense dictionary factor in place. Parameters ---------- dictionary : ndarray of shape (n_components, n_features) Value of the dictionary at the previous iteration. Y : ndarray of shape (n_samples, n_features) Data matrix. code : ndarray of shape (n_samples, n_components) Sparse coding of the data against which to optimize the dictionary. A : ndarray of shape (n_components, n_components), default=None Together with `B`, sufficient stats of the online model to update the dictionary. B : ndarray of shape (n_features, n_components), default=None Together with `A`, sufficient stats of the online model to update the dictionary. verbose: bool, default=False Degree of output the procedure will print. random_state : int, RandomState instance or None, default=None Used for randomly initializing the dictionary. Pass an int for reproducible results across multiple function calls. See :term:`Glossary <random_state>`. positive : bool, default=False Whether to enforce positivity when finding the dictionary. .. versionadded:: 0.20 """ n_samples, n_components = code.shape random_state = check_random_state(random_state) if A is None: A = code.T @ code if B is None: B = Y.T @ code n_unused = 0 for k in range(n_components): if A[k, k] > 1e-6: # 1e-6 is arbitrary but consistent with the spams implementation dictionary[k] += (B[:, k] - A[k] @ dictionary) / A[k, k] else: # kth atom is almost never used -> sample a new one from the data newd = Y[random_state.choice(n_samples)] # add small noise to avoid making the sparse coding ill conditioned noise_level = 0.01 * (newd.std() or 1) # avoid 0 std noise = random_state.normal(0, noise_level, size=len(newd)) dictionary[k] = newd + noise code[:, k] = 0 n_unused += 1 if positive: np.clip(dictionary[k], 0, None, out=dictionary[k]) # Projection on the constraint set ||V_k|| == 1 dictionary[k] /= linalg.norm(dictionary[k]) if verbose and n_unused > 0: print(f"{n_unused} unused atoms resampled.") def dict_learning(X, n_components, *, alpha, max_iter=100, tol=1e-8, method='lars', n_jobs=None, dict_init=None, code_init=None, callback=None, verbose=False, random_state=None, return_n_iter=False, positive_dict=False, positive_code=False, method_max_iter=1000): """Solves a dictionary learning matrix factorization problem. Finds the best dictionary and the corresponding sparse code for approximating the data matrix X by solving:: (U^*, V^*) = argmin 0.5 || X - U V ||_2^2 + alpha * || U ||_1 (U,V) with || V_k ||_2 = 1 for all 0 <= k < n_components where V is the dictionary and U is the sparse code. Read more in the :ref:`User Guide <DictionaryLearning>`. Parameters ---------- X : ndarray of shape (n_samples, n_features) Data matrix. n_components : int Number of dictionary atoms to extract. alpha : int Sparsity controlling parameter. max_iter : int, default=100 Maximum number of iterations to perform. tol : float, default=1e-8 Tolerance for the stopping condition. method : {'lars', 'cd'}, default='lars' The method used: * `'lars'`: uses the least angle regression method to solve the lasso problem (`linear_model.lars_path`); * `'cd'`: uses the coordinate descent method to compute the Lasso solution (`linear_model.Lasso`). Lars will be faster if the estimated components are sparse. n_jobs : int, default=None Number of parallel jobs to run. ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. ``-1`` means using all processors. See :term:`Glossary <n_jobs>` for more details. dict_init : ndarray of shape (n_components, n_features), default=None Initial value for the dictionary for warm restart scenarios. Only used if `code_init` and `dict_init` are not None. code_init : ndarray of shape (n_samples, n_components), default=None Initial value for the sparse code for warm restart scenarios. Only used if `code_init` and `dict_init` are not None. callback : callable, default=None Callable that gets invoked every five iterations verbose : bool, default=False To control the verbosity of the procedure. random_state : int, RandomState instance or None, default=None Used for randomly initializing the dictionary. Pass an int for reproducible results across multiple function calls. See :term:`Glossary <random_state>`. return_n_iter : bool, default=False Whether or not to return the number of iterations. positive_dict : bool, default=False Whether to enforce positivity when finding the dictionary. .. versionadded:: 0.20 positive_code : bool, default=False Whether to enforce positivity when finding the code. .. versionadded:: 0.20 method_max_iter : int, default=1000 Maximum number of iterations to perform. .. versionadded:: 0.22 Returns ------- code : ndarray of shape (n_samples, n_components) The sparse code factor in the matrix factorization. dictionary : ndarray of shape (n_components, n_features), The dictionary factor in the matrix factorization. errors : array Vector of errors at each iteration. n_iter : int Number of iterations run. Returned only if `return_n_iter` is set to True. See Also -------- dict_learning_online DictionaryLearning MiniBatchDictionaryLearning SparsePCA MiniBatchSparsePCA """ if method not in ('lars', 'cd'): raise ValueError('Coding method %r not supported as a fit algorithm.' % method) _check_positive_coding(method, positive_code) method = 'lasso_' + method t0 = time.time() # Avoid integer division problems alpha = float(alpha) random_state = check_random_state(random_state) # Init the code and the dictionary with SVD of Y if code_init is not None and dict_init is not None: code = np.array(code_init, order='F') # Don't copy V, it will happen below dictionary = dict_init else: code, S, dictionary = linalg.svd(X, full_matrices=False) # flip the initial code's sign to enforce deterministic output code, dictionary = svd_flip(code, dictionary) dictionary = S[:, np.newaxis] * dictionary r = len(dictionary) if n_components <= r: # True even if n_components=None code = code[:, :n_components] dictionary = dictionary[:n_components, :] else: code = np.c_[code, np.zeros((len(code), n_components - r))] dictionary = np.r_[dictionary, np.zeros((n_components - r, dictionary.shape[1]))] # Fortran-order dict better suited for the sparse coding which is the # bottleneck of this algorithm. dictionary = np.asfortranarray(dictionary) errors = [] current_cost = np.nan if verbose == 1: print('[dict_learning]', end=' ') # If max_iter is 0, number of iterations returned should be zero ii = -1 for ii in range(max_iter): dt = (time.time() - t0) if verbose == 1: sys.stdout.write(".") sys.stdout.flush() elif verbose: print("Iteration % 3i " "(elapsed time: % 3is, % 4.1fmn, current cost % 7.3f)" % (ii, dt, dt / 60, current_cost)) # Update code code = sparse_encode(X, dictionary, algorithm=method, alpha=alpha, init=code, n_jobs=n_jobs, positive=positive_code, max_iter=method_max_iter, verbose=verbose) # Update dictionary in place _update_dict(dictionary, X, code, verbose=verbose, random_state=random_state, positive=positive_dict) # Cost function current_cost = (0.5 * np.sum((X - code @ dictionary)**2) + alpha * np.sum(np.abs(code))) errors.append(current_cost) if ii > 0: dE = errors[-2] - errors[-1] # assert(dE >= -tol * errors[-1]) if dE < tol * errors[-1]: if verbose == 1: # A line return print("") elif verbose: print("--- Convergence reached after %d iterations" % ii) break if ii % 5 == 0 and callback is not None: callback(locals()) if return_n_iter: return code, dictionary, errors, ii + 1 else: return code, dictionary, errors def dict_learning_online(X, n_components=2, *, alpha=1, n_iter=100, return_code=True, dict_init=None, callback=None, batch_size=3, verbose=False, shuffle=True, n_jobs=None, method='lars', iter_offset=0, random_state=None, return_inner_stats=False, inner_stats=None, return_n_iter=False, positive_dict=False, positive_code=False, method_max_iter=1000): """Solves a dictionary learning matrix factorization problem online. Finds the best dictionary and the corresponding sparse code for approximating the data matrix X by solving:: (U^*, V^*) = argmin 0.5 || X - U V ||_2^2 + alpha * || U ||_1 (U,V) with || V_k ||_2 = 1 for all 0 <= k < n_components where V is the dictionary and U is the sparse code. This is accomplished by repeatedly iterating over mini-batches by slicing the input data. Read more in the :ref:`User Guide <DictionaryLearning>`. Parameters ---------- X : ndarray of shape (n_samples, n_features) Data matrix. n_components : int, default=2 Number of dictionary atoms to extract. alpha : float, default=1 Sparsity controlling parameter. n_iter : int, default=100 Number of mini-batch iterations to perform. return_code : bool, default=True Whether to also return the code U or just the dictionary `V`. dict_init : ndarray of shape (n_components, n_features), default=None Initial value for the dictionary for warm restart scenarios. callback : callable, default=None callable that gets invoked every five iterations. batch_size : int, default=3 The number of samples to take in each batch. verbose : bool, default=False To control the verbosity of the procedure. shuffle : bool, default=True Whether to shuffle the data before splitting it in batches. n_jobs : int, default=None Number of parallel jobs to run. ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. ``-1`` means using all processors. See :term:`Glossary <n_jobs>` for more details. method : {'lars', 'cd'}, default='lars' * `'lars'`: uses the least angle regression method to solve the lasso problem (`linear_model.lars_path`); * `'cd'`: uses the coordinate descent method to compute the Lasso solution (`linear_model.Lasso`). Lars will be faster if the estimated components are sparse. iter_offset : int, default=0 Number of previous iterations completed on the dictionary used for initialization. random_state : int, RandomState instance or None, default=None Used for initializing the dictionary when ``dict_init`` is not specified, randomly shuffling the data when ``shuffle`` is set to ``True``, and updating the dictionary. Pass an int for reproducible results across multiple function calls. See :term:`Glossary <random_state>`. return_inner_stats : bool, default=False Return the inner statistics A (dictionary covariance) and B (data approximation). Useful to restart the algorithm in an online setting. If `return_inner_stats` is `True`, `return_code` is ignored. inner_stats : tuple of (A, B) ndarrays, default=None Inner sufficient statistics that are kept by the algorithm. Passing them at initialization is useful in online settings, to avoid losing the history of the evolution. `A` `(n_components, n_components)` is the dictionary covariance matrix. `B` `(n_features, n_components)` is the data approximation matrix. return_n_iter : bool, default=False Whether or not to return the number of iterations. positive_dict : bool, default=False Whether to enforce positivity when finding the dictionary. .. versionadded:: 0.20 positive_code : bool, default=False Whether to enforce positivity when finding the code. .. versionadded:: 0.20 method_max_iter : int, default=1000 Maximum number of iterations to perform when solving the lasso problem. .. versionadded:: 0.22 Returns ------- code : ndarray of shape (n_samples, n_components), The sparse code (only returned if `return_code=True`). dictionary : ndarray of shape (n_components, n_features), The solutions to the dictionary learning problem. n_iter : int Number of iterations run. Returned only if `return_n_iter` is set to `True`. See Also -------- dict_learning DictionaryLearning MiniBatchDictionaryLearning SparsePCA MiniBatchSparsePCA """ if n_components is None: n_components = X.shape[1] if method not in ('lars', 'cd'): raise ValueError('Coding method not supported as a fit algorithm.') _check_positive_coding(method, positive_code) method = 'lasso_' + method t0 = time.time() n_samples, n_features = X.shape # Avoid integer division problems alpha = float(alpha) random_state = check_random_state(random_state) # Init V with SVD of X if dict_init is not None: dictionary = dict_init else: _, S, dictionary = randomized_svd(X, n_components, random_state=random_state) dictionary = S[:, np.newaxis] * dictionary r = len(dictionary) if n_components <= r: dictionary = dictionary[:n_components, :] else: dictionary = np.r_[dictionary, np.zeros((n_components - r, dictionary.shape[1]))] if verbose == 1: print('[dict_learning]', end=' ') if shuffle: X_train = X.copy() random_state.shuffle(X_train) else: X_train = X # Fortran-order dict better suited for the sparse coding which is the # bottleneck of this algorithm. dictionary = check_array(dictionary, order='F', dtype=np.float64, copy=False) dictionary = np.require(dictionary, requirements='W') X_train = check_array(X_train, order='C', dtype=np.float64, copy=False) batches = gen_batches(n_samples, batch_size) batches = itertools.cycle(batches) # The covariance of the dictionary if inner_stats is None: A = np.zeros((n_components, n_components)) # The data approximation B = np.zeros((n_features, n_components)) else: A = inner_stats[0].copy() B = inner_stats[1].copy() # If n_iter is zero, we need to return zero. ii = iter_offset - 1 for ii, batch in zip(range(iter_offset, iter_offset + n_iter), batches): this_X = X_train[batch] dt = (time.time() - t0) if verbose == 1: sys.stdout.write(".") sys.stdout.flush() elif verbose: if verbose > 10 or ii % ceil(100. / verbose) == 0: print("Iteration % 3i (elapsed time: % 3is, % 4.1fmn)" % (ii, dt, dt / 60)) this_code = sparse_encode(this_X, dictionary, algorithm=method, alpha=alpha, n_jobs=n_jobs, check_input=False, positive=positive_code, max_iter=method_max_iter, verbose=verbose) # Update the auxiliary variables if ii < batch_size - 1: theta = float((ii + 1) * batch_size) else: theta = float(batch_size ** 2 + ii + 1 - batch_size) beta = (theta + 1 - batch_size) / (theta + 1) A *= beta A += np.dot(this_code.T, this_code) B *= beta B += np.dot(this_X.T, this_code) # Update dictionary in place _update_dict(dictionary, this_X, this_code, A, B, verbose=verbose, random_state=random_state, positive=positive_dict) # Maybe we need a stopping criteria based on the amount of # modification in the dictionary if callback is not None: callback(locals()) if return_inner_stats: if return_n_iter: return dictionary, (A, B), ii - iter_offset + 1 else: return dictionary, (A, B) if return_code: if verbose > 1: print('Learning code...', end=' ') elif verbose == 1: print('|', end=' ') code = sparse_encode(X, dictionary, algorithm=method, alpha=alpha, n_jobs=n_jobs, check_input=False, positive=positive_code, max_iter=method_max_iter, verbose=verbose) if verbose > 1: dt = (time.time() - t0) print('done (total time: % 3is, % 4.1fmn)' % (dt, dt / 60)) if return_n_iter: return code, dictionary, ii - iter_offset + 1 else: return code, dictionary if return_n_iter: return dictionary, ii - iter_offset + 1 else: return dictionary class _BaseSparseCoding(TransformerMixin): """Base class from SparseCoder and DictionaryLearning algorithms.""" def __init__(self, transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter): self.transform_algorithm = transform_algorithm self.transform_n_nonzero_coefs = transform_n_nonzero_coefs self.transform_alpha = transform_alpha self.transform_max_iter = transform_max_iter self.split_sign = split_sign self.n_jobs = n_jobs self.positive_code = positive_code def _transform(self, X, dictionary): """Private method allowing to accomodate both DictionaryLearning and SparseCoder.""" X = self._validate_data(X, reset=False) # transform_alpha has to be changed in _transform # this is done for consistency with the value of alpha if (hasattr(self, "alpha") and self.alpha != 1. and self.transform_alpha is None): warnings.warn("By default transform_alpha will be equal to" "alpha instead of 1.0 starting from version 1.2", FutureWarning) transform_alpha = 1. # TODO change to self.alpha in 1.2 else: transform_alpha = self.transform_alpha code = sparse_encode( X, dictionary, algorithm=self.transform_algorithm, n_nonzero_coefs=self.transform_n_nonzero_coefs, alpha=transform_alpha, max_iter=self.transform_max_iter, n_jobs=self.n_jobs, positive=self.positive_code) if self.split_sign: # feature vector is split into a positive and negative side n_samples, n_features = code.shape split_code = np.empty((n_samples, 2 * n_features)) split_code[:, :n_features] = np.maximum(code, 0) split_code[:, n_features:] = -np.minimum(code, 0) code = split_code return code def transform(self, X): """Encode the data as a sparse combination of the dictionary atoms. Coding method is determined by the object parameter `transform_algorithm`. Parameters ---------- X : ndarray of shape (n_samples, n_features) Test data to be transformed, must have the same number of features as the data used to train the model. Returns ------- X_new : ndarray of shape (n_samples, n_components) Transformed data. """ check_is_fitted(self) return self._transform(X, self.components_) class SparseCoder(_BaseSparseCoding, BaseEstimator): """Sparse coding Finds a sparse representation of data against a fixed, precomputed dictionary. Each row of the result is the solution to a sparse coding problem. The goal is to find a sparse array `code` such that:: X ~= code * dictionary Read more in the :ref:`User Guide <SparseCoder>`. Parameters ---------- dictionary : ndarray of shape (n_components, n_features) The dictionary atoms used for sparse coding. Lines are assumed to be normalized to unit norm. transform_algorithm : {'lasso_lars', 'lasso_cd', 'lars', 'omp', \ 'threshold'}, default='omp' Algorithm used to transform the data: - `'lars'`: uses the least angle regression method (`linear_model.lars_path`); - `'lasso_lars'`: uses Lars to compute the Lasso solution; - `'lasso_cd'`: uses the coordinate descent method to compute the Lasso solution (linear_model.Lasso). `'lasso_lars'` will be faster if the estimated components are sparse; - `'omp'`: uses orthogonal matching pursuit to estimate the sparse solution; - `'threshold'`: squashes to zero all coefficients less than alpha from the projection ``dictionary * X'``. transform_n_nonzero_coefs : int, default=None Number of nonzero coefficients to target in each column of the solution. This is only used by `algorithm='lars'` and `algorithm='omp'` and is overridden by `alpha` in the `omp` case. If `None`, then `transform_n_nonzero_coefs=int(n_features / 10)`. transform_alpha : float, default=None If `algorithm='lasso_lars'` or `algorithm='lasso_cd'`, `alpha` is the penalty applied to the L1 norm. If `algorithm='threshold'`, `alpha` is the absolute value of the threshold below which coefficients will be squashed to zero. If `algorithm='omp'`, `alpha` is the tolerance parameter: the value of the reconstruction error targeted. In this case, it overrides `n_nonzero_coefs`. If `None`, default to 1. split_sign : bool, default=False Whether to split the sparse feature vector into the concatenation of its negative part and its positive part. This can improve the performance of downstream classifiers. n_jobs : int, default=None Number of parallel jobs to run. ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. ``-1`` means using all processors. See :term:`Glossary <n_jobs>` for more details. positive_code : bool, default=False Whether to enforce positivity when finding the code. .. versionadded:: 0.20 transform_max_iter : int, default=1000 Maximum number of iterations to perform if `algorithm='lasso_cd'` or `lasso_lars`. .. versionadded:: 0.22 Attributes ---------- components_ : ndarray of shape (n_components, n_features) The unchanged dictionary atoms. .. deprecated:: 0.24 This attribute is deprecated in 0.24 and will be removed in 1.1 (renaming of 0.26). Use `dictionary` instead. Examples -------- >>> import numpy as np >>> from sklearn.decomposition import SparseCoder >>> X = np.array([[-1, -1, -1], [0, 0, 3]]) >>> dictionary = np.array( ... [[0, 1, 0], ... [-1, -1, 2], ... [1, 1, 1], ... [0, 1, 1], ... [0, 2, 1]], ... dtype=np.float64 ... ) >>> coder = SparseCoder( ... dictionary=dictionary, transform_algorithm='lasso_lars', ... transform_alpha=1e-10, ... ) >>> coder.transform(X) array([[ 0., 0., -1., 0., 0.], [ 0., 1., 1., 0., 0.]]) See Also -------- DictionaryLearning MiniBatchDictionaryLearning SparsePCA MiniBatchSparsePCA sparse_encode """ _required_parameters = ["dictionary"] def __init__(self, dictionary, *, transform_algorithm='omp', transform_n_nonzero_coefs=None, transform_alpha=None, split_sign=False, n_jobs=None, positive_code=False, transform_max_iter=1000): super().__init__( transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter ) self.dictionary = dictionary def fit(self, X, y=None): """Do nothing and return the estimator unchanged. This method is just there to implement the usual API and hence work in pipelines. Parameters ---------- X : Ignored y : Ignored Returns ------- self : object """ return self @deprecated("The attribute 'components_' is deprecated " # type: ignore "in 0.24 and will be removed in 1.1 (renaming of 0.26). Use " "the 'dictionary' instead.") @property def components_(self): return self.dictionary def transform(self, X, y=None): """Encode the data as a sparse combination of the dictionary atoms. Coding method is determined by the object parameter `transform_algorithm`. Parameters ---------- X : ndarray of shape (n_samples, n_features) Test data to be transformed, must have the same number of features as the data used to train the model. y : Ignored Returns ------- X_new : ndarray of shape (n_samples, n_components) Transformed data. """ return super()._transform(X, self.dictionary) def _more_tags(self): return {"requires_fit": False} @property def n_components_(self): return self.dictionary.shape[0] @property def n_features_in_(self): return self.dictionary.shape[1] class DictionaryLearning(_BaseSparseCoding, BaseEstimator): """Dictionary learning Finds a dictionary (a set of atoms) that can best be used to represent data using a sparse code. Solves the optimization problem:: (U^*,V^*) = argmin 0.5 || X - U V ||_2^2 + alpha * || U ||_1 (U,V) with || V_k ||_2 = 1 for all 0 <= k < n_components Read more in the :ref:`User Guide <DictionaryLearning>`. Parameters ---------- n_components : int, default=n_features Number of dictionary elements to extract. alpha : float, default=1.0 Sparsity controlling parameter. max_iter : int, default=1000 Maximum number of iterations to perform. tol : float, default=1e-8 Tolerance for numerical error. fit_algorithm : {'lars', 'cd'}, default='lars' * `'lars'`: uses the least angle regression method to solve the lasso problem (:func:`~sklearn.linear_model.lars_path`); * `'cd'`: uses the coordinate descent method to compute the Lasso solution (:class:`~sklearn.linear_model.Lasso`). Lars will be faster if the estimated components are sparse. .. versionadded:: 0.17 *cd* coordinate descent method to improve speed. transform_algorithm : {'lasso_lars', 'lasso_cd', 'lars', 'omp', \ 'threshold'}, default='omp' Algorithm used to transform the data: - `'lars'`: uses the least angle regression method (:func:`~sklearn.linear_model.lars_path`); - `'lasso_lars'`: uses Lars to compute the Lasso solution. - `'lasso_cd'`: uses the coordinate descent method to compute the Lasso solution (:class:`~sklearn.linear_model.Lasso`). `'lasso_lars'` will be faster if the estimated components are sparse. - `'omp'`: uses orthogonal matching pursuit to estimate the sparse solution. - `'threshold'`: squashes to zero all coefficients less than alpha from the projection ``dictionary * X'``. .. versionadded:: 0.17 *lasso_cd* coordinate descent method to improve speed. transform_n_nonzero_coefs : int, default=None Number of nonzero coefficients to target in each column of the solution. This is only used by `algorithm='lars'` and `algorithm='omp'`. If `None`, then `transform_n_nonzero_coefs=int(n_features / 10)`. transform_alpha : float, default=None If `algorithm='lasso_lars'` or `algorithm='lasso_cd'`, `alpha` is the penalty applied to the L1 norm. If `algorithm='threshold'`, `alpha` is the absolute value of the threshold below which coefficients will be squashed to zero. If `None`, defaults to `alpha`. n_jobs : int or None, default=None Number of parallel jobs to run. ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. ``-1`` means using all processors. See :term:`Glossary <n_jobs>` for more details. code_init : ndarray of shape (n_samples, n_components), default=None Initial value for the code, for warm restart. Only used if `code_init` and `dict_init` are not None. dict_init : ndarray of shape (n_components, n_features), default=None Initial values for the dictionary, for warm restart. Only used if `code_init` and `dict_init` are not None. verbose : bool, default=False To control the verbosity of the procedure. split_sign : bool, default=False Whether to split the sparse feature vector into the concatenation of its negative part and its positive part. This can improve the performance of downstream classifiers. random_state : int, RandomState instance or None, default=None Used for initializing the dictionary when ``dict_init`` is not specified, randomly shuffling the data when ``shuffle`` is set to ``True``, and updating the dictionary. Pass an int for reproducible results across multiple function calls. See :term:`Glossary <random_state>`. positive_code : bool, default=False Whether to enforce positivity when finding the code. .. versionadded:: 0.20 positive_dict : bool, default=False Whether to enforce positivity when finding the dictionary .. versionadded:: 0.20 transform_max_iter : int, default=1000 Maximum number of iterations to perform if `algorithm='lasso_cd'` or `'lasso_lars'`. .. versionadded:: 0.22 Attributes ---------- components_ : ndarray of shape (n_components, n_features) dictionary atoms extracted from the data error_ : array vector of errors at each iteration n_iter_ : int Number of iterations run. Examples -------- >>> import numpy as np >>> from sklearn.datasets import make_sparse_coded_signal >>> from sklearn.decomposition import DictionaryLearning >>> X, dictionary, code = make_sparse_coded_signal( ... n_samples=100, n_components=15, n_features=20, n_nonzero_coefs=10, ... random_state=42, ... ) >>> dict_learner = DictionaryLearning( ... n_components=15, transform_algorithm='lasso_lars', random_state=42, ... ) >>> X_transformed = dict_learner.fit_transform(X) We can check the level of sparsity of `X_transformed`: >>> np.mean(X_transformed == 0) 0.87... We can compare the average squared euclidean norm of the reconstruction error of the sparse coded signal relative to the squared euclidean norm of the original signal: >>> X_hat = X_transformed @ dict_learner.components_ >>> np.mean(np.sum((X_hat - X) ** 2, axis=1) / np.sum(X ** 2, axis=1)) 0.08... Notes ----- **References:** J. Mairal, F. Bach, J. Ponce, G. Sapiro, 2009: Online dictionary learning for sparse coding (https://www.di.ens.fr/sierra/pdfs/icml09.pdf) See Also -------- SparseCoder MiniBatchDictionaryLearning SparsePCA MiniBatchSparsePCA """ def __init__(self, n_components=None, *, alpha=1, max_iter=1000, tol=1e-8, fit_algorithm='lars', transform_algorithm='omp', transform_n_nonzero_coefs=None, transform_alpha=None, n_jobs=None, code_init=None, dict_init=None, verbose=False, split_sign=False, random_state=None, positive_code=False, positive_dict=False, transform_max_iter=1000): super().__init__( transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter ) self.n_components = n_components self.alpha = alpha self.max_iter = max_iter self.tol = tol self.fit_algorithm = fit_algorithm self.code_init = code_init self.dict_init = dict_init self.verbose = verbose self.random_state = random_state self.positive_dict = positive_dict def fit(self, X, y=None): """Fit the model from data in X. Parameters ---------- X : array-like of shape (n_samples, n_features) Training vector, where `n_samples` in the number of samples and `n_features` is the number of features. y : Ignored Returns ------- self : object Returns the object itself. """ random_state = check_random_state(self.random_state) X = self._validate_data(X) if self.n_components is None: n_components = X.shape[1] else: n_components = self.n_components V, U, E, self.n_iter_ = dict_learning( X, n_components, alpha=self.alpha, tol=self.tol, max_iter=self.max_iter, method=self.fit_algorithm, method_max_iter=self.transform_max_iter, n_jobs=self.n_jobs, code_init=self.code_init, dict_init=self.dict_init, verbose=self.verbose, random_state=random_state, return_n_iter=True, positive_dict=self.positive_dict, positive_code=self.positive_code) self.components_ = U self.error_ = E return self class MiniBatchDictionaryLearning(_BaseSparseCoding, BaseEstimator): """Mini-batch dictionary learning Finds a dictionary (a set of atoms) that can best be used to represent data using a sparse code. Solves the optimization problem:: (U^*,V^*) = argmin 0.5 || X - U V ||_2^2 + alpha * || U ||_1 (U,V) with || V_k ||_2 = 1 for all 0 <= k < n_components Read more in the :ref:`User Guide <DictionaryLearning>`. Parameters ---------- n_components : int, default=None Number of dictionary elements to extract. alpha : float, default=1 Sparsity controlling parameter. n_iter : int, default=1000 Total number of iterations to perform. fit_algorithm : {'lars', 'cd'}, default='lars' The algorithm used: - `'lars'`: uses the least angle regression method to solve the lasso problem (`linear_model.lars_path`) - `'cd'`: uses the coordinate descent method to compute the Lasso solution (`linear_model.Lasso`). Lars will be faster if the estimated components are sparse. n_jobs : int, default=None Number of parallel jobs to run. ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. ``-1`` means using all processors. See :term:`Glossary <n_jobs>` for more details. batch_size : int, default=3 Number of samples in each mini-batch. shuffle : bool, default=True Whether to shuffle the samples before forming batches. dict_init : ndarray of shape (n_components, n_features), default=None initial value of the dictionary for warm restart scenarios transform_algorithm : {'lasso_lars', 'lasso_cd', 'lars', 'omp', \ 'threshold'}, default='omp' Algorithm used to transform the data: - `'lars'`: uses the least angle regression method (`linear_model.lars_path`); - `'lasso_lars'`: uses Lars to compute the Lasso solution. - `'lasso_cd'`: uses the coordinate descent method to compute the Lasso solution (`linear_model.Lasso`). `'lasso_lars'` will be faster if the estimated components are sparse. - `'omp'`: uses orthogonal matching pursuit to estimate the sparse solution. - `'threshold'`: squashes to zero all coefficients less than alpha from the projection ``dictionary * X'``. transform_n_nonzero_coefs : int, default=None Number of nonzero coefficients to target in each column of the solution. This is only used by `algorithm='lars'` and `algorithm='omp'`. If `None`, then `transform_n_nonzero_coefs=int(n_features / 10)`. transform_alpha : float, default=None If `algorithm='lasso_lars'` or `algorithm='lasso_cd'`, `alpha` is the penalty applied to the L1 norm. If `algorithm='threshold'`, `alpha` is the absolute value of the threshold below which coefficients will be squashed to zero. If `None`, defaults to `alpha`. verbose : bool, default=False To control the verbosity of the procedure. split_sign : bool, default=False Whether to split the sparse feature vector into the concatenation of its negative part and its positive part. This can improve the performance of downstream classifiers. random_state : int, RandomState instance or None, default=None Used for initializing the dictionary when ``dict_init`` is not specified, randomly shuffling the data when ``shuffle`` is set to ``True``, and updating the dictionary. Pass an int for reproducible results across multiple function calls. See :term:`Glossary <random_state>`. positive_code : bool, default=False Whether to enforce positivity when finding the code. .. versionadded:: 0.20 positive_dict : bool, default=False Whether to enforce positivity when finding the dictionary. .. versionadded:: 0.20 transform_max_iter : int, default=1000 Maximum number of iterations to perform if `algorithm='lasso_cd'` or `'lasso_lars'`. .. versionadded:: 0.22 Attributes ---------- components_ : ndarray of shape (n_components, n_features) Components extracted from the data. inner_stats_ : tuple of (A, B) ndarrays Internal sufficient statistics that are kept by the algorithm. Keeping them is useful in online settings, to avoid losing the history of the evolution, but they shouldn't have any use for the end user. `A` `(n_components, n_components)` is the dictionary covariance matrix. `B` `(n_features, n_components)` is the data approximation matrix. n_iter_ : int Number of iterations run. iter_offset_ : int The number of iteration on data batches that has been performed before. random_state_ : RandomState instance RandomState instance that is generated either from a seed, the random number generattor or by `np.random`. Examples -------- >>> import numpy as np >>> from sklearn.datasets import make_sparse_coded_signal >>> from sklearn.decomposition import MiniBatchDictionaryLearning >>> X, dictionary, code = make_sparse_coded_signal( ... n_samples=100, n_components=15, n_features=20, n_nonzero_coefs=10, ... random_state=42) >>> dict_learner = MiniBatchDictionaryLearning( ... n_components=15, transform_algorithm='lasso_lars', random_state=42, ... ) >>> X_transformed = dict_learner.fit_transform(X) We can check the level of sparsity of `X_transformed`: >>> np.mean(X_transformed == 0) 0.86... We can compare the average squared euclidean norm of the reconstruction error of the sparse coded signal relative to the squared euclidean norm of the original signal: >>> X_hat = X_transformed @ dict_learner.components_ >>> np.mean(np.sum((X_hat - X) ** 2, axis=1) / np.sum(X ** 2, axis=1)) 0.07... Notes ----- **References:** J. Mairal, F. Bach, J. Ponce, G. Sapiro, 2009: Online dictionary learning for sparse coding (https://www.di.ens.fr/sierra/pdfs/icml09.pdf) See Also -------- SparseCoder DictionaryLearning SparsePCA MiniBatchSparsePCA """ def __init__(self, n_components=None, *, alpha=1, n_iter=1000, fit_algorithm='lars', n_jobs=None, batch_size=3, shuffle=True, dict_init=None, transform_algorithm='omp', transform_n_nonzero_coefs=None, transform_alpha=None, verbose=False, split_sign=False, random_state=None, positive_code=False, positive_dict=False, transform_max_iter=1000): super().__init__( transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter ) self.n_components = n_components self.alpha = alpha self.n_iter = n_iter self.fit_algorithm = fit_algorithm self.dict_init = dict_init self.verbose = verbose self.shuffle = shuffle self.batch_size = batch_size self.split_sign = split_sign self.random_state = random_state self.positive_dict = positive_dict def fit(self, X, y=None): """Fit the model from data in X. Parameters ---------- X : array-like of shape (n_samples, n_features) Training vector, where n_samples in the number of samples and n_features is the number of features. y : Ignored Returns ------- self : object Returns the instance itself. """ random_state = check_random_state(self.random_state) X = self._validate_data(X) U, (A, B), self.n_iter_ = dict_learning_online( X, self.n_components, alpha=self.alpha, n_iter=self.n_iter, return_code=False, method=self.fit_algorithm, method_max_iter=self.transform_max_iter, n_jobs=self.n_jobs, dict_init=self.dict_init, batch_size=self.batch_size, shuffle=self.shuffle, verbose=self.verbose, random_state=random_state, return_inner_stats=True, return_n_iter=True, positive_dict=self.positive_dict, positive_code=self.positive_code) self.components_ = U # Keep track of the state of the algorithm to be able to do # some online fitting (partial_fit) self.inner_stats_ = (A, B) self.iter_offset_ = self.n_iter self.random_state_ = random_state return self def partial_fit(self, X, y=None, iter_offset=None): """Updates the model using the data in X as a mini-batch. Parameters ---------- X : array-like of shape (n_samples, n_features) Training vector, where n_samples in the number of samples and n_features is the number of features. y : Ignored iter_offset : int, default=None The number of iteration on data batches that has been performed before this call to partial_fit. This is optional: if no number is passed, the memory of the object is used. Returns ------- self : object Returns the instance itself. """ if not hasattr(self, 'random_state_'): self.random_state_ = check_random_state(self.random_state) if hasattr(self, 'components_'): dict_init = self.components_ else: dict_init = self.dict_init inner_stats = getattr(self, 'inner_stats_', None) if iter_offset is None: iter_offset = getattr(self, 'iter_offset_', 0) X = self._validate_data(X, reset=(iter_offset == 0)) U, (A, B) = dict_learning_online( X, self.n_components, alpha=self.alpha, n_iter=1, method=self.fit_algorithm, method_max_iter=self.transform_max_iter, n_jobs=self.n_jobs, dict_init=dict_init, batch_size=len(X), shuffle=False, verbose=self.verbose, return_code=False, iter_offset=iter_offset, random_state=self.random_state_, return_inner_stats=True, inner_stats=inner_stats, positive_dict=self.positive_dict, positive_code=self.positive_code) self.components_ = U # Keep track of the state of the algorithm to be able to do # some online fitting (partial_fit) self.inner_stats_ = (A, B) self.iter_offset_ = iter_offset + 1 return self
36.228155
79
0.62222
import time import sys import itertools import warnings from math import ceil import numpy as np from scipy import linalg from joblib import Parallel, effective_n_jobs from ..base import BaseEstimator, TransformerMixin from ..utils import deprecated from ..utils import (check_array, check_random_state, gen_even_slices, gen_batches) from ..utils.extmath import randomized_svd, row_norms, svd_flip from ..utils.validation import check_is_fitted from ..utils.fixes import delayed from ..linear_model import Lasso, orthogonal_mp_gram, LassoLars, Lars def _check_positive_coding(method, positive): if positive and method in ["omp", "lars"]: raise ValueError( "Positive constraint not supported for '{}' " "coding method.".format(method) ) def _sparse_encode(X, dictionary, gram, cov=None, algorithm='lasso_lars', regularization=None, copy_cov=True, init=None, max_iter=1000, check_input=True, verbose=0, positive=False): if X.ndim == 1: X = X[:, np.newaxis] n_samples, n_features = X.shape n_components = dictionary.shape[0] if dictionary.shape[1] != X.shape[1]: raise ValueError("Dictionary and X have different numbers of features:" "dictionary.shape: {} X.shape{}".format( dictionary.shape, X.shape)) if cov is None and algorithm != 'lasso_cd': copy_cov = False cov = np.dot(dictionary, X.T) _check_positive_coding(algorithm, positive) if algorithm == 'lasso_lars': alpha = float(regularization) / n_features try: err_mgt = np.seterr(all='ignore') lasso_lars = LassoLars(alpha=alpha, fit_intercept=False, verbose=verbose, normalize=False, precompute=gram, fit_path=False, positive=positive, max_iter=max_iter) lasso_lars.fit(dictionary.T, X.T, Xy=cov) new_code = lasso_lars.coef_ finally: np.seterr(**err_mgt) elif algorithm == 'lasso_cd': alpha = float(regularization) / n_features clf = Lasso(alpha=alpha, fit_intercept=False, normalize=False, precompute=gram, max_iter=max_iter, warm_start=True, positive=positive) if init is not None: clf.coef_ = init clf.fit(dictionary.T, X.T, check_input=check_input) new_code = clf.coef_ elif algorithm == 'lars': try: err_mgt = np.seterr(all='ignore') lars = Lars(fit_intercept=False, verbose=verbose, normalize=False, precompute=gram, n_nonzero_coefs=int(regularization), fit_path=False) lars.fit(dictionary.T, X.T, Xy=cov) new_code = lars.coef_ finally: np.seterr(**err_mgt) elif algorithm == 'threshold': new_code = ((np.sign(cov) * np.maximum(np.abs(cov) - regularization, 0)).T) if positive: np.clip(new_code, 0, None, out=new_code) elif algorithm == 'omp': new_code = orthogonal_mp_gram( Gram=gram, Xy=cov, n_nonzero_coefs=int(regularization), tol=None, norms_squared=row_norms(X, squared=True), copy_Xy=copy_cov).T else: raise ValueError('Sparse coding method must be "lasso_lars" ' '"lasso_cd", "lasso", "threshold" or "omp", got %s.' % algorithm) if new_code.ndim != 2: return new_code.reshape(n_samples, n_components) return new_code def sparse_encode(X, dictionary, *, gram=None, cov=None, algorithm='lasso_lars', n_nonzero_coefs=None, alpha=None, copy_cov=True, init=None, max_iter=1000, n_jobs=None, check_input=True, verbose=0, positive=False): if check_input: if algorithm == 'lasso_cd': dictionary = check_array(dictionary, order='C', dtype='float64') X = check_array(X, order='C', dtype='float64') else: dictionary = check_array(dictionary) X = check_array(X) n_samples, n_features = X.shape n_components = dictionary.shape[0] if gram is None and algorithm != 'threshold': gram = np.dot(dictionary, dictionary.T) if cov is None and algorithm != 'lasso_cd': copy_cov = False cov = np.dot(dictionary, X.T) if algorithm in ('lars', 'omp'): regularization = n_nonzero_coefs if regularization is None: regularization = min(max(n_features / 10, 1), n_components) else: regularization = alpha if regularization is None: regularization = 1. if effective_n_jobs(n_jobs) == 1 or algorithm == 'threshold': code = _sparse_encode(X, dictionary, gram, cov=cov, algorithm=algorithm, regularization=regularization, copy_cov=copy_cov, init=init, max_iter=max_iter, check_input=False, verbose=verbose, positive=positive) return code code = np.empty((n_samples, n_components)) slices = list(gen_even_slices(n_samples, effective_n_jobs(n_jobs))) code_views = Parallel(n_jobs=n_jobs, verbose=verbose)( delayed(_sparse_encode)( X[this_slice], dictionary, gram, cov[:, this_slice] if cov is not None else None, algorithm, regularization=regularization, copy_cov=copy_cov, init=init[this_slice] if init is not None else None, max_iter=max_iter, check_input=False, verbose=verbose, positive=positive) for this_slice in slices) for this_slice, this_view in zip(slices, code_views): code[this_slice] = this_view return code def _update_dict(dictionary, Y, code, A=None, B=None, verbose=False, random_state=None, positive=False): n_samples, n_components = code.shape random_state = check_random_state(random_state) if A is None: A = code.T @ code if B is None: B = Y.T @ code n_unused = 0 for k in range(n_components): if A[k, k] > 1e-6: dictionary[k] += (B[:, k] - A[k] @ dictionary) / A[k, k] else: newd = Y[random_state.choice(n_samples)] noise_level = 0.01 * (newd.std() or 1) noise = random_state.normal(0, noise_level, size=len(newd)) dictionary[k] = newd + noise code[:, k] = 0 n_unused += 1 if positive: np.clip(dictionary[k], 0, None, out=dictionary[k]) dictionary[k] /= linalg.norm(dictionary[k]) if verbose and n_unused > 0: print(f"{n_unused} unused atoms resampled.") def dict_learning(X, n_components, *, alpha, max_iter=100, tol=1e-8, method='lars', n_jobs=None, dict_init=None, code_init=None, callback=None, verbose=False, random_state=None, return_n_iter=False, positive_dict=False, positive_code=False, method_max_iter=1000): if method not in ('lars', 'cd'): raise ValueError('Coding method %r not supported as a fit algorithm.' % method) _check_positive_coding(method, positive_code) method = 'lasso_' + method t0 = time.time() alpha = float(alpha) random_state = check_random_state(random_state) if code_init is not None and dict_init is not None: code = np.array(code_init, order='F') dictionary = dict_init else: code, S, dictionary = linalg.svd(X, full_matrices=False) # flip the initial code's sign to enforce deterministic output code, dictionary = svd_flip(code, dictionary) dictionary = S[:, np.newaxis] * dictionary r = len(dictionary) if n_components <= r: code = code[:, :n_components] dictionary = dictionary[:n_components, :] else: code = np.c_[code, np.zeros((len(code), n_components - r))] dictionary = np.r_[dictionary, np.zeros((n_components - r, dictionary.shape[1]))] dictionary = np.asfortranarray(dictionary) errors = [] current_cost = np.nan if verbose == 1: print('[dict_learning]', end=' ') ii = -1 for ii in range(max_iter): dt = (time.time() - t0) if verbose == 1: sys.stdout.write(".") sys.stdout.flush() elif verbose: print("Iteration % 3i " "(elapsed time: % 3is, % 4.1fmn, current cost % 7.3f)" % (ii, dt, dt / 60, current_cost)) code = sparse_encode(X, dictionary, algorithm=method, alpha=alpha, init=code, n_jobs=n_jobs, positive=positive_code, max_iter=method_max_iter, verbose=verbose) _update_dict(dictionary, X, code, verbose=verbose, random_state=random_state, positive=positive_dict) current_cost = (0.5 * np.sum((X - code @ dictionary)**2) + alpha * np.sum(np.abs(code))) errors.append(current_cost) if ii > 0: dE = errors[-2] - errors[-1] if dE < tol * errors[-1]: if verbose == 1: print("") elif verbose: print("--- Convergence reached after %d iterations" % ii) break if ii % 5 == 0 and callback is not None: callback(locals()) if return_n_iter: return code, dictionary, errors, ii + 1 else: return code, dictionary, errors def dict_learning_online(X, n_components=2, *, alpha=1, n_iter=100, return_code=True, dict_init=None, callback=None, batch_size=3, verbose=False, shuffle=True, n_jobs=None, method='lars', iter_offset=0, random_state=None, return_inner_stats=False, inner_stats=None, return_n_iter=False, positive_dict=False, positive_code=False, method_max_iter=1000): if n_components is None: n_components = X.shape[1] if method not in ('lars', 'cd'): raise ValueError('Coding method not supported as a fit algorithm.') _check_positive_coding(method, positive_code) method = 'lasso_' + method t0 = time.time() n_samples, n_features = X.shape alpha = float(alpha) random_state = check_random_state(random_state) if dict_init is not None: dictionary = dict_init else: _, S, dictionary = randomized_svd(X, n_components, random_state=random_state) dictionary = S[:, np.newaxis] * dictionary r = len(dictionary) if n_components <= r: dictionary = dictionary[:n_components, :] else: dictionary = np.r_[dictionary, np.zeros((n_components - r, dictionary.shape[1]))] if verbose == 1: print('[dict_learning]', end=' ') if shuffle: X_train = X.copy() random_state.shuffle(X_train) else: X_train = X dictionary = check_array(dictionary, order='F', dtype=np.float64, copy=False) dictionary = np.require(dictionary, requirements='W') X_train = check_array(X_train, order='C', dtype=np.float64, copy=False) batches = gen_batches(n_samples, batch_size) batches = itertools.cycle(batches) if inner_stats is None: A = np.zeros((n_components, n_components)) B = np.zeros((n_features, n_components)) else: A = inner_stats[0].copy() B = inner_stats[1].copy() ii = iter_offset - 1 for ii, batch in zip(range(iter_offset, iter_offset + n_iter), batches): this_X = X_train[batch] dt = (time.time() - t0) if verbose == 1: sys.stdout.write(".") sys.stdout.flush() elif verbose: if verbose > 10 or ii % ceil(100. / verbose) == 0: print("Iteration % 3i (elapsed time: % 3is, % 4.1fmn)" % (ii, dt, dt / 60)) this_code = sparse_encode(this_X, dictionary, algorithm=method, alpha=alpha, n_jobs=n_jobs, check_input=False, positive=positive_code, max_iter=method_max_iter, verbose=verbose) if ii < batch_size - 1: theta = float((ii + 1) * batch_size) else: theta = float(batch_size ** 2 + ii + 1 - batch_size) beta = (theta + 1 - batch_size) / (theta + 1) A *= beta A += np.dot(this_code.T, this_code) B *= beta B += np.dot(this_X.T, this_code) _update_dict(dictionary, this_X, this_code, A, B, verbose=verbose, random_state=random_state, positive=positive_dict) if callback is not None: callback(locals()) if return_inner_stats: if return_n_iter: return dictionary, (A, B), ii - iter_offset + 1 else: return dictionary, (A, B) if return_code: if verbose > 1: print('Learning code...', end=' ') elif verbose == 1: print('|', end=' ') code = sparse_encode(X, dictionary, algorithm=method, alpha=alpha, n_jobs=n_jobs, check_input=False, positive=positive_code, max_iter=method_max_iter, verbose=verbose) if verbose > 1: dt = (time.time() - t0) print('done (total time: % 3is, % 4.1fmn)' % (dt, dt / 60)) if return_n_iter: return code, dictionary, ii - iter_offset + 1 else: return code, dictionary if return_n_iter: return dictionary, ii - iter_offset + 1 else: return dictionary class _BaseSparseCoding(TransformerMixin): def __init__(self, transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter): self.transform_algorithm = transform_algorithm self.transform_n_nonzero_coefs = transform_n_nonzero_coefs self.transform_alpha = transform_alpha self.transform_max_iter = transform_max_iter self.split_sign = split_sign self.n_jobs = n_jobs self.positive_code = positive_code def _transform(self, X, dictionary): X = self._validate_data(X, reset=False) if (hasattr(self, "alpha") and self.alpha != 1. and self.transform_alpha is None): warnings.warn("By default transform_alpha will be equal to" "alpha instead of 1.0 starting from version 1.2", FutureWarning) transform_alpha = 1. else: transform_alpha = self.transform_alpha code = sparse_encode( X, dictionary, algorithm=self.transform_algorithm, n_nonzero_coefs=self.transform_n_nonzero_coefs, alpha=transform_alpha, max_iter=self.transform_max_iter, n_jobs=self.n_jobs, positive=self.positive_code) if self.split_sign: n_samples, n_features = code.shape split_code = np.empty((n_samples, 2 * n_features)) split_code[:, :n_features] = np.maximum(code, 0) split_code[:, n_features:] = -np.minimum(code, 0) code = split_code return code def transform(self, X): check_is_fitted(self) return self._transform(X, self.components_) class SparseCoder(_BaseSparseCoding, BaseEstimator): _required_parameters = ["dictionary"] def __init__(self, dictionary, *, transform_algorithm='omp', transform_n_nonzero_coefs=None, transform_alpha=None, split_sign=False, n_jobs=None, positive_code=False, transform_max_iter=1000): super().__init__( transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter ) self.dictionary = dictionary def fit(self, X, y=None): return self @deprecated("The attribute 'components_' is deprecated " "in 0.24 and will be removed in 1.1 (renaming of 0.26). Use " "the 'dictionary' instead.") @property def components_(self): return self.dictionary def transform(self, X, y=None): return super()._transform(X, self.dictionary) def _more_tags(self): return {"requires_fit": False} @property def n_components_(self): return self.dictionary.shape[0] @property def n_features_in_(self): return self.dictionary.shape[1] class DictionaryLearning(_BaseSparseCoding, BaseEstimator): def __init__(self, n_components=None, *, alpha=1, max_iter=1000, tol=1e-8, fit_algorithm='lars', transform_algorithm='omp', transform_n_nonzero_coefs=None, transform_alpha=None, n_jobs=None, code_init=None, dict_init=None, verbose=False, split_sign=False, random_state=None, positive_code=False, positive_dict=False, transform_max_iter=1000): super().__init__( transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter ) self.n_components = n_components self.alpha = alpha self.max_iter = max_iter self.tol = tol self.fit_algorithm = fit_algorithm self.code_init = code_init self.dict_init = dict_init self.verbose = verbose self.random_state = random_state self.positive_dict = positive_dict def fit(self, X, y=None): random_state = check_random_state(self.random_state) X = self._validate_data(X) if self.n_components is None: n_components = X.shape[1] else: n_components = self.n_components V, U, E, self.n_iter_ = dict_learning( X, n_components, alpha=self.alpha, tol=self.tol, max_iter=self.max_iter, method=self.fit_algorithm, method_max_iter=self.transform_max_iter, n_jobs=self.n_jobs, code_init=self.code_init, dict_init=self.dict_init, verbose=self.verbose, random_state=random_state, return_n_iter=True, positive_dict=self.positive_dict, positive_code=self.positive_code) self.components_ = U self.error_ = E return self class MiniBatchDictionaryLearning(_BaseSparseCoding, BaseEstimator): def __init__(self, n_components=None, *, alpha=1, n_iter=1000, fit_algorithm='lars', n_jobs=None, batch_size=3, shuffle=True, dict_init=None, transform_algorithm='omp', transform_n_nonzero_coefs=None, transform_alpha=None, verbose=False, split_sign=False, random_state=None, positive_code=False, positive_dict=False, transform_max_iter=1000): super().__init__( transform_algorithm, transform_n_nonzero_coefs, transform_alpha, split_sign, n_jobs, positive_code, transform_max_iter ) self.n_components = n_components self.alpha = alpha self.n_iter = n_iter self.fit_algorithm = fit_algorithm self.dict_init = dict_init self.verbose = verbose self.shuffle = shuffle self.batch_size = batch_size self.split_sign = split_sign self.random_state = random_state self.positive_dict = positive_dict def fit(self, X, y=None): random_state = check_random_state(self.random_state) X = self._validate_data(X) U, (A, B), self.n_iter_ = dict_learning_online( X, self.n_components, alpha=self.alpha, n_iter=self.n_iter, return_code=False, method=self.fit_algorithm, method_max_iter=self.transform_max_iter, n_jobs=self.n_jobs, dict_init=self.dict_init, batch_size=self.batch_size, shuffle=self.shuffle, verbose=self.verbose, random_state=random_state, return_inner_stats=True, return_n_iter=True, positive_dict=self.positive_dict, positive_code=self.positive_code) self.components_ = U self.inner_stats_ = (A, B) self.iter_offset_ = self.n_iter self.random_state_ = random_state return self def partial_fit(self, X, y=None, iter_offset=None): if not hasattr(self, 'random_state_'): self.random_state_ = check_random_state(self.random_state) if hasattr(self, 'components_'): dict_init = self.components_ else: dict_init = self.dict_init inner_stats = getattr(self, 'inner_stats_', None) if iter_offset is None: iter_offset = getattr(self, 'iter_offset_', 0) X = self._validate_data(X, reset=(iter_offset == 0)) U, (A, B) = dict_learning_online( X, self.n_components, alpha=self.alpha, n_iter=1, method=self.fit_algorithm, method_max_iter=self.transform_max_iter, n_jobs=self.n_jobs, dict_init=dict_init, batch_size=len(X), shuffle=False, verbose=self.verbose, return_code=False, iter_offset=iter_offset, random_state=self.random_state_, return_inner_stats=True, inner_stats=inner_stats, positive_dict=self.positive_dict, positive_code=self.positive_code) self.components_ = U self.inner_stats_ = (A, B) self.iter_offset_ = iter_offset + 1 return self
true
true
1c485451846b20e51102909630e63adf8b06d286
366
py
Python
UchuujinPatcher/patch_eboot.py
colebob9/UchuujinPatcher
1f5880240e7b3da329d4c8334fc23df92eece402
[ "MIT" ]
null
null
null
UchuujinPatcher/patch_eboot.py
colebob9/UchuujinPatcher
1f5880240e7b3da329d4c8334fc23df92eece402
[ "MIT" ]
null
null
null
UchuujinPatcher/patch_eboot.py
colebob9/UchuujinPatcher
1f5880240e7b3da329d4c8334fc23df92eece402
[ "MIT" ]
null
null
null
# https://pypi.org/project/bsdiff4/ # Use bsdiff and eboot patch in main repo # Needs wheel and VC++ 2015 v140 toolset (choco install vcbuildtools) # on Windows, look into Linux / Docker import bsdiff4 bsdiff4.file_patch("isofiles/EBOOT.BIN", "isofiles/EBOOT_patched.BIN", "main_repo/EBOOT.BIN.patch" )
26.142857
69
0.63388
import bsdiff4 bsdiff4.file_patch("isofiles/EBOOT.BIN", "isofiles/EBOOT_patched.BIN", "main_repo/EBOOT.BIN.patch" )
true
true
1c485455fea34089e7998b97dab5da52c8ae328a
493
py
Python
queries/q5.py
csruiliu/tpch-pyspark
ec707ddd8a5e917b08e0ee1ce320b826fa6aa977
[ "MIT" ]
null
null
null
queries/q5.py
csruiliu/tpch-pyspark
ec707ddd8a5e917b08e0ee1ce320b826fa6aa977
[ "MIT" ]
null
null
null
queries/q5.py
csruiliu/tpch-pyspark
ec707ddd8a5e917b08e0ee1ce320b826fa6aa977
[ "MIT" ]
null
null
null
query = """ SELECT N_NAME, sum(L_EXTENDEDPRICE * (1 - L_DISCOUNT)) AS REVENUE FROM customer, orders, lineitem, supplier, nation, region WHERE C_CUSTKEY = O_CUSTKEY AND L_ORDERKEY = O_ORDERKEY AND L_SUPPKEY = S_SUPPKEY AND C_NATIONKEY = S_NATIONKEY AND S_NATIONKEY = N_NATIONKEY AND N_REGIONKEY = R_REGIONKEY AND R_NAME = 'ASIA' AND O_ORDERDATE >= '1994-01-01' AND O_ORDERDATE < '1995-01-01' GROUP BY N_NAME ORDER BY REVENUE desc """
22.409091
59
0.6714
query = """ SELECT N_NAME, sum(L_EXTENDEDPRICE * (1 - L_DISCOUNT)) AS REVENUE FROM customer, orders, lineitem, supplier, nation, region WHERE C_CUSTKEY = O_CUSTKEY AND L_ORDERKEY = O_ORDERKEY AND L_SUPPKEY = S_SUPPKEY AND C_NATIONKEY = S_NATIONKEY AND S_NATIONKEY = N_NATIONKEY AND N_REGIONKEY = R_REGIONKEY AND R_NAME = 'ASIA' AND O_ORDERDATE >= '1994-01-01' AND O_ORDERDATE < '1995-01-01' GROUP BY N_NAME ORDER BY REVENUE desc """
true
true
1c48545cf2834159a424d70b7811f00db3a47e6c
3,089
py
Python
bassl/pretrain/utils/metric.py
kakaobrain/bassl
551fe94343debf60a64c787be6752284153a0f7a
[ "Apache-2.0" ]
55
2022-01-17T02:18:40.000Z
2022-03-25T08:24:28.000Z
bassl/pretrain/utils/metric.py
kakaobrain/bassl
551fe94343debf60a64c787be6752284153a0f7a
[ "Apache-2.0" ]
5
2022-01-18T01:59:49.000Z
2022-03-24T00:20:35.000Z
bassl/pretrain/utils/metric.py
kakaobrain/bassl
551fe94343debf60a64c787be6752284153a0f7a
[ "Apache-2.0" ]
1
2022-01-23T10:50:15.000Z
2022-01-23T10:50:15.000Z
""" - kNN Precision """ from collections import defaultdict import torch import torchmetrics class KnnPrecisionMetric(torchmetrics.Metric): def __init__(self, top_k_list): super().__init__(compute_on_step=False, dist_sync_on_step=True) self.add_state("feat_data", default=[], dist_reduce_fx=None) self.add_state("vids_data", default=[], dist_reduce_fx=None) self.add_state("scene_data", default=[], dist_reduce_fx=None) self.top_k_list = set(top_k_list) self.max_k = max(self.top_k_list) def update(self, vid, invideo_scene_id, feat): assert isinstance(invideo_scene_id, torch.Tensor) assert isinstance(vid, torch.Tensor) assert isinstance(feat, torch.Tensor) self.feat_data.append(feat) self.vids_data.append(vid) self.scene_data.append(invideo_scene_id) def compute(self) -> torch.Tensor: score = defaultdict(dict) pool_feats = defaultdict(list) pool_invideo_scene_id = defaultdict(list) pool_gts = defaultdict(dict) num_data = 0 for vid, invideo_scene_id, gathered_feat in zip( self.vids_data, self.scene_data, self.feat_data ): vid = vid.item() invideo_scene_id = invideo_scene_id.item() if invideo_scene_id not in pool_gts[vid]: pool_gts[vid][invideo_scene_id] = set() pool_gts[vid][invideo_scene_id].add(len(pool_feats[vid])) pool_invideo_scene_id[vid].append(invideo_scene_id) pool_feats[vid].append(gathered_feat) num_data += 1 for top_k in self.top_k_list: score[top_k] = {"correct": 0, "total": 0} for vid, gt in pool_gts.items(): X = torch.stack(pool_feats[vid]) sim = torch.matmul(X, X.t()) sim = sim - 999 * torch.eye(sim.shape[0]).type_as(sim) # exclude self indices = torch.argsort(sim, descending=True) assert indices.shape[1] >= self.max_k, f"{indices.shape[1]} >= {self.max_k}" indices = indices[:, : self.max_k] for j in range(indices.shape[0]): _cache = {"correct": 0, "total": 0} _query_scene_id = pool_invideo_scene_id[vid][j] for k in range(self.max_k): if _query_scene_id in gt: if indices[j][k].item() in gt[_query_scene_id]: _cache["correct"] += 1 _cache["total"] += 1 if k + 1 in self.top_k_list and len(gt[_query_scene_id]) > k: score[k + 1]["correct"] += _cache["correct"] score[k + 1]["total"] += _cache["total"] for top_k in self.top_k_list: assert score[top_k]["total"] > 0 score[top_k]["precision"] = ( 100.0 * score[top_k]["correct"] / score[top_k]["total"] ) del X, sim, indices, pool_feats, pool_invideo_scene_id, pool_gts torch.cuda.empty_cache() return score
39.602564
88
0.583684
from collections import defaultdict import torch import torchmetrics class KnnPrecisionMetric(torchmetrics.Metric): def __init__(self, top_k_list): super().__init__(compute_on_step=False, dist_sync_on_step=True) self.add_state("feat_data", default=[], dist_reduce_fx=None) self.add_state("vids_data", default=[], dist_reduce_fx=None) self.add_state("scene_data", default=[], dist_reduce_fx=None) self.top_k_list = set(top_k_list) self.max_k = max(self.top_k_list) def update(self, vid, invideo_scene_id, feat): assert isinstance(invideo_scene_id, torch.Tensor) assert isinstance(vid, torch.Tensor) assert isinstance(feat, torch.Tensor) self.feat_data.append(feat) self.vids_data.append(vid) self.scene_data.append(invideo_scene_id) def compute(self) -> torch.Tensor: score = defaultdict(dict) pool_feats = defaultdict(list) pool_invideo_scene_id = defaultdict(list) pool_gts = defaultdict(dict) num_data = 0 for vid, invideo_scene_id, gathered_feat in zip( self.vids_data, self.scene_data, self.feat_data ): vid = vid.item() invideo_scene_id = invideo_scene_id.item() if invideo_scene_id not in pool_gts[vid]: pool_gts[vid][invideo_scene_id] = set() pool_gts[vid][invideo_scene_id].add(len(pool_feats[vid])) pool_invideo_scene_id[vid].append(invideo_scene_id) pool_feats[vid].append(gathered_feat) num_data += 1 for top_k in self.top_k_list: score[top_k] = {"correct": 0, "total": 0} for vid, gt in pool_gts.items(): X = torch.stack(pool_feats[vid]) sim = torch.matmul(X, X.t()) sim = sim - 999 * torch.eye(sim.shape[0]).type_as(sim) indices = torch.argsort(sim, descending=True) assert indices.shape[1] >= self.max_k, f"{indices.shape[1]} >= {self.max_k}" indices = indices[:, : self.max_k] for j in range(indices.shape[0]): _cache = {"correct": 0, "total": 0} _query_scene_id = pool_invideo_scene_id[vid][j] for k in range(self.max_k): if _query_scene_id in gt: if indices[j][k].item() in gt[_query_scene_id]: _cache["correct"] += 1 _cache["total"] += 1 if k + 1 in self.top_k_list and len(gt[_query_scene_id]) > k: score[k + 1]["correct"] += _cache["correct"] score[k + 1]["total"] += _cache["total"] for top_k in self.top_k_list: assert score[top_k]["total"] > 0 score[top_k]["precision"] = ( 100.0 * score[top_k]["correct"] / score[top_k]["total"] ) del X, sim, indices, pool_feats, pool_invideo_scene_id, pool_gts torch.cuda.empty_cache() return score
true
true
1c48545f0baaba0b798a29f73f40c45f2e843e2d
29,175
py
Python
venv/lib/python3.8/site-packages/mpl_toolkits/axes_grid1/axes_divider.py
willBear/willBear-Fundamental_Analysis
bc67eb1e69dcf6765c0b77314d37f7f165a7318f
[ "MIT" ]
15
2020-06-29T08:33:39.000Z
2022-02-12T00:28:51.000Z
venv/lib/python3.8/site-packages/mpl_toolkits/axes_grid1/axes_divider.py
willBear/willBear-Fundamental_Analysis
bc67eb1e69dcf6765c0b77314d37f7f165a7318f
[ "MIT" ]
30
2020-04-15T19:37:40.000Z
2020-04-22T21:19:35.000Z
venv/lib/python3.8/site-packages/mpl_toolkits/axes_grid1/axes_divider.py
willBear/willBear-Fundamental_Analysis
bc67eb1e69dcf6765c0b77314d37f7f165a7318f
[ "MIT" ]
11
2019-01-21T17:51:48.000Z
2021-08-10T07:04:33.000Z
""" The axes_divider module provides helper classes to adjust the positions of multiple axes at drawing time. Divider: this is the class that is used to calculate the axes position. It divides the given rectangular area into several sub rectangles. You initialize the divider by setting the horizontal and vertical lists of sizes that the division will be based on. You then use the new_locator method, whose return value is a callable object that can be used to set the axes_locator of the axes. """ from matplotlib import cbook from matplotlib.axes import SubplotBase from matplotlib.gridspec import SubplotSpec, GridSpec import matplotlib.transforms as mtransforms from . import axes_size as Size class Divider: """ This class calculates the axes position. It divides the given rectangular area into several sub-rectangles. You initialize the divider by setting the horizontal and vertical lists of sizes (:mod:`mpl_toolkits.axes_grid.axes_size`) that the division will be based on. You then use the new_locator method to create a callable object that can be used as the axes_locator of the axes. """ def __init__(self, fig, pos, horizontal, vertical, aspect=None, anchor="C"): """ Parameters ---------- fig : Figure pos : tuple of 4 floats position of the rectangle that will be divided horizontal : list of :mod:`~mpl_toolkits.axes_grid.axes_size` sizes for horizontal division vertical : list of :mod:`~mpl_toolkits.axes_grid.axes_size` sizes for vertical division aspect : bool if True, the overall rectangular area is reduced so that the relative part of the horizontal and vertical scales have the same scale. anchor : {'C', 'SW', 'S', 'SE', 'E', 'NE', 'N', 'NW', 'W'} placement of the reduced rectangle when *aspect* is True """ self._fig = fig self._pos = pos self._horizontal = horizontal self._vertical = vertical self._anchor = anchor self._aspect = aspect self._xrefindex = 0 self._yrefindex = 0 self._locator = None def get_horizontal_sizes(self, renderer): return [s.get_size(renderer) for s in self.get_horizontal()] def get_vertical_sizes(self, renderer): return [s.get_size(renderer) for s in self.get_vertical()] def get_vsize_hsize(self): from .axes_size import AddList vsize = AddList(self.get_vertical()) hsize = AddList(self.get_horizontal()) return vsize, hsize @staticmethod def _calc_k(l, total_size): rs_sum, as_sum = 0., 0. for _rs, _as in l: rs_sum += _rs as_sum += _as if rs_sum != 0.: k = (total_size - as_sum) / rs_sum return k else: return 0. @staticmethod def _calc_offsets(l, k): offsets = [0.] for _rs, _as in l: offsets.append(offsets[-1] + _rs*k + _as) return offsets def set_position(self, pos): """ set the position of the rectangle. Parameters ---------- pos : tuple of 4 floats position of the rectangle that will be divided """ self._pos = pos def get_position(self): "return the position of the rectangle." return self._pos def set_anchor(self, anchor): """ Parameters ---------- anchor : {'C', 'SW', 'S', 'SE', 'E', 'NE', 'N', 'NW', 'W'} anchor position ===== ============ value description ===== ============ 'C' Center 'SW' bottom left 'S' bottom 'SE' bottom right 'E' right 'NE' top right 'N' top 'NW' top left 'W' left ===== ============ """ if len(anchor) != 2: cbook._check_in_list(mtransforms.Bbox.coefs, anchor=anchor) self._anchor = anchor def get_anchor(self): "return the anchor" return self._anchor def set_horizontal(self, h): """ Parameters ---------- h : list of :mod:`~mpl_toolkits.axes_grid.axes_size` sizes for horizontal division """ self._horizontal = h def get_horizontal(self): "return horizontal sizes" return self._horizontal def set_vertical(self, v): """ Parameters ---------- v : list of :mod:`~mpl_toolkits.axes_grid.axes_size` sizes for vertical division """ self._vertical = v def get_vertical(self): "return vertical sizes" return self._vertical def set_aspect(self, aspect=False): """ Parameters ---------- aspect : bool """ self._aspect = aspect def get_aspect(self): "return aspect" return self._aspect def set_locator(self, _locator): self._locator = _locator def get_locator(self): return self._locator def get_position_runtime(self, ax, renderer): if self._locator is None: return self.get_position() else: return self._locator(ax, renderer).bounds def locate(self, nx, ny, nx1=None, ny1=None, axes=None, renderer=None): """ Parameters ---------- nx, nx1 : int Integers specifying the column-position of the cell. When *nx1* is None, a single *nx*-th column is specified. Otherwise location of columns spanning between *nx* to *nx1* (but excluding *nx1*-th column) is specified. ny, ny1 : int Same as *nx* and *nx1*, but for row positions. axes renderer """ figW, figH = self._fig.get_size_inches() x, y, w, h = self.get_position_runtime(axes, renderer) hsizes = self.get_horizontal_sizes(renderer) vsizes = self.get_vertical_sizes(renderer) k_h = self._calc_k(hsizes, figW*w) k_v = self._calc_k(vsizes, figH*h) if self.get_aspect(): k = min(k_h, k_v) ox = self._calc_offsets(hsizes, k) oy = self._calc_offsets(vsizes, k) ww = (ox[-1] - ox[0])/figW hh = (oy[-1] - oy[0])/figH pb = mtransforms.Bbox.from_bounds(x, y, w, h) pb1 = mtransforms.Bbox.from_bounds(x, y, ww, hh) pb1_anchored = pb1.anchored(self.get_anchor(), pb) x0, y0 = pb1_anchored.x0, pb1_anchored.y0 else: ox = self._calc_offsets(hsizes, k_h) oy = self._calc_offsets(vsizes, k_v) x0, y0 = x, y if nx1 is None: nx1 = nx+1 if ny1 is None: ny1 = ny+1 x1, w1 = x0 + ox[nx]/figW, (ox[nx1] - ox[nx])/figW y1, h1 = y0 + oy[ny]/figH, (oy[ny1] - oy[ny])/figH return mtransforms.Bbox.from_bounds(x1, y1, w1, h1) def new_locator(self, nx, ny, nx1=None, ny1=None): """ Returns a new locator (:class:`mpl_toolkits.axes_grid.axes_divider.AxesLocator`) for specified cell. Parameters ---------- nx, nx1 : int Integers specifying the column-position of the cell. When *nx1* is None, a single *nx*-th column is specified. Otherwise location of columns spanning between *nx* to *nx1* (but excluding *nx1*-th column) is specified. ny, ny1 : int Same as *nx* and *nx1*, but for row positions. """ return AxesLocator(self, nx, ny, nx1, ny1) def append_size(self, position, size): if position == "left": self._horizontal.insert(0, size) self._xrefindex += 1 elif position == "right": self._horizontal.append(size) elif position == "bottom": self._vertical.insert(0, size) self._yrefindex += 1 elif position == "top": self._vertical.append(size) else: cbook._check_in_list(["left", "right", "bottom", "top"], position=position) def add_auto_adjustable_area(self, use_axes, pad=0.1, adjust_dirs=None, ): if adjust_dirs is None: adjust_dirs = ["left", "right", "bottom", "top"] from .axes_size import Padded, SizeFromFunc, GetExtentHelper for d in adjust_dirs: helper = GetExtentHelper(use_axes, d) size = SizeFromFunc(helper) padded_size = Padded(size, pad) # pad in inch self.append_size(d, padded_size) class AxesLocator: """ A simple callable object, initialized with AxesDivider class, returns the position and size of the given cell. """ def __init__(self, axes_divider, nx, ny, nx1=None, ny1=None): """ Parameters ---------- axes_divider : AxesDivider nx, nx1 : int Integers specifying the column-position of the cell. When *nx1* is None, a single *nx*-th column is specified. Otherwise location of columns spanning between *nx* to *nx1* (but excluding *nx1*-th column) is specified. ny, ny1 : int Same as *nx* and *nx1*, but for row positions. """ self._axes_divider = axes_divider _xrefindex = axes_divider._xrefindex _yrefindex = axes_divider._yrefindex self._nx, self._ny = nx - _xrefindex, ny - _yrefindex if nx1 is None: nx1 = nx+1 if ny1 is None: ny1 = ny+1 self._nx1 = nx1 - _xrefindex self._ny1 = ny1 - _yrefindex def __call__(self, axes, renderer): _xrefindex = self._axes_divider._xrefindex _yrefindex = self._axes_divider._yrefindex return self._axes_divider.locate(self._nx + _xrefindex, self._ny + _yrefindex, self._nx1 + _xrefindex, self._ny1 + _yrefindex, axes, renderer) def get_subplotspec(self): if hasattr(self._axes_divider, "get_subplotspec"): return self._axes_divider.get_subplotspec() else: return None class SubplotDivider(Divider): """ The Divider class whose rectangle area is specified as a subplot geometry. """ def __init__(self, fig, *args, horizontal=None, vertical=None, aspect=None, anchor='C'): """ Parameters ---------- fig : `matplotlib.figure.Figure` *args : tuple (*nrows*, *ncols*, *index*) or int The array of subplots in the figure has dimensions ``(nrows, ncols)``, and *index* is the index of the subplot being created. *index* starts at 1 in the upper left corner and increases to the right. If *nrows*, *ncols*, and *index* are all single digit numbers, then *args* can be passed as a single 3-digit number (e.g. 234 for (2, 3, 4)). """ self.figure = fig if len(args) == 1: if isinstance(args[0], SubplotSpec): self._subplotspec = args[0] else: try: s = str(int(args[0])) rows, cols, num = map(int, s) except ValueError: raise ValueError( 'Single argument to subplot must be a 3-digit integer') self._subplotspec = GridSpec(rows, cols)[num-1] # num - 1 for converting from MATLAB to python indexing elif len(args) == 3: rows, cols, num = args rows = int(rows) cols = int(cols) if isinstance(num, tuple) and len(num) == 2: num = [int(n) for n in num] self._subplotspec = GridSpec(rows, cols)[num[0]-1:num[1]] else: self._subplotspec = GridSpec(rows, cols)[int(num)-1] # num - 1 for converting from MATLAB to python indexing else: raise ValueError(f'Illegal argument(s) to subplot: {args}') # total = rows*cols # num -= 1 # convert from matlab to python indexing # # i.e., num in range(0, total) # if num >= total: # raise ValueError( 'Subplot number exceeds total subplots') # self._rows = rows # self._cols = cols # self._num = num # self.update_params() # sets self.fixbox self.update_params() pos = self.figbox.bounds Divider.__init__(self, fig, pos, horizontal or [], vertical or [], aspect=aspect, anchor=anchor) def get_position(self): "return the bounds of the subplot box" self.update_params() # update self.figbox return self.figbox.bounds # def update_params(self): # 'update the subplot position from fig.subplotpars' # rows = self._rows # cols = self._cols # num = self._num # pars = self.figure.subplotpars # left = pars.left # right = pars.right # bottom = pars.bottom # top = pars.top # wspace = pars.wspace # hspace = pars.hspace # totWidth = right-left # totHeight = top-bottom # figH = totHeight/(rows + hspace*(rows-1)) # sepH = hspace*figH # figW = totWidth/(cols + wspace*(cols-1)) # sepW = wspace*figW # rowNum, colNum = divmod(num, cols) # figBottom = top - (rowNum+1)*figH - rowNum*sepH # figLeft = left + colNum*(figW + sepW) # self.figbox = mtransforms.Bbox.from_bounds(figLeft, figBottom, # figW, figH) def update_params(self): """Update the subplot position from fig.subplotpars.""" self.figbox = self.get_subplotspec().get_position(self.figure) def get_geometry(self): """Get the subplot geometry, e.g., (2, 2, 3).""" rows, cols, num1, num2 = self.get_subplotspec().get_geometry() return rows, cols, num1+1 # for compatibility # COVERAGE NOTE: Never used internally or from examples def change_geometry(self, numrows, numcols, num): """Change subplot geometry, e.g., from (1, 1, 1) to (2, 2, 3).""" self._subplotspec = GridSpec(numrows, numcols)[num-1] self.update_params() self.set_position(self.figbox) def get_subplotspec(self): """Get the SubplotSpec instance.""" return self._subplotspec def set_subplotspec(self, subplotspec): """Set the SubplotSpec instance.""" self._subplotspec = subplotspec class AxesDivider(Divider): """ Divider based on the pre-existing axes. """ def __init__(self, axes, xref=None, yref=None): """ Parameters ---------- axes : :class:`~matplotlib.axes.Axes` xref yref """ self._axes = axes if xref is None: self._xref = Size.AxesX(axes) else: self._xref = xref if yref is None: self._yref = Size.AxesY(axes) else: self._yref = yref Divider.__init__(self, fig=axes.get_figure(), pos=None, horizontal=[self._xref], vertical=[self._yref], aspect=None, anchor="C") def _get_new_axes(self, *, axes_class=None, **kwargs): axes = self._axes if axes_class is None: if isinstance(axes, SubplotBase): axes_class = axes._axes_class else: axes_class = type(axes) return axes_class(axes.get_figure(), axes.get_position(original=True), **kwargs) def new_horizontal(self, size, pad=None, pack_start=False, **kwargs): """ Add a new axes on the right (or left) side of the main axes. Parameters ---------- size : :mod:`~mpl_toolkits.axes_grid.axes_size` or float or str A width of the axes. If float or string is given, *from_any* function is used to create the size, with *ref_size* set to AxesX instance of the current axes. pad : :mod:`~mpl_toolkits.axes_grid.axes_size` or float or str Pad between the axes. It takes same argument as *size*. pack_start : bool If False, the new axes is appended at the end of the list, i.e., it became the right-most axes. If True, it is inserted at the start of the list, and becomes the left-most axes. **kwargs All extra keywords arguments are passed to the created axes. If *axes_class* is given, the new axes will be created as an instance of the given class. Otherwise, the same class of the main axes will be used. """ if pad is None: cbook.warn_deprecated( "3.2", message="In a future version, 'pad' will default to " "rcParams['figure.subplot.wspace']. Set pad=0 to keep the " "old behavior.") if pad: if not isinstance(pad, Size._Base): pad = Size.from_any(pad, fraction_ref=self._xref) if pack_start: self._horizontal.insert(0, pad) self._xrefindex += 1 else: self._horizontal.append(pad) if not isinstance(size, Size._Base): size = Size.from_any(size, fraction_ref=self._xref) if pack_start: self._horizontal.insert(0, size) self._xrefindex += 1 locator = self.new_locator(nx=0, ny=self._yrefindex) else: self._horizontal.append(size) locator = self.new_locator( nx=len(self._horizontal) - 1, ny=self._yrefindex) ax = self._get_new_axes(**kwargs) ax.set_axes_locator(locator) return ax def new_vertical(self, size, pad=None, pack_start=False, **kwargs): """ Add a new axes on the top (or bottom) side of the main axes. Parameters ---------- size : :mod:`~mpl_toolkits.axes_grid.axes_size` or float or str A height of the axes. If float or string is given, *from_any* function is used to create the size, with *ref_size* set to AxesX instance of the current axes. pad : :mod:`~mpl_toolkits.axes_grid.axes_size` or float or str Pad between the axes. It takes same argument as *size*. pack_start : bool If False, the new axes is appended at the end of the list, i.e., it became the right-most axes. If True, it is inserted at the start of the list, and becomes the left-most axes. **kwargs All extra keywords arguments are passed to the created axes. If *axes_class* is given, the new axes will be created as an instance of the given class. Otherwise, the same class of the main axes will be used. """ if pad is None: cbook.warn_deprecated( "3.2", message="In a future version, 'pad' will default to " "rcParams['figure.subplot.hspace']. Set pad=0 to keep the " "old behavior.") if pad: if not isinstance(pad, Size._Base): pad = Size.from_any(pad, fraction_ref=self._yref) if pack_start: self._vertical.insert(0, pad) self._yrefindex += 1 else: self._vertical.append(pad) if not isinstance(size, Size._Base): size = Size.from_any(size, fraction_ref=self._yref) if pack_start: self._vertical.insert(0, size) self._yrefindex += 1 locator = self.new_locator(nx=self._xrefindex, ny=0) else: self._vertical.append(size) locator = self.new_locator( nx=self._xrefindex, ny=len(self._vertical)-1) ax = self._get_new_axes(**kwargs) ax.set_axes_locator(locator) return ax def append_axes(self, position, size, pad=None, add_to_figure=True, **kwargs): """ Create an axes at the given *position* with the same height (or width) of the main axes. *position* ["left"|"right"|"bottom"|"top"] *size* and *pad* should be axes_grid.axes_size compatible. """ if position == "left": ax = self.new_horizontal(size, pad, pack_start=True, **kwargs) elif position == "right": ax = self.new_horizontal(size, pad, pack_start=False, **kwargs) elif position == "bottom": ax = self.new_vertical(size, pad, pack_start=True, **kwargs) elif position == "top": ax = self.new_vertical(size, pad, pack_start=False, **kwargs) else: cbook._check_in_list(["left", "right", "bottom", "top"], position=position) if add_to_figure: self._fig.add_axes(ax) return ax def get_aspect(self): if self._aspect is None: aspect = self._axes.get_aspect() if aspect == "auto": return False else: return True else: return self._aspect def get_position(self): if self._pos is None: bbox = self._axes.get_position(original=True) return bbox.bounds else: return self._pos def get_anchor(self): if self._anchor is None: return self._axes.get_anchor() else: return self._anchor def get_subplotspec(self): if hasattr(self._axes, "get_subplotspec"): return self._axes.get_subplotspec() else: return None class HBoxDivider(SubplotDivider): def __init__(self, fig, *args, **kwargs): SubplotDivider.__init__(self, fig, *args, **kwargs) @staticmethod def _determine_karray(equivalent_sizes, appended_sizes, max_equivalent_size, total_appended_size): n = len(equivalent_sizes) import numpy as np A = np.mat(np.zeros((n+1, n+1), dtype="d")) B = np.zeros((n+1), dtype="d") # AxK = B # populated A for i, (r, a) in enumerate(equivalent_sizes): A[i, i] = r A[i, -1] = -1 B[i] = -a A[-1, :-1] = [r for r, a in appended_sizes] B[-1] = total_appended_size - sum([a for rs, a in appended_sizes]) karray_H = (A.I*np.mat(B).T).A1 karray = karray_H[:-1] H = karray_H[-1] if H > max_equivalent_size: karray = ((max_equivalent_size - np.array([a for r, a in equivalent_sizes])) / np.array([r for r, a in equivalent_sizes])) return karray @staticmethod def _calc_offsets(appended_sizes, karray): offsets = [0.] for (r, a), k in zip(appended_sizes, karray): offsets.append(offsets[-1] + r*k + a) return offsets def new_locator(self, nx, nx1=None): """ Create a new `~mpl_toolkits.axes_grid.axes_divider.AxesLocator` for the specified cell. Parameters ---------- nx, nx1 : int Integers specifying the column-position of the cell. When *nx1* is None, a single *nx*-th column is specified. Otherwise location of columns spanning between *nx* to *nx1* (but excluding *nx1*-th column) is specified. ny, ny1 : int Same as *nx* and *nx1*, but for row positions. """ return AxesLocator(self, nx, 0, nx1, None) def _locate(self, x, y, w, h, y_equivalent_sizes, x_appended_sizes, figW, figH): """ Parameters ---------- x y w h y_equivalent_sizes x_appended_sizes figW figH """ equivalent_sizes = y_equivalent_sizes appended_sizes = x_appended_sizes max_equivalent_size = figH*h total_appended_size = figW*w karray = self._determine_karray(equivalent_sizes, appended_sizes, max_equivalent_size, total_appended_size) ox = self._calc_offsets(appended_sizes, karray) ww = (ox[-1] - ox[0])/figW ref_h = equivalent_sizes[0] hh = (karray[0]*ref_h[0] + ref_h[1])/figH pb = mtransforms.Bbox.from_bounds(x, y, w, h) pb1 = mtransforms.Bbox.from_bounds(x, y, ww, hh) pb1_anchored = pb1.anchored(self.get_anchor(), pb) x0, y0 = pb1_anchored.x0, pb1_anchored.y0 return x0, y0, ox, hh def locate(self, nx, ny, nx1=None, ny1=None, axes=None, renderer=None): """ Parameters ---------- axes_divider : AxesDivider nx, nx1 : int Integers specifying the column-position of the cell. When *nx1* is None, a single *nx*-th column is specified. Otherwise location of columns spanning between *nx* to *nx1* (but excluding *nx1*-th column) is specified. ny, ny1 : int Same as *nx* and *nx1*, but for row positions. axes renderer """ figW, figH = self._fig.get_size_inches() x, y, w, h = self.get_position_runtime(axes, renderer) y_equivalent_sizes = self.get_vertical_sizes(renderer) x_appended_sizes = self.get_horizontal_sizes(renderer) x0, y0, ox, hh = self._locate(x, y, w, h, y_equivalent_sizes, x_appended_sizes, figW, figH) if nx1 is None: nx1 = nx+1 x1, w1 = x0 + ox[nx]/figW, (ox[nx1] - ox[nx])/figW y1, h1 = y0, hh return mtransforms.Bbox.from_bounds(x1, y1, w1, h1) class VBoxDivider(HBoxDivider): """ The Divider class whose rectangle area is specified as a subplot geometry. """ def new_locator(self, ny, ny1=None): """ Create a new `~mpl_toolkits.axes_grid.axes_divider.AxesLocator` for the specified cell. Parameters ---------- ny, ny1 : int Integers specifying the row-position of the cell. When *ny1* is None, a single *ny*-th row is specified. Otherwise location of rows spanning between *ny* to *ny1* (but excluding *ny1*-th row) is specified. """ return AxesLocator(self, 0, ny, None, ny1) def locate(self, nx, ny, nx1=None, ny1=None, axes=None, renderer=None): """ Parameters ---------- axes_divider : AxesDivider nx, nx1 : int Integers specifying the column-position of the cell. When *nx1* is None, a single *nx*-th column is specified. Otherwise location of columns spanning between *nx* to *nx1* (but excluding *nx1*-th column) is specified. ny, ny1 : int Same as *nx* and *nx1*, but for row positions. axes renderer """ figW, figH = self._fig.get_size_inches() x, y, w, h = self.get_position_runtime(axes, renderer) x_equivalent_sizes = self.get_horizontal_sizes(renderer) y_appended_sizes = self.get_vertical_sizes(renderer) y0, x0, oy, ww = self._locate(y, x, h, w, x_equivalent_sizes, y_appended_sizes, figH, figW) if ny1 is None: ny1 = ny+1 x1, w1 = x0, ww y1, h1 = y0 + oy[ny]/figH, (oy[ny1] - oy[ny])/figH return mtransforms.Bbox.from_bounds(x1, y1, w1, h1) def make_axes_locatable(axes): divider = AxesDivider(axes) locator = divider.new_locator(nx=0, ny=0) axes.set_axes_locator(locator) return divider def make_axes_area_auto_adjustable(ax, use_axes=None, pad=0.1, adjust_dirs=None): if adjust_dirs is None: adjust_dirs = ["left", "right", "bottom", "top"] divider = make_axes_locatable(ax) if use_axes is None: use_axes = ax divider.add_auto_adjustable_area(use_axes=use_axes, pad=pad, adjust_dirs=adjust_dirs)
33.650519
79
0.55078
from matplotlib import cbook from matplotlib.axes import SubplotBase from matplotlib.gridspec import SubplotSpec, GridSpec import matplotlib.transforms as mtransforms from . import axes_size as Size class Divider: def __init__(self, fig, pos, horizontal, vertical, aspect=None, anchor="C"): self._fig = fig self._pos = pos self._horizontal = horizontal self._vertical = vertical self._anchor = anchor self._aspect = aspect self._xrefindex = 0 self._yrefindex = 0 self._locator = None def get_horizontal_sizes(self, renderer): return [s.get_size(renderer) for s in self.get_horizontal()] def get_vertical_sizes(self, renderer): return [s.get_size(renderer) for s in self.get_vertical()] def get_vsize_hsize(self): from .axes_size import AddList vsize = AddList(self.get_vertical()) hsize = AddList(self.get_horizontal()) return vsize, hsize @staticmethod def _calc_k(l, total_size): rs_sum, as_sum = 0., 0. for _rs, _as in l: rs_sum += _rs as_sum += _as if rs_sum != 0.: k = (total_size - as_sum) / rs_sum return k else: return 0. @staticmethod def _calc_offsets(l, k): offsets = [0.] for _rs, _as in l: offsets.append(offsets[-1] + _rs*k + _as) return offsets def set_position(self, pos): self._pos = pos def get_position(self): return self._pos def set_anchor(self, anchor): if len(anchor) != 2: cbook._check_in_list(mtransforms.Bbox.coefs, anchor=anchor) self._anchor = anchor def get_anchor(self): return self._anchor def set_horizontal(self, h): self._horizontal = h def get_horizontal(self): return self._horizontal def set_vertical(self, v): self._vertical = v def get_vertical(self): return self._vertical def set_aspect(self, aspect=False): self._aspect = aspect def get_aspect(self): return self._aspect def set_locator(self, _locator): self._locator = _locator def get_locator(self): return self._locator def get_position_runtime(self, ax, renderer): if self._locator is None: return self.get_position() else: return self._locator(ax, renderer).bounds def locate(self, nx, ny, nx1=None, ny1=None, axes=None, renderer=None): figW, figH = self._fig.get_size_inches() x, y, w, h = self.get_position_runtime(axes, renderer) hsizes = self.get_horizontal_sizes(renderer) vsizes = self.get_vertical_sizes(renderer) k_h = self._calc_k(hsizes, figW*w) k_v = self._calc_k(vsizes, figH*h) if self.get_aspect(): k = min(k_h, k_v) ox = self._calc_offsets(hsizes, k) oy = self._calc_offsets(vsizes, k) ww = (ox[-1] - ox[0])/figW hh = (oy[-1] - oy[0])/figH pb = mtransforms.Bbox.from_bounds(x, y, w, h) pb1 = mtransforms.Bbox.from_bounds(x, y, ww, hh) pb1_anchored = pb1.anchored(self.get_anchor(), pb) x0, y0 = pb1_anchored.x0, pb1_anchored.y0 else: ox = self._calc_offsets(hsizes, k_h) oy = self._calc_offsets(vsizes, k_v) x0, y0 = x, y if nx1 is None: nx1 = nx+1 if ny1 is None: ny1 = ny+1 x1, w1 = x0 + ox[nx]/figW, (ox[nx1] - ox[nx])/figW y1, h1 = y0 + oy[ny]/figH, (oy[ny1] - oy[ny])/figH return mtransforms.Bbox.from_bounds(x1, y1, w1, h1) def new_locator(self, nx, ny, nx1=None, ny1=None): return AxesLocator(self, nx, ny, nx1, ny1) def append_size(self, position, size): if position == "left": self._horizontal.insert(0, size) self._xrefindex += 1 elif position == "right": self._horizontal.append(size) elif position == "bottom": self._vertical.insert(0, size) self._yrefindex += 1 elif position == "top": self._vertical.append(size) else: cbook._check_in_list(["left", "right", "bottom", "top"], position=position) def add_auto_adjustable_area(self, use_axes, pad=0.1, adjust_dirs=None, ): if adjust_dirs is None: adjust_dirs = ["left", "right", "bottom", "top"] from .axes_size import Padded, SizeFromFunc, GetExtentHelper for d in adjust_dirs: helper = GetExtentHelper(use_axes, d) size = SizeFromFunc(helper) padded_size = Padded(size, pad) self.append_size(d, padded_size) class AxesLocator: def __init__(self, axes_divider, nx, ny, nx1=None, ny1=None): self._axes_divider = axes_divider _xrefindex = axes_divider._xrefindex _yrefindex = axes_divider._yrefindex self._nx, self._ny = nx - _xrefindex, ny - _yrefindex if nx1 is None: nx1 = nx+1 if ny1 is None: ny1 = ny+1 self._nx1 = nx1 - _xrefindex self._ny1 = ny1 - _yrefindex def __call__(self, axes, renderer): _xrefindex = self._axes_divider._xrefindex _yrefindex = self._axes_divider._yrefindex return self._axes_divider.locate(self._nx + _xrefindex, self._ny + _yrefindex, self._nx1 + _xrefindex, self._ny1 + _yrefindex, axes, renderer) def get_subplotspec(self): if hasattr(self._axes_divider, "get_subplotspec"): return self._axes_divider.get_subplotspec() else: return None class SubplotDivider(Divider): def __init__(self, fig, *args, horizontal=None, vertical=None, aspect=None, anchor='C'): self.figure = fig if len(args) == 1: if isinstance(args[0], SubplotSpec): self._subplotspec = args[0] else: try: s = str(int(args[0])) rows, cols, num = map(int, s) except ValueError: raise ValueError( 'Single argument to subplot must be a 3-digit integer') self._subplotspec = GridSpec(rows, cols)[num-1] elif len(args) == 3: rows, cols, num = args rows = int(rows) cols = int(cols) if isinstance(num, tuple) and len(num) == 2: num = [int(n) for n in num] self._subplotspec = GridSpec(rows, cols)[num[0]-1:num[1]] else: self._subplotspec = GridSpec(rows, cols)[int(num)-1] else: raise ValueError(f'Illegal argument(s) to subplot: {args}') self.update_params() pos = self.figbox.bounds Divider.__init__(self, fig, pos, horizontal or [], vertical or [], aspect=aspect, anchor=anchor) def get_position(self): self.update_params() return self.figbox.bounds def update_params(self): self.figbox = self.get_subplotspec().get_position(self.figure) def get_geometry(self): rows, cols, num1, num2 = self.get_subplotspec().get_geometry() return rows, cols, num1+1 def change_geometry(self, numrows, numcols, num): self._subplotspec = GridSpec(numrows, numcols)[num-1] self.update_params() self.set_position(self.figbox) def get_subplotspec(self): return self._subplotspec def set_subplotspec(self, subplotspec): self._subplotspec = subplotspec class AxesDivider(Divider): def __init__(self, axes, xref=None, yref=None): self._axes = axes if xref is None: self._xref = Size.AxesX(axes) else: self._xref = xref if yref is None: self._yref = Size.AxesY(axes) else: self._yref = yref Divider.__init__(self, fig=axes.get_figure(), pos=None, horizontal=[self._xref], vertical=[self._yref], aspect=None, anchor="C") def _get_new_axes(self, *, axes_class=None, **kwargs): axes = self._axes if axes_class is None: if isinstance(axes, SubplotBase): axes_class = axes._axes_class else: axes_class = type(axes) return axes_class(axes.get_figure(), axes.get_position(original=True), **kwargs) def new_horizontal(self, size, pad=None, pack_start=False, **kwargs): if pad is None: cbook.warn_deprecated( "3.2", message="In a future version, 'pad' will default to " "rcParams['figure.subplot.wspace']. Set pad=0 to keep the " "old behavior.") if pad: if not isinstance(pad, Size._Base): pad = Size.from_any(pad, fraction_ref=self._xref) if pack_start: self._horizontal.insert(0, pad) self._xrefindex += 1 else: self._horizontal.append(pad) if not isinstance(size, Size._Base): size = Size.from_any(size, fraction_ref=self._xref) if pack_start: self._horizontal.insert(0, size) self._xrefindex += 1 locator = self.new_locator(nx=0, ny=self._yrefindex) else: self._horizontal.append(size) locator = self.new_locator( nx=len(self._horizontal) - 1, ny=self._yrefindex) ax = self._get_new_axes(**kwargs) ax.set_axes_locator(locator) return ax def new_vertical(self, size, pad=None, pack_start=False, **kwargs): if pad is None: cbook.warn_deprecated( "3.2", message="In a future version, 'pad' will default to " "rcParams['figure.subplot.hspace']. Set pad=0 to keep the " "old behavior.") if pad: if not isinstance(pad, Size._Base): pad = Size.from_any(pad, fraction_ref=self._yref) if pack_start: self._vertical.insert(0, pad) self._yrefindex += 1 else: self._vertical.append(pad) if not isinstance(size, Size._Base): size = Size.from_any(size, fraction_ref=self._yref) if pack_start: self._vertical.insert(0, size) self._yrefindex += 1 locator = self.new_locator(nx=self._xrefindex, ny=0) else: self._vertical.append(size) locator = self.new_locator( nx=self._xrefindex, ny=len(self._vertical)-1) ax = self._get_new_axes(**kwargs) ax.set_axes_locator(locator) return ax def append_axes(self, position, size, pad=None, add_to_figure=True, **kwargs): if position == "left": ax = self.new_horizontal(size, pad, pack_start=True, **kwargs) elif position == "right": ax = self.new_horizontal(size, pad, pack_start=False, **kwargs) elif position == "bottom": ax = self.new_vertical(size, pad, pack_start=True, **kwargs) elif position == "top": ax = self.new_vertical(size, pad, pack_start=False, **kwargs) else: cbook._check_in_list(["left", "right", "bottom", "top"], position=position) if add_to_figure: self._fig.add_axes(ax) return ax def get_aspect(self): if self._aspect is None: aspect = self._axes.get_aspect() if aspect == "auto": return False else: return True else: return self._aspect def get_position(self): if self._pos is None: bbox = self._axes.get_position(original=True) return bbox.bounds else: return self._pos def get_anchor(self): if self._anchor is None: return self._axes.get_anchor() else: return self._anchor def get_subplotspec(self): if hasattr(self._axes, "get_subplotspec"): return self._axes.get_subplotspec() else: return None class HBoxDivider(SubplotDivider): def __init__(self, fig, *args, **kwargs): SubplotDivider.__init__(self, fig, *args, **kwargs) @staticmethod def _determine_karray(equivalent_sizes, appended_sizes, max_equivalent_size, total_appended_size): n = len(equivalent_sizes) import numpy as np A = np.mat(np.zeros((n+1, n+1), dtype="d")) B = np.zeros((n+1), dtype="d") for i, (r, a) in enumerate(equivalent_sizes): A[i, i] = r A[i, -1] = -1 B[i] = -a A[-1, :-1] = [r for r, a in appended_sizes] B[-1] = total_appended_size - sum([a for rs, a in appended_sizes]) karray_H = (A.I*np.mat(B).T).A1 karray = karray_H[:-1] H = karray_H[-1] if H > max_equivalent_size: karray = ((max_equivalent_size - np.array([a for r, a in equivalent_sizes])) / np.array([r for r, a in equivalent_sizes])) return karray @staticmethod def _calc_offsets(appended_sizes, karray): offsets = [0.] for (r, a), k in zip(appended_sizes, karray): offsets.append(offsets[-1] + r*k + a) return offsets def new_locator(self, nx, nx1=None): return AxesLocator(self, nx, 0, nx1, None) def _locate(self, x, y, w, h, y_equivalent_sizes, x_appended_sizes, figW, figH): equivalent_sizes = y_equivalent_sizes appended_sizes = x_appended_sizes max_equivalent_size = figH*h total_appended_size = figW*w karray = self._determine_karray(equivalent_sizes, appended_sizes, max_equivalent_size, total_appended_size) ox = self._calc_offsets(appended_sizes, karray) ww = (ox[-1] - ox[0])/figW ref_h = equivalent_sizes[0] hh = (karray[0]*ref_h[0] + ref_h[1])/figH pb = mtransforms.Bbox.from_bounds(x, y, w, h) pb1 = mtransforms.Bbox.from_bounds(x, y, ww, hh) pb1_anchored = pb1.anchored(self.get_anchor(), pb) x0, y0 = pb1_anchored.x0, pb1_anchored.y0 return x0, y0, ox, hh def locate(self, nx, ny, nx1=None, ny1=None, axes=None, renderer=None): figW, figH = self._fig.get_size_inches() x, y, w, h = self.get_position_runtime(axes, renderer) y_equivalent_sizes = self.get_vertical_sizes(renderer) x_appended_sizes = self.get_horizontal_sizes(renderer) x0, y0, ox, hh = self._locate(x, y, w, h, y_equivalent_sizes, x_appended_sizes, figW, figH) if nx1 is None: nx1 = nx+1 x1, w1 = x0 + ox[nx]/figW, (ox[nx1] - ox[nx])/figW y1, h1 = y0, hh return mtransforms.Bbox.from_bounds(x1, y1, w1, h1) class VBoxDivider(HBoxDivider): def new_locator(self, ny, ny1=None): return AxesLocator(self, 0, ny, None, ny1) def locate(self, nx, ny, nx1=None, ny1=None, axes=None, renderer=None): figW, figH = self._fig.get_size_inches() x, y, w, h = self.get_position_runtime(axes, renderer) x_equivalent_sizes = self.get_horizontal_sizes(renderer) y_appended_sizes = self.get_vertical_sizes(renderer) y0, x0, oy, ww = self._locate(y, x, h, w, x_equivalent_sizes, y_appended_sizes, figH, figW) if ny1 is None: ny1 = ny+1 x1, w1 = x0, ww y1, h1 = y0 + oy[ny]/figH, (oy[ny1] - oy[ny])/figH return mtransforms.Bbox.from_bounds(x1, y1, w1, h1) def make_axes_locatable(axes): divider = AxesDivider(axes) locator = divider.new_locator(nx=0, ny=0) axes.set_axes_locator(locator) return divider def make_axes_area_auto_adjustable(ax, use_axes=None, pad=0.1, adjust_dirs=None): if adjust_dirs is None: adjust_dirs = ["left", "right", "bottom", "top"] divider = make_axes_locatable(ax) if use_axes is None: use_axes = ax divider.add_auto_adjustable_area(use_axes=use_axes, pad=pad, adjust_dirs=adjust_dirs)
true
true
1c4854884ba21127c83dd2876aef95bd6d1e9f13
1,785
py
Python
data/p4VQE/R4/benchmark/startCirq672.py
UCLA-SEAL/QDiff
d968cbc47fe926b7f88b4adf10490f1edd6f8819
[ "BSD-3-Clause" ]
null
null
null
data/p4VQE/R4/benchmark/startCirq672.py
UCLA-SEAL/QDiff
d968cbc47fe926b7f88b4adf10490f1edd6f8819
[ "BSD-3-Clause" ]
null
null
null
data/p4VQE/R4/benchmark/startCirq672.py
UCLA-SEAL/QDiff
d968cbc47fe926b7f88b4adf10490f1edd6f8819
[ "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Time : 5/15/20 4:49 PM # @File : grover.py # qubit number=4 # total number=12 import cirq import cirq.google as cg from typing import Optional import sys from math import log2 import numpy as np #thatsNoCode from cirq.contrib.svg import SVGCircuit # Symbols for the rotation angles in the QAOA circuit. def make_circuit(n: int, input_qubit): c = cirq.Circuit() # circuit begin c.append(cirq.H.on(input_qubit[0])) # number=1 c.append(cirq.CNOT.on(input_qubit[3],input_qubit[0])) # number=9 c.append(cirq.Z.on(input_qubit[3])) # number=10 c.append(cirq.CNOT.on(input_qubit[3],input_qubit[0])) # number=11 c.append(cirq.Z.on(input_qubit[1])) # number=8 c.append(cirq.H.on(input_qubit[2])) # number=3 c.append(cirq.SWAP.on(input_qubit[3],input_qubit[0])) # number=5 c.append(cirq.SWAP.on(input_qubit[3],input_qubit[0])) # number=6 # circuit end c.append(cirq.measure(*input_qubit, key='result')) return c def bitstring(bits): return ''.join(str(int(b)) for b in bits) if __name__ == '__main__': qubit_count = 4 input_qubits = [cirq.GridQubit(i, 0) for i in range(qubit_count)] circuit = make_circuit(qubit_count,input_qubits) circuit = cg.optimized_for_sycamore(circuit, optimizer_type='sqrt_iswap') circuit_sample_count =2000 simulator = cirq.Simulator() result = simulator.run(circuit, repetitions=circuit_sample_count) frequencies = result.histogram(key='result', fold_func=bitstring) writefile = open("../data/startCirq672.csv","w+") print(format(frequencies),file=writefile) print("results end", file=writefile) print(circuit.__len__(), file=writefile) print(circuit,file=writefile) writefile.close()
28.333333
77
0.696359
import cirq import cirq.google as cg from typing import Optional import sys from math import log2 import numpy as np from cirq.contrib.svg import SVGCircuit def make_circuit(n: int, input_qubit): c = cirq.Circuit() c.append(cirq.H.on(input_qubit[0])) c.append(cirq.CNOT.on(input_qubit[3],input_qubit[0])) c.append(cirq.Z.on(input_qubit[3])) c.append(cirq.CNOT.on(input_qubit[3],input_qubit[0])) c.append(cirq.Z.on(input_qubit[1])) c.append(cirq.H.on(input_qubit[2])) c.append(cirq.SWAP.on(input_qubit[3],input_qubit[0])) c.append(cirq.SWAP.on(input_qubit[3],input_qubit[0])) c.append(cirq.measure(*input_qubit, key='result')) return c def bitstring(bits): return ''.join(str(int(b)) for b in bits) if __name__ == '__main__': qubit_count = 4 input_qubits = [cirq.GridQubit(i, 0) for i in range(qubit_count)] circuit = make_circuit(qubit_count,input_qubits) circuit = cg.optimized_for_sycamore(circuit, optimizer_type='sqrt_iswap') circuit_sample_count =2000 simulator = cirq.Simulator() result = simulator.run(circuit, repetitions=circuit_sample_count) frequencies = result.histogram(key='result', fold_func=bitstring) writefile = open("../data/startCirq672.csv","w+") print(format(frequencies),file=writefile) print("results end", file=writefile) print(circuit.__len__(), file=writefile) print(circuit,file=writefile) writefile.close()
true
true
1c4855277424f0ca199fd32a8a23f150c55dae3d
17,289
py
Python
utils/fileutils.py
sag-tgo/xpybuild
f114043e441681b1d0c35f9833d02f7fd8a52329
[ "Apache-2.0" ]
null
null
null
utils/fileutils.py
sag-tgo/xpybuild
f114043e441681b1d0c35f9833d02f7fd8a52329
[ "Apache-2.0" ]
null
null
null
utils/fileutils.py
sag-tgo/xpybuild
f114043e441681b1d0c35f9833d02f7fd8a52329
[ "Apache-2.0" ]
null
null
null
# fileutils - helper methods related to the file system # # Copyright (c) 2013 - 2019 Software AG, Darmstadt, Germany and/or its licensors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # $Id: fileutils.py 301527 2017-02-06 15:31:43Z matj $ # """ @undocumented: _getStatCacheSize """ import shutil, os, os.path, time, platform, threading import stat, sys from utils.flatten import getStringList import subprocess, errno import logging log = logging.getLogger('fileutils') __isWindows = platform.system()=='Windows' if __isWindows: # ugly ugly hacks due to stupid windows filesystem semantics. See http://bugs.python.org/issue4944 try: import win32file class openForWrite: def __init__(self, dest, mode): assert mode == 'wb' # the Win32 API only has binary mode, not text mode, so only accept opening in this mode. This also means you must deal with any line ending issues yourself when using this. self.dest = dest def __enter__(self): self.Fd = win32file.CreateFile(self.dest, win32file.GENERIC_WRITE, win32file.FILE_SHARE_READ | win32file.FILE_SHARE_WRITE | win32file.FILE_SHARE_DELETE, None, win32file.CREATE_ALWAYS, win32file.FILE_ATTRIBUTE_NORMAL, None) return self def write(self, string): if isinstance(string, unicode): # must pass byte strings not unicode objects when writing in binary mode; for backwards compatibility, automatically convert 7-bit ascii (which is what the Unix impl will do), but give errors for anything else as user should be explicitly specifying their encoding string = string.encode('ascii', errors='strict') assert isinstance(string, str), 'cannot write unicode character string to binary file; please use byte str instead: %s'%repr(string) win32file.WriteFile(self.Fd, string) def writelines(self, lines): for l in lines: self.write(l) def close(self): win32file.CloseHandle(self.Fd) def __exit__(self, ex_type, ex_val, tb): self.close() except: openForWrite = open else: openForWrite = open def mkdir(newdir): """ Recursively create the specified directory if it doesn't already exist. If it does, exit without error. @param newdir: The path to create. @return: newdir, to allow fluent use of this method. """ origdir = newdir newdir=normLongPath(newdir) if os.path.isdir(newdir): # already exists return origdir if os.path.isfile(newdir): raise IOError("A file with the same name as the desired dir, '%s', already exists" % newdir) #when multiple threads/processes are creating directories #at the same time, it can be a race try: os.makedirs(newdir) except Exception, e: if os.path.isdir(newdir): pass else: raise IOError('Problem creating directory %s: %s' % (newdir, e)) return origdir def deleteDir(path, allowRetry=True): """ Recursively delete the contents of a directory. Contains magic hacks so it works even on paths that exceed the Windows MAX_PATH 260 character length. @param path: the path to delete. @param allowRetry: set to False to disable automatic retry of the deletion after a few seconds (in case the error was transient) """ def handleRemoveReadonly(func, path, exc): # once we've got this working reliably, might reduce the level of some of these log statements excvalue = exc[1] log.info("handleRemoveReadonly: error removing path %s (%s %s), will try harder; exists=%s" % (path,errno.errorcode.get(excvalue.errno, "EUNKNOWN"), func, os.path.exists(path))) if func in (os.rmdir, os.remove): if not os.path.exists(path): # no idea why this happens, but on windows it does log.info("handleRemoveReadonly: suppressing spurious remove exception for already-deleted path: %s", path) return if excvalue.errno == errno.EACCES: # access denied, make it writable first try: os.chmod(path, stat.S_IRWXU| stat.S_IRWXG| stat.S_IRWXO) # 0777 func(path) log.info("handleRemoveReadonly: fixed by chmod: %s", path) return except Exception: log.exception('handleRemoveReadonly error while trying to handle EACCES: ') if not os.path.exists(path): log.info('handleRemoveReadonly gone now') # surely this never happens? if it does, change the code below raise elif excvalue.errno == errno.ENOTEMPTY: # directory not empty, try again try: log.info("handleRemoveReadonly: ENOTEMPTY dir - has contents: %s", os.listdir(path)) except Exception, e: log.info("handleRemoveReadonly: ENOTEMPTY dir, could not get contents: %s"%e) if allowRetry: # avoid danger of infinite recursion if things are going really wrong deleteDir(path, allowRetry=False) log.info("handleRemoveReadonly: fixed by retrying rmtree for ENOTEMPTY: %s", path) return elif excvalue.errno == errno.ENOENT: # maybe windows went mad and deleted it anyway log.error("handleRemoveReadonly: ENOTENT error was raised by path that still exists: %s"%path) raise # if we didn't manage to handle this, rethrow log.warning("handleRemoveReadonly: still failed to remove path %s (%s %s); exists=%s" % (path,errno.errorcode.get(excvalue.errno, "EUNKNOWN"), func, os.path.exists(path))) raise path = normLongPath(path) if not os.path.exists(path): return try: shutil.rmtree(path, ignore_errors=False, onerror=handleRemoveReadonly) except OSError as e: if os.path.isfile(path): raise OSError, "Unable to delete dir %s as this is a file not a directory" % (path) if allowRetry: log.warn("Failed to delete dir %s (%s), will retry in 10 seconds" %(path, e)) # todo: remove these debug comments in time #handleslog = os.path.normpath('openhandles_%s.txt'%os.path.basename(path)) #with open(handleslog, 'w') as f: # handlecmd = [os.path.normpath('c:/dev/apama-lib2/win/all/sysinternals/handle.exe'), '-u', 'c:\\dev\\5.2.0.x\\apama-src', '/accepteula'] # #print 'running: ', ' '.join(handlecmd) # subprocess.call(args=handlecmd, stdout=f) # maybe it was a transient error, so try again a little later time.sleep(10.0) # on windows, try again using a separate process, just in case that # helps to avoid problems with virus checkers, etc if __isWindows: rmdirresult = os.system(u'rmdir /s /q "%s" 2>1 > /dev/nul'%path) log.info("Directory deletion retry using rmdir returned code %d: %s", rmdirresult, path) # continue to run deleteDir regardless of result, to check it's # really gone, and to give better error messages if we still # can't delete for any reason deleteDir(path, allowRetry=False) log.info("Deleted successfully on retry: %s", path) else: if os.path.exists(path): # maybe logging this is overkill, consider removing in future log.info("Unable to delete dir %s - original exception is: " % (path), exc_info=sys.exc_info()) raise OSError("Unable to delete dir %s: %s" % (path, e)) def deleteFile(path, allowRetry=True): """Delete the specified file, with the option of automatically retrying a few times if the first attempt fails (to get around Windows weirdness), throwing an exception if the file still exists at the end of retrying. Use this instead of os.remove for improved robustness. Does nothing if the file doesn't already exist. Contains magic hacks so it works even on paths that exceed the Windows MAX_PATH 260 character length. @param path: The path to delete. @param allowRetry: If true, wait for a bit and retry the removal if it fails (default: true) """ path = normLongPath(path) try: if not os.path.lexists(path): return # use lexists in case we're deleting a symlink try: os.remove(path) except Exception: if os.path.lexists(path): raise except OSError, e: if os.path.isdir(path): raise OSError, "Unable to delete file %s as this is a directory not a file" % (path) if allowRetry: log.debug("Failed to delete file %s on first attempt (%s), will retry in 5 seconds", path, e) # maybe it was a transient error, so try again a little later # on contended windows machines a 5 second sleep isn't always sufficient to prevent error 32 time.sleep(10.0) deleteFile(path, allowRetry=False) log.debug("Deleted file successfully on retry: %s", path) else: if os.path.lexists(path): if os.path.basename(path) in ('%s'%e): raise else: raise OSError("Unable to delete file %s: %s" % (path, e)) def parsePropertiesFile(lines, excludeLines=None): """ Parse the contents of the specified properties file or line list, and return an ordered list of (key,value,lineno) pairs. @param lines: an open file handle or a sequence that can be iterated over to get each line in the file. @param excludeLines: a string of list of strings to search for, any KEY containing these strings will be ignored >>> parsePropertiesFile(['a','b=c',' z = x', 'a=d #foo', '#g=h']) [('b', 'c', 2), ('z', 'x', 3), ('a', 'd', 4)] >>> parsePropertiesFile(['a=b','c=d#foo','XfooX=e', 'f=h'], excludeLines='foo') [('a', 'b', 1), ('c', 'd', 2), ('f', 'h', 4)] >>> parsePropertiesFile(['a=b','c=d#foo','XfooX=e', 'f=h'], excludeLines=['foo','h']) [('a', 'b', 1), ('c', 'd', 2), ('f', 'h', 4)] """ excludeLines = getStringList(excludeLines) result = [] lineNo = 0 for line in lines: lineNo += 1 if '#' in line: line = line[:line.find('#')].strip() line = line.strip() if not line or line.startswith('#') or line.startswith('//') or not '=' in line: continue key = line[:line.find('=')].strip() value = line[line.find('=')+1:].strip() if filter(lambda x: x in key, excludeLines): log.debug('Ignoring property line due to exclusion: %s', line) continue # NB: we don't have a full implementation of .properties escaping yet (e.g. \n but not \\n etc) # but this is all we need for now value = value.replace('\\\\','\\') result.append((key,value, lineNo)) return result if os.sep == '\\': def isDirPath(path): """ Returns true if the path is a directory (ends with / or \\). >>> isDirPath(None) False >>> isDirPath('/') True >>> isDirPath('a/') True >>> isDirPath('a'+os.sep) True """ try: return path[-1] in ('/', '\\') except Exception: return False else: def isDirPath(path): """ Returns true if the path is a directory (ends with / or \\). >>> isDirPath(None) False >>> isDirPath('/') True >>> isDirPath('a/') True >>> isDirPath('a'+os.sep) True """ try: return path[-1] == '/' except Exception: return False __longPathCache = {} # GIL protects integrity of dict, no need for extra locking as it's only a cache def toLongPathSafe(path, force=False): """ Converts the specified path string to a form suitable for passing to API calls if it exceeds the maximum path length on this OS. Currently, this is necessary only on Windows, where a unicode string starting with \\?\ must be used to get correct behaviour for long paths. Unlike L{normLongPath} which also performs the long path conversion, this function does NOT convert to a canonical form, normalize slashes or remove '..' elements (unless required for long path support). It is therefore faster. @param path: A path. Must not be a relative path. Can be None/empty. Can contain ".." sequences, though performance is a lot lower if it does. @param force: Normally the long path support is added only if this path exceeds the maximum length on this OS (e.g. 256 chars) or ends with a directory slash. Set force to True to add long path support regardless of length, which allows extra characters to be added on to the end of the string (e.g. ".log" or a directory filename) safely. @return: The passed-in path, possibly with a "\\?\" prefix added and forward slashes converted to backslashes on Windows. Any trailing slash is preserved by this function (though will be converted to a backslash). """ if (not __isWindows) or (not path): return path if (force or len(path)>255 or isDirPath(path)) and not path.startswith('\\\\?\\'): if path in __longPathCache: return __longPathCache[path] inputpath = path # ".." is not permitted in \\?\ paths; normpath is expensive so don't do this unless we have to if '.' in path: path = os.path.normpath(path)+('\\' if isDirPath(path) else '') else: # path is most likely to contain / so more efficient to conditionalize this path = path.replace('/','\\') if '\\\\' in path: # consecutive \ separators are not permitted in \\?\ paths path = path.replace('\\\\','\\') try: if path.startswith('\\\\'): path = u'\\\\?\\UNC\\'+path.lstrip('\\') # \\?\UNC\server\share Oh My else: path = u'\\\\?\\'+path except Exception: # can throw an exception if path is a bytestring containing non-ascii characters # to be safe, fallback to original string, just hoping it isn't both # international AND long # could try converting using a default encoding, but slightly error-prone pass __longPathCache[inputpath] = path return path __normLongPathCache = {} # GIL protects integrity of dict, no need for extra locking as it's only a cache def normLongPath(path): """ Normalizes and absolutizes a path (os.path.abspath), converts to a canonical form (e.g. normalizing the case of the drive letter on Windows), and on windows adds the "\\?\" prefix needed to force correct handling of long (>256 chars) paths (same as L{toLongPathSafe}). @param path: the absolute path to be converted should be a unicode string where possible, as specifying a byte string will not work if the path contains non-ascii characters. """ if not path: return path # profiling shows normLongPath is surprisingly costly; caching results reduces dep checking by 2-3x if path in __normLongPathCache: return __normLongPathCache[path] inputpath = path # currently there is some duplication between this and buildcommon.normpath which we ought to fix at some point # normpath does nothing to normalize case, and windows seems to be quite random about upper/lower case # for drive letters (more so than directory names), with different cmd prompts frequently using different # capitalization, so normalize at least that bit, to prevent spurious rebuilding from different prompts if __isWindows and len(path)>2 and path[1] == ':' and path[0] >= 'A' and path[0] <= 'Z': path = path[0].lower()+path[1:] if __isWindows and path and path.startswith('\\\\?\\'): path = path.replace('/', '\\') else: # abspath also normalizes slashes path = os.path.abspath(path)+(os.path.sep if isDirPath(path) else '') if __isWindows and path and not path.startswith('\\\\?\\'): try: if path.startswith('\\\\'): path = u'\\\\?\\UNC\\'+path.lstrip('\\') # \\?\UNC\server\share Oh My else: path = u'\\\\?\\'+path except Exception: # can throw an exception if path is a bytestring containing non-ascii characters # to be safe, fallback to original string, just hoping it isn't both # international AND long # could try converting using a default encoding, but slightly error-prone pass __normLongPathCache[inputpath] = path return path __statcache = {} __statcache_get = __statcache.get def getstat(path, errorIfMissing=False): """ Cached-once os.stat (DO NOT USE if you expect it to change after startup). Returns False if missing. """ st = __statcache_get(path, None) if st is None: try: st = os.stat(path) except os.error: # mean file doesn't exist st = False __statcache[path] = st if st is False and errorIfMissing: raise Exception('Cannot find path "%s"'%path) return st def getmtime(path): """ Cached-once os.getmtime (DO NOT USE if you expect it to change after startup) """ return getstat(path, errorIfMissing=True).st_mtime def getsize(path): """ Cached-once os.path.getsize (DO NOT USE if you expect it to change after startup) """ return getstat(path, errorIfMissing=True).st_size def exists(path): """ Cached-once os.path.exists (DO NOT USE if you expect it to change after startup) """ return getstat(path) is not False def isfile(path): """ Cached-once os.path.isfile (DO NOT USE if you expect it to change after startup) """ st = getstat(path) return (st is not False) and stat.S_ISREG(st.st_mode) def isdir(path): """ Cached-once os.path.isdir (DO NOT USE if you expect it to change after startup) """ st = getstat(path) return (st is not False) and stat.S_ISDIR(st.st_mode) def _getStatCacheSize(): """ Internal diagnostic method for getting the number of entries we've stat'ed so far. """ return len(__statcache) def resetStatCache(): """ Resets cached stat data """ __statcache.clear()
37.341253
269
0.695876
""" @undocumented: _getStatCacheSize """ import shutil, os, os.path, time, platform, threading import stat, sys from utils.flatten import getStringList import subprocess, errno import logging log = logging.getLogger('fileutils') __isWindows = platform.system()=='Windows' if __isWindows: try: import win32file class openForWrite: def __init__(self, dest, mode): assert mode == 'wb' self.dest = dest def __enter__(self): self.Fd = win32file.CreateFile(self.dest, win32file.GENERIC_WRITE, win32file.FILE_SHARE_READ | win32file.FILE_SHARE_WRITE | win32file.FILE_SHARE_DELETE, None, win32file.CREATE_ALWAYS, win32file.FILE_ATTRIBUTE_NORMAL, None) return self def write(self, string): if isinstance(string, unicode): string = string.encode('ascii', errors='strict') assert isinstance(string, str), 'cannot write unicode character string to binary file; please use byte str instead: %s'%repr(string) win32file.WriteFile(self.Fd, string) def writelines(self, lines): for l in lines: self.write(l) def close(self): win32file.CloseHandle(self.Fd) def __exit__(self, ex_type, ex_val, tb): self.close() except: openForWrite = open else: openForWrite = open def mkdir(newdir): """ Recursively create the specified directory if it doesn't already exist. If it does, exit without error. @param newdir: The path to create. @return: newdir, to allow fluent use of this method. """ origdir = newdir newdir=normLongPath(newdir) if os.path.isdir(newdir): # already exists return origdir if os.path.isfile(newdir): raise IOError("A file with the same name as the desired dir, '%s', already exists" % newdir) #when multiple threads/processes are creating directories #at the same time, it can be a race try: os.makedirs(newdir) except Exception, e: if os.path.isdir(newdir): pass else: raise IOError('Problem creating directory %s: %s' % (newdir, e)) return origdir def deleteDir(path, allowRetry=True): """ Recursively delete the contents of a directory. Contains magic hacks so it works even on paths that exceed the Windows MAX_PATH 260 character length. @param path: the path to delete. @param allowRetry: set to False to disable automatic retry of the deletion after a few seconds (in case the error was transient) """ def handleRemoveReadonly(func, path, exc): # once we've got this working reliably, might reduce the level of some of these log statements excvalue = exc[1] log.info("handleRemoveReadonly: error removing path %s (%s %s), will try harder; exists=%s" % (path,errno.errorcode.get(excvalue.errno, "EUNKNOWN"), func, os.path.exists(path))) if func in (os.rmdir, os.remove): if not os.path.exists(path): log.info("handleRemoveReadonly: suppressing spurious remove exception for already-deleted path: %s", path) return if excvalue.errno == errno.EACCES: try: os.chmod(path, stat.S_IRWXU| stat.S_IRWXG| stat.S_IRWXO) func(path) log.info("handleRemoveReadonly: fixed by chmod: %s", path) return except Exception: log.exception('handleRemoveReadonly error while trying to handle EACCES: ') if not os.path.exists(path): log.info('handleRemoveReadonly gone now') raise elif excvalue.errno == errno.ENOTEMPTY: try: log.info("handleRemoveReadonly: ENOTEMPTY dir - has contents: %s", os.listdir(path)) except Exception, e: log.info("handleRemoveReadonly: ENOTEMPTY dir, could not get contents: %s"%e) if allowRetry: deleteDir(path, allowRetry=False) log.info("handleRemoveReadonly: fixed by retrying rmtree for ENOTEMPTY: %s", path) return elif excvalue.errno == errno.ENOENT: log.error("handleRemoveReadonly: ENOTENT error was raised by path that still exists: %s"%path) raise log.warning("handleRemoveReadonly: still failed to remove path %s (%s %s); exists=%s" % (path,errno.errorcode.get(excvalue.errno, "EUNKNOWN"), func, os.path.exists(path))) raise path = normLongPath(path) if not os.path.exists(path): return try: shutil.rmtree(path, ignore_errors=False, onerror=handleRemoveReadonly) except OSError as e: if os.path.isfile(path): raise OSError, "Unable to delete dir %s as this is a file not a directory" % (path) if allowRetry: log.warn("Failed to delete dir %s (%s), will retry in 10 seconds" %(path, e)) # todo: remove these debug comments in time #handleslog = os.path.normpath('openhandles_%s.txt'%os.path.basename(path)) #with open(handleslog, 'w') as f: # handlecmd = [os.path.normpath('c:/dev/apama-lib2/win/all/sysinternals/handle.exe'), '-u', 'c:\\dev\\5.2.0.x\\apama-src', '/accepteula'] # #print 'running: ', ' '.join(handlecmd) # subprocess.call(args=handlecmd, stdout=f) # maybe it was a transient error, so try again a little later time.sleep(10.0) # on windows, try again using a separate process, just in case that # helps to avoid problems with virus checkers, etc if __isWindows: rmdirresult = os.system(u'rmdir /s /q "%s" 2>1 > /dev/nul'%path) log.info("Directory deletion retry using rmdir returned code %d: %s", rmdirresult, path) # continue to run deleteDir regardless of result, to check it's deleteDir(path, allowRetry=False) log.info("Deleted successfully on retry: %s", path) else: if os.path.exists(path): # maybe logging this is overkill, consider removing in future log.info("Unable to delete dir %s - original exception is: " % (path), exc_info=sys.exc_info()) raise OSError("Unable to delete dir %s: %s" % (path, e)) def deleteFile(path, allowRetry=True): """Delete the specified file, with the option of automatically retrying a few times if the first attempt fails (to get around Windows weirdness), throwing an exception if the file still exists at the end of retrying. Use this instead of os.remove for improved robustness. Does nothing if the file doesn't already exist. Contains magic hacks so it works even on paths that exceed the Windows MAX_PATH 260 character length. @param path: The path to delete. @param allowRetry: If true, wait for a bit and retry the removal if it fails (default: true) """ path = normLongPath(path) try: if not os.path.lexists(path): return try: os.remove(path) except Exception: if os.path.lexists(path): raise except OSError, e: if os.path.isdir(path): raise OSError, "Unable to delete file %s as this is a directory not a file" % (path) if allowRetry: log.debug("Failed to delete file %s on first attempt (%s), will retry in 5 seconds", path, e) # maybe it was a transient error, so try again a little later # on contended windows machines a 5 second sleep isn't always sufficient to prevent error 32 time.sleep(10.0) deleteFile(path, allowRetry=False) log.debug("Deleted file successfully on retry: %s", path) else: if os.path.lexists(path): if os.path.basename(path) in ('%s'%e): raise else: raise OSError("Unable to delete file %s: %s" % (path, e)) def parsePropertiesFile(lines, excludeLines=None): """ Parse the contents of the specified properties file or line list, and return an ordered list of (key,value,lineno) pairs. @param lines: an open file handle or a sequence that can be iterated over to get each line in the file. @param excludeLines: a string of list of strings to search for, any KEY containing these strings will be ignored >>> parsePropertiesFile(['a','b=c',' z = x', 'a=d #foo', '#g=h']) [('b', 'c', 2), ('z', 'x', 3), ('a', 'd', 4)] >>> parsePropertiesFile(['a=b','c=d#foo','XfooX=e', 'f=h'], excludeLines='foo') [('a', 'b', 1), ('c', 'd', 2), ('f', 'h', 4)] >>> parsePropertiesFile(['a=b','c=d#foo','XfooX=e', 'f=h'], excludeLines=['foo','h']) [('a', 'b', 1), ('c', 'd', 2), ('f', 'h', 4)] """ excludeLines = getStringList(excludeLines) result = [] lineNo = 0 for line in lines: lineNo += 1 if '#' in line: line = line[:line.find('#')].strip() line = line.strip() if not line or line.startswith('#') or line.startswith('//') or not '=' in line: continue key = line[:line.find('=')].strip() value = line[line.find('=')+1:].strip() if filter(lambda x: x in key, excludeLines): log.debug('Ignoring property line due to exclusion: %s', line) continue # but this is all we need for now value = value.replace('\\\\','\\') result.append((key,value, lineNo)) return result if os.sep == '\\': def isDirPath(path): """ Returns true if the path is a directory (ends with / or \\). >>> isDirPath(None) False >>> isDirPath('/') True >>> isDirPath('a/') True >>> isDirPath('a'+os.sep) True """ try: return path[-1] in ('/', '\\') except Exception: return False else: def isDirPath(path): """ Returns true if the path is a directory (ends with / or \\). >>> isDirPath(None) False >>> isDirPath('/') True >>> isDirPath('a/') True >>> isDirPath('a'+os.sep) True """ try: return path[-1] == '/' except Exception: return False __longPathCache = {} # GIL protects integrity of dict, no need for extra locking as it's only a cache def toLongPathSafe(path, force=False): """ Converts the specified path string to a form suitable for passing to API calls if it exceeds the maximum path length on this OS. Currently, this is necessary only on Windows, where a unicode string starting with \\?\ must be used to get correct behaviour for long paths. Unlike L{normLongPath} which also performs the long path conversion, this function does NOT convert to a canonical form, normalize slashes or remove '..' elements (unless required for long path support). It is therefore faster. @param path: A path. Must not be a relative path. Can be None/empty. Can contain ".." sequences, though performance is a lot lower if it does. @param force: Normally the long path support is added only if this path exceeds the maximum length on this OS (e.g. 256 chars) or ends with a directory slash. Set force to True to add long path support regardless of length, which allows extra characters to be added on to the end of the string (e.g. ".log" or a directory filename) safely. @return: The passed-in path, possibly with a "\\?\" prefix added and forward slashes converted to backslashes on Windows. Any trailing slash is preserved by this function (though will be converted to a backslash). """ if (not __isWindows) or (not path): return path if (force or len(path)>255 or isDirPath(path)) and not path.startswith('\\\\?\\'): if path in __longPathCache: return __longPathCache[path] inputpath = path if '.' in path: path = os.path.normpath(path)+('\\' if isDirPath(path) else '') else: # path is most likely to contain / so more efficient to conditionalize this path = path.replace('/','\\') if '\\\\' in path: # consecutive \ separators are not permitted in \\?\ paths path = path.replace('\\\\','\\') try: if path.startswith('\\\\'): path = u'\\\\?\\UNC\\'+path.lstrip('\\') # \\?\UNC\server\share Oh My else: path = u'\\\\?\\'+path except Exception: # can throw an exception if path is a bytestring containing non-ascii characters # to be safe, fallback to original string, just hoping it isn't both pass __longPathCache[inputpath] = path return path __normLongPathCache = {} def normLongPath(path): """ Normalizes and absolutizes a path (os.path.abspath), converts to a canonical form (e.g. normalizing the case of the drive letter on Windows), and on windows adds the "\\?\" prefix needed to force correct handling of long (>256 chars) paths (same as L{toLongPathSafe}). @param path: the absolute path to be converted should be a unicode string where possible, as specifying a byte string will not work if the path contains non-ascii characters. """ if not path: return path # profiling shows normLongPath is surprisingly costly; caching results reduces dep checking by 2-3x if path in __normLongPathCache: return __normLongPathCache[path] inputpath = path # currently there is some duplication between this and buildcommon.normpath which we ought to fix at some point # normpath does nothing to normalize case, and windows seems to be quite random about upper/lower case # for drive letters (more so than directory names), with different cmd prompts frequently using different # capitalization, so normalize at least that bit, to prevent spurious rebuilding from different prompts if __isWindows and len(path)>2 and path[1] == ':' and path[0] >= 'A' and path[0] <= 'Z': path = path[0].lower()+path[1:] if __isWindows and path and path.startswith('\\\\?\\'): path = path.replace('/', '\\') else: # abspath also normalizes slashes path = os.path.abspath(path)+(os.path.sep if isDirPath(path) else '') if __isWindows and path and not path.startswith('\\\\?\\'): try: if path.startswith('\\\\'): path = u'\\\\?\\UNC\\'+path.lstrip('\\') # \\?\UNC\server\share Oh My else: path = u'\\\\?\\'+path except Exception: # can throw an exception if path is a bytestring containing non-ascii characters # to be safe, fallback to original string, just hoping it isn't both pass __normLongPathCache[inputpath] = path return path __statcache = {} __statcache_get = __statcache.get def getstat(path, errorIfMissing=False): """ Cached-once os.stat (DO NOT USE if you expect it to change after startup). Returns False if missing. """ st = __statcache_get(path, None) if st is None: try: st = os.stat(path) except os.error: st = False __statcache[path] = st if st is False and errorIfMissing: raise Exception('Cannot find path "%s"'%path) return st def getmtime(path): """ Cached-once os.getmtime (DO NOT USE if you expect it to change after startup) """ return getstat(path, errorIfMissing=True).st_mtime def getsize(path): """ Cached-once os.path.getsize (DO NOT USE if you expect it to change after startup) """ return getstat(path, errorIfMissing=True).st_size def exists(path): """ Cached-once os.path.exists (DO NOT USE if you expect it to change after startup) """ return getstat(path) is not False def isfile(path): """ Cached-once os.path.isfile (DO NOT USE if you expect it to change after startup) """ st = getstat(path) return (st is not False) and stat.S_ISREG(st.st_mode) def isdir(path): """ Cached-once os.path.isdir (DO NOT USE if you expect it to change after startup) """ st = getstat(path) return (st is not False) and stat.S_ISDIR(st.st_mode) def _getStatCacheSize(): """ Internal diagnostic method for getting the number of entries we've stat'ed so far. """ return len(__statcache) def resetStatCache(): """ Resets cached stat data """ __statcache.clear()
false
true
1c4856012f6f123905c3fbdfe922f2752e1467ef
3,974
py
Python
ProjectEuler/Problems_051_100/P070_TotientPermutation.py
mqq-marek/ProjectEuler
3a865b32a655c5ba39bf58a4cb96cef0ffeccbbd
[ "MIT" ]
null
null
null
ProjectEuler/Problems_051_100/P070_TotientPermutation.py
mqq-marek/ProjectEuler
3a865b32a655c5ba39bf58a4cb96cef0ffeccbbd
[ "MIT" ]
null
null
null
ProjectEuler/Problems_051_100/P070_TotientPermutation.py
mqq-marek/ProjectEuler
3a865b32a655c5ba39bf58a4cb96cef0ffeccbbd
[ "MIT" ]
null
null
null
""" Euler's Totient function, phi(n), is used to determine the number of positive numbers less than or equal to n which are relatively prime to n. For example, as 1, 2, 4, 5, 7, and 8, are all less than nine and relatively prime to nine, phi(9)=6. The number 1 is considered to be relatively prime to every positive number, so phi(1)=1. Interestingly, phi(87109)=79180, and it can be seen that 87109 is a permutation of 79180. Find the value of n, 1 < n < 10**7, for which phi(n) is a permutation of n and the ratio n/phi(n) produces a minimum. """ import time from bisect import bisect_right from fractions import Fraction import math from collections import Counter from typing import Iterator def eratosthenes_sieve(n): """Return primes <= n.""" def add_prime(k): """Add founded prime.""" p = k + k + 3 primes.append(p) pos = k + p while pos <= n: numbers[pos] = 1 pos += p numbers = [0] * (n + 1) primes = [2] for i in range(n): if not numbers[i]: add_prime(i) return primes primes = eratosthenes_sieve(10 ** 4) def prime_divisors(num: int) -> Iterator[int]: """ Get all num prime divisors. :param num: number for which we yields prime divisors :yields: num prime divisors """ assert num > 0 start_num = num sqrt_num = int(math.sqrt(num)) + 1 counter = 0 for p in primes: while num % p == 0: yield p counter += 1 num //= p if num == 1 or counter > 3: return if p > sqrt_num: yield num return raise Exception(f"Primes too short for {start_num} -> {num}, Primes{len(primes)}/{primes[-1]}") def totient(n): """ Compute totient. totient(prime**k) = p**k - p**(k-1) totient(n*m) = totient(n) * totient(m) if n,m coprime. """ result = list(prime_divisors(n)) prime_power = Counter(result) res = 1 for p, cnt in prime_power.items(): if cnt == 1: res *= (p - 1) else: res = pow(p, cnt - 1) * (p - 1) return res, result def totient_loop_naive(n): index = 0 ratio = Fraction(2, 1) start = time.perf_counter() cnt = 0 for i in range(17, n): tt = totient(i) t = tt[0] if sorted(str(i)) == sorted(str(t)): new_ratio = Fraction(i, t) if new_ratio < ratio: # print(cnt, i, i - index, t, tt[1], round(float(new_ratio),3), round(time.perf_counter() - start)) cnt += 1 start = time.perf_counter() ratio = new_ratio index = i return index def totient_guesser(n): ratio = n index = 0 n_sqrt = int(math.sqrt(n)) + 1 start = primes.index(149) end = bisect_right(primes, n_sqrt) + 1 for i1 in range(start, end): p1 = primes[i1] for i2 in range(i1, len(primes)): p2 = primes[i2] p12 = p1 * p2 if p12 >= n: break phi = (p1-1) * (p2 - 1) new_ratio = p12 / phi if new_ratio < ratio and sorted(str(p12)) == sorted(str(phi)): # print(p1, p2, new_ratio, p12) ratio = new_ratio index = p12 for i3 in range(i2, len(primes)): p3 = primes[i3] p123 = p12 * p3 if p123 >= n: break phi *= (p3 - 1) new_ratio = p123 / phi if new_ratio < ratio and sorted(str(p123)) == sorted(str(phi)): # print(p1, p2, new_ratio, p12) ratio = new_ratio index = p123 return index def totient_solver(n): if n <= 76000: return totient_loop_naive(n) return totient_guesser(n) print(primes) #n = 10**7 n = int(input()) print(totient_solver(n))
26.671141
117
0.530448
import time from bisect import bisect_right from fractions import Fraction import math from collections import Counter from typing import Iterator def eratosthenes_sieve(n): def add_prime(k): p = k + k + 3 primes.append(p) pos = k + p while pos <= n: numbers[pos] = 1 pos += p numbers = [0] * (n + 1) primes = [2] for i in range(n): if not numbers[i]: add_prime(i) return primes primes = eratosthenes_sieve(10 ** 4) def prime_divisors(num: int) -> Iterator[int]: assert num > 0 start_num = num sqrt_num = int(math.sqrt(num)) + 1 counter = 0 for p in primes: while num % p == 0: yield p counter += 1 num //= p if num == 1 or counter > 3: return if p > sqrt_num: yield num return raise Exception(f"Primes too short for {start_num} -> {num}, Primes{len(primes)}/{primes[-1]}") def totient(n): result = list(prime_divisors(n)) prime_power = Counter(result) res = 1 for p, cnt in prime_power.items(): if cnt == 1: res *= (p - 1) else: res = pow(p, cnt - 1) * (p - 1) return res, result def totient_loop_naive(n): index = 0 ratio = Fraction(2, 1) start = time.perf_counter() cnt = 0 for i in range(17, n): tt = totient(i) t = tt[0] if sorted(str(i)) == sorted(str(t)): new_ratio = Fraction(i, t) if new_ratio < ratio: cnt += 1 start = time.perf_counter() ratio = new_ratio index = i return index def totient_guesser(n): ratio = n index = 0 n_sqrt = int(math.sqrt(n)) + 1 start = primes.index(149) end = bisect_right(primes, n_sqrt) + 1 for i1 in range(start, end): p1 = primes[i1] for i2 in range(i1, len(primes)): p2 = primes[i2] p12 = p1 * p2 if p12 >= n: break phi = (p1-1) * (p2 - 1) new_ratio = p12 / phi if new_ratio < ratio and sorted(str(p12)) == sorted(str(phi)): ratio = new_ratio index = p12 for i3 in range(i2, len(primes)): p3 = primes[i3] p123 = p12 * p3 if p123 >= n: break phi *= (p3 - 1) new_ratio = p123 / phi if new_ratio < ratio and sorted(str(p123)) == sorted(str(phi)): ratio = new_ratio index = p123 return index def totient_solver(n): if n <= 76000: return totient_loop_naive(n) return totient_guesser(n) print(primes) n = int(input()) print(totient_solver(n))
true
true
1c4857b3876c7f9d4b85021c0fe07b0789cd8808
842
py
Python
sls_api/scripts/reset_user_projects.py
slsfi/sls_gde_api
68c6342cc3af95d9cf5b87cf096fc03b7fd5e67d
[ "Apache-2.0" ]
null
null
null
sls_api/scripts/reset_user_projects.py
slsfi/sls_gde_api
68c6342cc3af95d9cf5b87cf096fc03b7fd5e67d
[ "Apache-2.0" ]
null
null
null
sls_api/scripts/reset_user_projects.py
slsfi/sls_gde_api
68c6342cc3af95d9cf5b87cf096fc03b7fd5e67d
[ "Apache-2.0" ]
null
null
null
import argparse import sys from sls_api.models import User from sls_api import app if __name__ == "__main__": with app.app_context(): parser = argparse.ArgumentParser(description="Helper script to reset a Users projects") parser.add_argument("email", help="User email address") parser.add_argument("projects", help="User projects") args = parser.parse_args() success = User.reset_projects(args.email, args.projects) if success is None: print("Error during projects reset! Check API backend logs.") sys.exit(1) elif success: print(f"Projects for user {args.email} successfully changed to {args.projects}!") sys.exit(0) else: print(f"No user with the email {args.email} could be found!") sys.exit(1)
33.68
95
0.640143
import argparse import sys from sls_api.models import User from sls_api import app if __name__ == "__main__": with app.app_context(): parser = argparse.ArgumentParser(description="Helper script to reset a Users projects") parser.add_argument("email", help="User email address") parser.add_argument("projects", help="User projects") args = parser.parse_args() success = User.reset_projects(args.email, args.projects) if success is None: print("Error during projects reset! Check API backend logs.") sys.exit(1) elif success: print(f"Projects for user {args.email} successfully changed to {args.projects}!") sys.exit(0) else: print(f"No user with the email {args.email} could be found!") sys.exit(1)
true
true
1c48585dd10eb78791eb0b974d23a9c5313b1493
3,340
py
Python
producerapril19/boto3/__init__.py
drwitt/AWS-lambda-NLP-project-4
a1cdcaee5cb8679bb86a25811e8323abd40fffcf
[ "Apache-2.0" ]
null
null
null
producerapril19/boto3/__init__.py
drwitt/AWS-lambda-NLP-project-4
a1cdcaee5cb8679bb86a25811e8323abd40fffcf
[ "Apache-2.0" ]
9
2021-03-19T03:06:53.000Z
2022-03-12T00:37:04.000Z
myvenv/lib/python3.6/site-packages/boto3/__init__.py
yog240597/saleor
b75a23827a4ec2ce91637f0afe6808c9d09da00a
[ "CC-BY-4.0" ]
1
2021-04-06T15:08:09.000Z
2021-04-06T15:08:09.000Z
# Copyright 2014 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"). You # may not use this file except in compliance with the License. A copy of # the License is located at # # http://aws.amazon.com/apache2.0/ # # or in the "license" file accompanying this file. This file is # distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF # ANY KIND, either express or implied. See the License for the specific # language governing permissions and limitations under the License. import logging from boto3.session import Session __author__ = 'Amazon Web Services' __version__ = '1.12.42' # The default Boto3 session; autoloaded when needed. DEFAULT_SESSION = None def setup_default_session(**kwargs): """ Set up a default session, passing through any parameters to the session constructor. There is no need to call this unless you wish to pass custom parameters, because a default session will be created for you. """ global DEFAULT_SESSION DEFAULT_SESSION = Session(**kwargs) def set_stream_logger(name='boto3', level=logging.DEBUG, format_string=None): """ Add a stream handler for the given name and level to the logging module. By default, this logs all boto3 messages to ``stdout``. >>> import boto3 >>> boto3.set_stream_logger('boto3.resources', logging.INFO) For debugging purposes a good choice is to set the stream logger to ``''`` which is equivalent to saying "log everything". .. WARNING:: Be aware that when logging anything from ``'botocore'`` the full wire trace will appear in your logs. If your payloads contain sensitive data this should not be used in production. :type name: string :param name: Log name :type level: int :param level: Logging level, e.g. ``logging.INFO`` :type format_string: str :param format_string: Log message format """ if format_string is None: format_string = "%(asctime)s %(name)s [%(levelname)s] %(message)s" logger = logging.getLogger(name) logger.setLevel(level) handler = logging.StreamHandler() handler.setLevel(level) formatter = logging.Formatter(format_string) handler.setFormatter(formatter) logger.addHandler(handler) def _get_default_session(): """ Get the default session, creating one if needed. :rtype: :py:class:`~boto3.session.Session` :return: The default session """ if DEFAULT_SESSION is None: setup_default_session() return DEFAULT_SESSION def client(*args, **kwargs): """ Create a low-level service client by name using the default session. See :py:meth:`boto3.session.Session.client`. """ return _get_default_session().client(*args, **kwargs) def resource(*args, **kwargs): """ Create a resource service client by name using the default session. See :py:meth:`boto3.session.Session.resource`. """ return _get_default_session().resource(*args, **kwargs) # Set up logging to ``/dev/null`` like a library is supposed to. # http://docs.python.org/3.3/howto/logging.html#configuring-logging-for-a-library class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('boto3').addHandler(NullHandler())
30.09009
81
0.703593
import logging from boto3.session import Session __author__ = 'Amazon Web Services' __version__ = '1.12.42' DEFAULT_SESSION = None def setup_default_session(**kwargs): global DEFAULT_SESSION DEFAULT_SESSION = Session(**kwargs) def set_stream_logger(name='boto3', level=logging.DEBUG, format_string=None): if format_string is None: format_string = "%(asctime)s %(name)s [%(levelname)s] %(message)s" logger = logging.getLogger(name) logger.setLevel(level) handler = logging.StreamHandler() handler.setLevel(level) formatter = logging.Formatter(format_string) handler.setFormatter(formatter) logger.addHandler(handler) def _get_default_session(): if DEFAULT_SESSION is None: setup_default_session() return DEFAULT_SESSION def client(*args, **kwargs): return _get_default_session().client(*args, **kwargs) def resource(*args, **kwargs): return _get_default_session().resource(*args, **kwargs) ): def emit(self, record): pass logging.getLogger('boto3').addHandler(NullHandler())
true
true
1c48586d5ac96b2e1374dcfec7c9d9e588473c51
3,235
py
Python
shcomplete/shell_scraper/settings.py
gy741/shell-complete
20ad82eb45015a79afc734f4cce2201b5fba3785
[ "Apache-2.0" ]
null
null
null
shcomplete/shell_scraper/settings.py
gy741/shell-complete
20ad82eb45015a79afc734f4cce2201b5fba3785
[ "Apache-2.0" ]
null
null
null
shcomplete/shell_scraper/settings.py
gy741/shell-complete
20ad82eb45015a79afc734f4cce2201b5fba3785
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- # Scrapy settings for shell_scraper project # # For simplicity, this file contains only settings considered important or # commonly used. You can find more settings consulting the documentation: # # http://doc.scrapy.org/en/latest/topics/settings.html # http://scrapy.readthedocs.org/en/latest/topics/downloader-middleware.html # http://scrapy.readthedocs.org/en/latest/topics/spider-middleware.html BOT_NAME = 'shell_scraper' SPIDER_MODULES = ['shell_scraper.spiders'] NEWSPIDER_MODULE = 'shell_scraper.spiders' # Crawl responsibly by identifying yourself # (and your website) on the user-agent # USER_AGENT = 'shell_scraper (+http://www.yourdomain.com)' # Obey robots.txt rules ROBOTSTXT_OBEY = False # Configure maximum concurrent requests performed by Scrapy (default: 16) # CONCURRENT_REQUESTS = 32 # Configure a delay for requests for the same website (default: 0) # See http://scrapy.readthedocs.org/en/latest # /topics/settings.html#download-delay # See also autothrottle settings and docs # DOWNLOAD_DELAY = 3 # The download delay setting will honor only one of: # CONCURRENT_REQUESTS_PER_DOMAIN = 16 # CONCURRENT_REQUESTS_PER_IP = 16 # Disable cookies (enabled by default) # COOKIES_ENABLED = False # Disable Telnet Console (enabled by default) # TELNETCONSOLE_ENABLED = False # Override the default request headers: # DEFAULT_REQUEST_HEADERS = { # 'Accept': 'text/html,application/xhtml+xml, # application/xml;q=0.9,*/*;q=0.8', # 'Accept-Language': 'en', # } # Enable or disable spider middlewares # See http://scrapy.readthedocs.org/en/latest/topics/spider-middleware.html # SPIDER_MIDDLEWARES = { # 'shell_scraper.middlewares.ShellScraperSpiderMiddleware': 543, # } # Enable or disable downloader middlewares # See http://scrapy.readthedocs.org/en/latest/topics/downloader-middleware.html # DOWNLOADER_MIDDLEWARES = { # 'shell_scraper.middlewares.MyCustomDownloaderMiddleware': 543, # } # Enable or disable extensions # See http://scrapy.readthedocs.org/en/latest/topics/extensions.html # EXTENSIONS = { # 'scrapy.extensions.telnet.TelnetConsole': None, # } # Configure item pipelines # See http://scrapy.readthedocs.org/en/latest/topics/item-pipeline.html # ITEM_PIPELINES = { # 'shell_scraper.pipelines.ShellScraperPipeline': 300, # } # Enable and configure the AutoThrottle extension (disabled by default) # See http://doc.scrapy.org/en/latest/topics/autothrottle.html # AUTOTHROTTLE_ENABLED = True # The initial download delay # AUTOTHROTTLE_START_DELAY = 5 # The maximum download delay to be set in case of high latencies # AUTOTHROTTLE_MAX_DELAY = 60 # The average number of requests Scrapy should be sending in parallel to # each remote server # AUTOTHROTTLE_TARGET_CONCURRENCY = 1.0 # Enable showing throttling stats for every response received: # AUTOTHROTTLE_DEBUG = False # Enable and configure HTTP caching (disabled by default) # See http://scrapy.readthedocs.org/en/latest/topics # /downloader-middleware.html#httpcache-middleware-settings # HTTPCACHE_ENABLED = True # HTTPCACHE_EXPIRATION_SECS = 0 # HTTPCACHE_DIR = 'httpcache' # HTTPCACHE_IGNORE_HTTP_CODES = [] # HTTPCACHE_STORAGE = 'scrapy.extensions.httpcache.FilesystemCacheStorage'
34.052632
79
0.772179
BOT_NAME = 'shell_scraper' SPIDER_MODULES = ['shell_scraper.spiders'] NEWSPIDER_MODULE = 'shell_scraper.spiders' ROBOTSTXT_OBEY = False # application/xml;q=0.9,*/*;q=0.8',
true
true
1c485b47804f4492c2b8ffc80ef6f14d8497cc7c
1,004
py
Python
tests/test_config_load.py
voidpp/magrathea-python-tools
0fc7460c827b02d8914411cedddadc23ccb3cc73
[ "MIT" ]
null
null
null
tests/test_config_load.py
voidpp/magrathea-python-tools
0fc7460c827b02d8914411cedddadc23ccb3cc73
[ "MIT" ]
null
null
null
tests/test_config_load.py
voidpp/magrathea-python-tools
0fc7460c827b02d8914411cedddadc23ccb3cc73
[ "MIT" ]
null
null
null
import pytest from voidpp_tools.mocks.file_system import mockfs from voidpp_tools.json_config import JSONConfigLoader from voidpp_tools.config_loader import ConfigFileNotFoundException @mockfs(dict(etc = {'app1.json': u'{"the_answer": 42}'})) def test_load_config_from_etc(): # Arrange loader = JSONConfigLoader('') # Act data = loader.load("app1.json") # Assert assert data == dict(the_answer = 42) @mockfs() def test_load_config_file_not_found(): # Arrange loader = JSONConfigLoader('') # Act & Assert with pytest.raises(ConfigFileNotFoundException): loader.load("app1.json") @mockfs(dict( etc = {'app1.json': u'{"the_answer": 42}'}, home = dict(douglas = {'app1.json': u'{"the_question": "6*7"}'}) ), user = 'douglas') def test_load_config_nested(): # Arrange loader = JSONConfigLoader('', nested = True) # Act data = loader.load("app1.json") # Assert assert data == dict(the_answer = 42, the_question = "6*7")
24.487805
68
0.670319
import pytest from voidpp_tools.mocks.file_system import mockfs from voidpp_tools.json_config import JSONConfigLoader from voidpp_tools.config_loader import ConfigFileNotFoundException @mockfs(dict(etc = {'app1.json': u'{"the_answer": 42}'})) def test_load_config_from_etc(): loader = JSONConfigLoader('') data = loader.load("app1.json") assert data == dict(the_answer = 42) @mockfs() def test_load_config_file_not_found(): loader = JSONConfigLoader('') with pytest.raises(ConfigFileNotFoundException): loader.load("app1.json") @mockfs(dict( etc = {'app1.json': u'{"the_answer": 42}'}, home = dict(douglas = {'app1.json': u'{"the_question": "6*7"}'}) ), user = 'douglas') def test_load_config_nested(): loader = JSONConfigLoader('', nested = True) data = loader.load("app1.json") assert data == dict(the_answer = 42, the_question = "6*7")
true
true
1c485b6e79e47b54e18602864ca41cb848d6dcf1
477
py
Python
Python/Algorithms/1002.py
DimitrisJim/leetcode_solutions
765ea578748f8c9b21243dec9dc8a16163e85c0c
[ "Unlicense" ]
2
2021-01-15T17:22:54.000Z
2021-05-16T19:58:02.000Z
Python/Algorithms/1002.py
DimitrisJim/leetcode_solutions
765ea578748f8c9b21243dec9dc8a16163e85c0c
[ "Unlicense" ]
null
null
null
Python/Algorithms/1002.py
DimitrisJim/leetcode_solutions
765ea578748f8c9b21243dec9dc8a16163e85c0c
[ "Unlicense" ]
null
null
null
from collections import Counter class Solution: # 40 - 92.66, 14.3 - 54.19 def commonChars(self, A): # Build counter of characters Counter_ = Counter commons = Counter_(A[0]) for i in range(1, len(A)): # In-place intersection of minimum of elements. commons &= Counter_(A[i]) # bail whenever it becomes empty. if not commons: return [] return commons.elements()
28.058824
60
0.555556
from collections import Counter class Solution: def commonChars(self, A): Counter_ = Counter commons = Counter_(A[0]) for i in range(1, len(A)): commons &= Counter_(A[i]) if not commons: return [] return commons.elements()
true
true
1c485cf9254aebb6da185f4c1cd864b675899f50
3,250
py
Python
sdk/applicationinsights/azure-mgmt-applicationinsights/azure/mgmt/applicationinsights/v2015_05_01/aio/_configuration.py
rsdoherty/azure-sdk-for-python
6bba5326677468e6660845a703686327178bb7b1
[ "MIT" ]
207
2017-11-29T06:59:41.000Z
2022-03-31T10:00:53.000Z
sdk/applicationinsights/azure-mgmt-applicationinsights/azure/mgmt/applicationinsights/v2015_05_01/aio/_configuration.py
rsdoherty/azure-sdk-for-python
6bba5326677468e6660845a703686327178bb7b1
[ "MIT" ]
4,061
2017-10-27T23:19:56.000Z
2022-03-31T23:18:30.000Z
sdk/applicationinsights/azure-mgmt-applicationinsights/azure/mgmt/applicationinsights/v2015_05_01/aio/_configuration.py
rsdoherty/azure-sdk-for-python
6bba5326677468e6660845a703686327178bb7b1
[ "MIT" ]
802
2017-10-11T17:36:26.000Z
2022-03-31T22:24:32.000Z
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- from typing import Any, TYPE_CHECKING from azure.core.configuration import Configuration from azure.core.pipeline import policies from azure.mgmt.core.policies import ARMHttpLoggingPolicy if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials_async import AsyncTokenCredential VERSION = "unknown" class ApplicationInsightsManagementClientConfiguration(Configuration): """Configuration for ApplicationInsightsManagementClient. Note that all parameters used to create this instance are saved as instance attributes. :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: The ID of the target subscription. :type subscription_id: str """ def __init__( self, credential: "AsyncTokenCredential", subscription_id: str, **kwargs: Any ) -> None: if credential is None: raise ValueError("Parameter 'credential' must not be None.") if subscription_id is None: raise ValueError("Parameter 'subscription_id' must not be None.") super(ApplicationInsightsManagementClientConfiguration, self).__init__(**kwargs) self.credential = credential self.subscription_id = subscription_id self.api_version = "2015-05-01" self.credential_scopes = kwargs.pop('credential_scopes', ['https://management.azure.com/.default']) kwargs.setdefault('sdk_moniker', 'mgmt-applicationinsights/{}'.format(VERSION)) self._configure(**kwargs) def _configure( self, **kwargs: Any ) -> None: self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) self.http_logging_policy = kwargs.get('http_logging_policy') or ARMHttpLoggingPolicy(**kwargs) self.retry_policy = kwargs.get('retry_policy') or policies.AsyncRetryPolicy(**kwargs) self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) self.redirect_policy = kwargs.get('redirect_policy') or policies.AsyncRedirectPolicy(**kwargs) self.authentication_policy = kwargs.get('authentication_policy') if self.credential and not self.authentication_policy: self.authentication_policy = policies.AsyncBearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs)
48.507463
134
0.701846
from typing import Any, TYPE_CHECKING from azure.core.configuration import Configuration from azure.core.pipeline import policies from azure.mgmt.core.policies import ARMHttpLoggingPolicy if TYPE_CHECKING: from azure.core.credentials_async import AsyncTokenCredential VERSION = "unknown" class ApplicationInsightsManagementClientConfiguration(Configuration): def __init__( self, credential: "AsyncTokenCredential", subscription_id: str, **kwargs: Any ) -> None: if credential is None: raise ValueError("Parameter 'credential' must not be None.") if subscription_id is None: raise ValueError("Parameter 'subscription_id' must not be None.") super(ApplicationInsightsManagementClientConfiguration, self).__init__(**kwargs) self.credential = credential self.subscription_id = subscription_id self.api_version = "2015-05-01" self.credential_scopes = kwargs.pop('credential_scopes', ['https://management.azure.com/.default']) kwargs.setdefault('sdk_moniker', 'mgmt-applicationinsights/{}'.format(VERSION)) self._configure(**kwargs) def _configure( self, **kwargs: Any ) -> None: self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) self.http_logging_policy = kwargs.get('http_logging_policy') or ARMHttpLoggingPolicy(**kwargs) self.retry_policy = kwargs.get('retry_policy') or policies.AsyncRetryPolicy(**kwargs) self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) self.redirect_policy = kwargs.get('redirect_policy') or policies.AsyncRedirectPolicy(**kwargs) self.authentication_policy = kwargs.get('authentication_policy') if self.credential and not self.authentication_policy: self.authentication_policy = policies.AsyncBearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs)
true
true
1c485d1431e0cce094796e78eee222184717bdb4
2,473
py
Python
userbot/utils/tools.py
ronaldyganteng/WeebProject
d630cda9f79fafd83453650e414aa59ae136303e
[ "Naumen", "Condor-1.1", "MS-PL" ]
1
2021-05-29T05:31:53.000Z
2021-05-29T05:31:53.000Z
userbot/utils/tools.py
ronaldyganteng/WeebProject
d630cda9f79fafd83453650e414aa59ae136303e
[ "Naumen", "Condor-1.1", "MS-PL" ]
null
null
null
userbot/utils/tools.py
ronaldyganteng/WeebProject
d630cda9f79fafd83453650e414aa59ae136303e
[ "Naumen", "Condor-1.1", "MS-PL" ]
21
2021-02-01T14:01:42.000Z
2021-08-22T01:13:28.000Z
# Copyright (C) 2020 Adek Maulana # # SPDX-License-Identifier: GPL-3.0-or-later # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import asyncio import re import hashlib from typing import List async def md5(fname: str) -> str: hash_md5 = hashlib.md5() with open(fname, "rb") as f: for chunk in iter(lambda: f.read(4096), b""): hash_md5.update(chunk) return hash_md5.hexdigest() def humanbytes(size: int) -> str: if size is None or isinstance(size, str): return "" power = 2**10 raised_to_pow = 0 dict_power_n = {0: "", 1: "Ki", 2: "Mi", 3: "Gi", 4: "Ti"} while size > power: size /= power raised_to_pow += 1 return str(round(size, 2)) + " " + dict_power_n[raised_to_pow] + "B" def time_formatter(seconds: int) -> str: minutes, seconds = divmod(seconds, 60) hours, minutes = divmod(minutes, 60) days, hours = divmod(hours, 24) tmp = ( ((str(days) + " day(s), ") if days else "") + ((str(hours) + " hour(s), ") if hours else "") + ((str(minutes) + " minute(s), ") if minutes else "") + ((str(seconds) + " second(s), ") if seconds else "") ) return tmp[:-2] def human_to_bytes(size: str) -> int: units = { "M": 2**20, "MB": 2**20, "G": 2**30, "GB": 2**30, "T": 2**40, "TB": 2**40 } size = size.upper() if not re.match(r' ', size): size = re.sub(r'([KMGT])', r' \1', size) number, unit = [string.strip() for string in size.split()] return int(float(number) * units[unit]) async def run_cmd(cmd: List) -> (str, str): process = await asyncio.create_subprocess_exec( *cmd, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, ) out, err = await process.communicate() t_resp = out.strip() e_resp = err.strip() return t_resp, e_resp
30.158537
72
0.615851
import asyncio import re import hashlib from typing import List async def md5(fname: str) -> str: hash_md5 = hashlib.md5() with open(fname, "rb") as f: for chunk in iter(lambda: f.read(4096), b""): hash_md5.update(chunk) return hash_md5.hexdigest() def humanbytes(size: int) -> str: if size is None or isinstance(size, str): return "" power = 2**10 raised_to_pow = 0 dict_power_n = {0: "", 1: "Ki", 2: "Mi", 3: "Gi", 4: "Ti"} while size > power: size /= power raised_to_pow += 1 return str(round(size, 2)) + " " + dict_power_n[raised_to_pow] + "B" def time_formatter(seconds: int) -> str: minutes, seconds = divmod(seconds, 60) hours, minutes = divmod(minutes, 60) days, hours = divmod(hours, 24) tmp = ( ((str(days) + " day(s), ") if days else "") + ((str(hours) + " hour(s), ") if hours else "") + ((str(minutes) + " minute(s), ") if minutes else "") + ((str(seconds) + " second(s), ") if seconds else "") ) return tmp[:-2] def human_to_bytes(size: str) -> int: units = { "M": 2**20, "MB": 2**20, "G": 2**30, "GB": 2**30, "T": 2**40, "TB": 2**40 } size = size.upper() if not re.match(r' ', size): size = re.sub(r'([KMGT])', r' \1', size) number, unit = [string.strip() for string in size.split()] return int(float(number) * units[unit]) async def run_cmd(cmd: List) -> (str, str): process = await asyncio.create_subprocess_exec( *cmd, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, ) out, err = await process.communicate() t_resp = out.strip() e_resp = err.strip() return t_resp, e_resp
true
true
1c485e14fe7d5c3bbc2cb0ef4066486b1c2f9fc0
2,579
py
Python
pyad/pyadexceptions.py
sukhjinderpalsingh/pyad
d95ff67745065cafca4f2653aab1fbce2df91fb9
[ "Apache-2.0" ]
145
2015-01-14T21:53:35.000Z
2022-03-10T19:40:32.000Z
pyad/pyadexceptions.py
zakird/pyad
d95ff67745065cafca4f2653aab1fbce2df91fb9
[ "Apache-2.0" ]
104
2015-01-12T21:14:10.000Z
2022-03-02T12:38:41.000Z
pyad/pyadexceptions.py
sukhjinderpalsingh/pyad
d95ff67745065cafca4f2653aab1fbce2df91fb9
[ "Apache-2.0" ]
73
2015-03-27T07:36:47.000Z
2022-03-04T18:59:18.000Z
from __future__ import print_function from builtins import str class comException(Exception): def __init__(self, error_info, additional_info={}): self.error_info = error_info self.additional_info = additional_info def __str__(self): print("Error Constant: %s" % self.error_info['error_constant']) print("Error Code: %s" % str(self.error_info['error_code'])) #print "Error Message: %s" % self.error_info['error_message'] #print "type is ", self.error_info['error_message'].__class__ #return "%s (%s): %s" % (str(self.error_info['error_constant']), str(self.error_info['error_code']), str(self.error_info['error_message'])) class genericADSIException(comException): def __init__(self, error_info, additional_info={}): comException.__init__(error_info, additional_info) def __str__(self): return "%s (%s): %s" % (self.error_info['error_constant'], self.error_info['error_code'], self.error_info['error_message']) class win32Exception(comException): def __init__(self, error_info, additional_info={}): comException.__init__(self, error_info, additional_info) def __str__(self): return "%s: %s" % (self.error_info['error_code'], self.error_info['message']) class invalidOwnerException(Exception): def __str__(self): return "The submitted object is not eligible to own another object." class noObjectFoundException(Exception): def __str__(self): return "The requested object does not exist." class InvalidObjectException(noObjectFoundException, win32Exception): def __init__(self, error_info, additional_info): win32Exception.__init__(self, error_info, additional_info) class InvalidAttribute(AttributeError): def __init__(self, obj, attribute): self.obj, self.attribute = obj, attribute def __str__(self): return 'The attribute "%s" is not permitted by the schema definition of the object "%s" (the requested attribute does not exist).' % (self.attribute, self.obj) class noExecutedQuery(Exception): def __str__(self): return 'No query has been executed. Therefore there are no results to return. Execute a query before requesting results.' class invalidResults(Exception): def __init__(self, numberResults): self.__numberResults = numberResults def __str__(self): return 'The specified query returned %i results. getSingleResults only functions with a single result.' % self.__numberResults
39.676923
168
0.697945
from __future__ import print_function from builtins import str class comException(Exception): def __init__(self, error_info, additional_info={}): self.error_info = error_info self.additional_info = additional_info def __str__(self): print("Error Constant: %s" % self.error_info['error_constant']) print("Error Code: %s" % str(self.error_info['error_code'])) class genericADSIException(comException): def __init__(self, error_info, additional_info={}): comException.__init__(error_info, additional_info) def __str__(self): return "%s (%s): %s" % (self.error_info['error_constant'], self.error_info['error_code'], self.error_info['error_message']) class win32Exception(comException): def __init__(self, error_info, additional_info={}): comException.__init__(self, error_info, additional_info) def __str__(self): return "%s: %s" % (self.error_info['error_code'], self.error_info['message']) class invalidOwnerException(Exception): def __str__(self): return "The submitted object is not eligible to own another object." class noObjectFoundException(Exception): def __str__(self): return "The requested object does not exist." class InvalidObjectException(noObjectFoundException, win32Exception): def __init__(self, error_info, additional_info): win32Exception.__init__(self, error_info, additional_info) class InvalidAttribute(AttributeError): def __init__(self, obj, attribute): self.obj, self.attribute = obj, attribute def __str__(self): return 'The attribute "%s" is not permitted by the schema definition of the object "%s" (the requested attribute does not exist).' % (self.attribute, self.obj) class noExecutedQuery(Exception): def __str__(self): return 'No query has been executed. Therefore there are no results to return. Execute a query before requesting results.' class invalidResults(Exception): def __init__(self, numberResults): self.__numberResults = numberResults def __str__(self): return 'The specified query returned %i results. getSingleResults only functions with a single result.' % self.__numberResults
true
true
1c485ee284a25efec48426c3601ced0b86cf1c38
1,167
py
Python
axitom/phantoms.py
PolymerGuy/AXITOM
7682be5b21fa933b9bea4082fe9a830076431feb
[ "MIT" ]
4
2019-09-06T16:31:11.000Z
2022-02-04T12:18:47.000Z
axitom/phantoms.py
PolymerGuy/AXITOM
7682be5b21fa933b9bea4082fe9a830076431feb
[ "MIT" ]
1
2019-08-08T12:30:33.000Z
2019-08-08T12:34:55.000Z
axitom/phantoms.py
PolymerGuy/AXITOM
7682be5b21fa933b9bea4082fe9a830076431feb
[ "MIT" ]
7
2019-08-21T20:51:12.000Z
2020-02-04T14:20:42.000Z
import numpy as np """ Phantoms This module contains the phantoms that can be used for forward projection and virtual experiments """ def barrel(domain_size=128, outer_rad_fraction=0.7,center_val=None): """ Barrel shaped phantom with a linear density gradient The domain size is cubic with dimension "domain_size" along each axis Parameters ---------- domain_size : int The length of the sides of the domain outer_rad_fraction : float The diameter of the barrel given as a the fraction of the side length center_val : float The density value in the center of the barrel Returns ------- ndarray The phantom """ center = domain_size / 2. domain = np.zeros((domain_size, domain_size, domain_size), dtype=np.float64) xs, ys = np.meshgrid(np.arange(domain_size), np.arange(domain_size)) xs = xs - center ys = ys - center r = np.sqrt(xs ** 2. + ys ** 2.) domain[r < outer_rad_fraction * center, :] = 1. if center_val is not None: domain = domain * (center_val + (r / (outer_rad_fraction * center)) ** 2. * 0.5)[:, :, np.newaxis] return domain
28.463415
106
0.652099
import numpy as np def barrel(domain_size=128, outer_rad_fraction=0.7,center_val=None): center = domain_size / 2. domain = np.zeros((domain_size, domain_size, domain_size), dtype=np.float64) xs, ys = np.meshgrid(np.arange(domain_size), np.arange(domain_size)) xs = xs - center ys = ys - center r = np.sqrt(xs ** 2. + ys ** 2.) domain[r < outer_rad_fraction * center, :] = 1. if center_val is not None: domain = domain * (center_val + (r / (outer_rad_fraction * center)) ** 2. * 0.5)[:, :, np.newaxis] return domain
true
true
1c485f4180c62d34716606f40f2b0dda9ebcd895
10,087
py
Python
tests/integration/standard/test_cluster.py
josh-mckenzie/python-driver
472675c61664afa99d2c9eb6c32424c846c1a367
[ "Apache-2.0" ]
null
null
null
tests/integration/standard/test_cluster.py
josh-mckenzie/python-driver
472675c61664afa99d2c9eb6c32424c846c1a367
[ "Apache-2.0" ]
null
null
null
tests/integration/standard/test_cluster.py
josh-mckenzie/python-driver
472675c61664afa99d2c9eb6c32424c846c1a367
[ "Apache-2.0" ]
null
null
null
# Copyright 2013-2014 DataStax, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from tests.integration import PROTOCOL_VERSION try: import unittest2 as unittest except ImportError: import unittest # noqa import cassandra from cassandra.query import SimpleStatement, TraceUnavailable from cassandra.policies import RoundRobinPolicy, ExponentialReconnectionPolicy, RetryPolicy, SimpleConvictionPolicy, HostDistance from cassandra.cluster import Cluster, NoHostAvailable class ClusterTests(unittest.TestCase): def test_basic(self): """ Test basic connection and usage """ cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() result = session.execute( """ CREATE KEYSPACE clustertests WITH replication = {'class': 'SimpleStrategy', 'replication_factor': '1'} """) self.assertEqual(None, result) result = session.execute( """ CREATE TABLE clustertests.cf0 ( a text, b text, c text, PRIMARY KEY (a, b) ) """) self.assertEqual(None, result) result = session.execute( """ INSERT INTO clustertests.cf0 (a, b, c) VALUES ('a', 'b', 'c') """) self.assertEqual(None, result) result = session.execute("SELECT * FROM clustertests.cf0") self.assertEqual([('a', 'b', 'c')], result) cluster.shutdown() def test_connect_on_keyspace(self): """ Ensure clusters that connect on a keyspace, do """ cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() result = session.execute( """ INSERT INTO test3rf.test (k, v) VALUES (8889, 8889) """) self.assertEqual(None, result) result = session.execute("SELECT * FROM test3rf.test") self.assertEqual([(8889, 8889)], result) # test_connect_on_keyspace session2 = cluster.connect('test3rf') result2 = session2.execute("SELECT * FROM test") self.assertEqual(result, result2) def test_set_keyspace_twice(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() session.execute("USE system") session.execute("USE system") def test_default_connections(self): """ Ensure errors are not thrown when using non-default policies """ Cluster( load_balancing_policy=RoundRobinPolicy(), reconnection_policy=ExponentialReconnectionPolicy(1.0, 600.0), default_retry_policy=RetryPolicy(), conviction_policy_factory=SimpleConvictionPolicy, protocol_version=PROTOCOL_VERSION ) def test_connect_to_already_shutdown_cluster(self): """ Ensure you cannot connect to a cluster that's been shutdown """ cluster = Cluster(protocol_version=PROTOCOL_VERSION) cluster.shutdown() self.assertRaises(Exception, cluster.connect) def test_auth_provider_is_callable(self): """ Ensure that auth_providers are always callable """ self.assertRaises(TypeError, Cluster, auth_provider=1, protocol_version=1) c = Cluster(protocol_version=1) self.assertRaises(TypeError, setattr, c, 'auth_provider', 1) def test_v2_auth_provider(self): """ Check for v2 auth_provider compliance """ bad_auth_provider = lambda x: {'username': 'foo', 'password': 'bar'} self.assertRaises(TypeError, Cluster, auth_provider=bad_auth_provider, protocol_version=2) c = Cluster(protocol_version=2) self.assertRaises(TypeError, setattr, c, 'auth_provider', bad_auth_provider) def test_conviction_policy_factory_is_callable(self): """ Ensure that conviction_policy_factory are always callable """ self.assertRaises(ValueError, Cluster, conviction_policy_factory=1) def test_connect_to_bad_hosts(self): """ Ensure that a NoHostAvailable Exception is thrown when a cluster cannot connect to given hosts """ cluster = Cluster(['127.1.2.9', '127.1.2.10'], protocol_version=PROTOCOL_VERSION) self.assertRaises(NoHostAvailable, cluster.connect) def test_cluster_settings(self): """ Test connection setting getters and setters """ if PROTOCOL_VERSION >= 3: raise unittest.SkipTest("min/max requests and core/max conns aren't used with v3 protocol") cluster = Cluster(protocol_version=PROTOCOL_VERSION) min_requests_per_connection = cluster.get_min_requests_per_connection(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MIN_REQUESTS, min_requests_per_connection) cluster.set_min_requests_per_connection(HostDistance.LOCAL, min_requests_per_connection + 1) self.assertEqual(cluster.get_min_requests_per_connection(HostDistance.LOCAL), min_requests_per_connection + 1) max_requests_per_connection = cluster.get_max_requests_per_connection(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MAX_REQUESTS, max_requests_per_connection) cluster.set_max_requests_per_connection(HostDistance.LOCAL, max_requests_per_connection + 1) self.assertEqual(cluster.get_max_requests_per_connection(HostDistance.LOCAL), max_requests_per_connection + 1) core_connections_per_host = cluster.get_core_connections_per_host(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MIN_CONNECTIONS_PER_LOCAL_HOST, core_connections_per_host) cluster.set_core_connections_per_host(HostDistance.LOCAL, core_connections_per_host + 1) self.assertEqual(cluster.get_core_connections_per_host(HostDistance.LOCAL), core_connections_per_host + 1) max_connections_per_host = cluster.get_max_connections_per_host(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MAX_CONNECTIONS_PER_LOCAL_HOST, max_connections_per_host) cluster.set_max_connections_per_host(HostDistance.LOCAL, max_connections_per_host + 1) self.assertEqual(cluster.get_max_connections_per_host(HostDistance.LOCAL), max_connections_per_host + 1) def test_submit_schema_refresh(self): """ Ensure new new schema is refreshed after submit_schema_refresh() """ cluster = Cluster(protocol_version=PROTOCOL_VERSION) cluster.connect() self.assertNotIn("newkeyspace", cluster.metadata.keyspaces) other_cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = other_cluster.connect() session.execute( """ CREATE KEYSPACE newkeyspace WITH replication = {'class': 'SimpleStrategy', 'replication_factor': '1'} """) future = cluster.submit_schema_refresh() future.result() self.assertIn("newkeyspace", cluster.metadata.keyspaces) def test_trace(self): """ Ensure trace can be requested for async and non-async queries """ cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() self.assertRaises(TypeError, session.execute, "SELECT * FROM system.local", trace=True) def check_trace(trace): self.assertIsNot(None, trace.request_type) self.assertIsNot(None, trace.duration) self.assertIsNot(None, trace.started_at) self.assertIsNot(None, trace.coordinator) self.assertIsNot(None, trace.events) query = "SELECT * FROM system.local" statement = SimpleStatement(query) session.execute(statement, trace=True) check_trace(statement.trace) query = "SELECT * FROM system.local" statement = SimpleStatement(query) session.execute(statement) self.assertEqual(None, statement.trace) statement2 = SimpleStatement(query) future = session.execute_async(statement2, trace=True) future.result() check_trace(future.get_query_trace()) statement2 = SimpleStatement(query) future = session.execute_async(statement2) future.result() self.assertEqual(None, future.get_query_trace()) prepared = session.prepare("SELECT * FROM system.local") future = session.execute_async(prepared, parameters=(), trace=True) future.result() check_trace(future.get_query_trace()) def test_trace_timeout(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() query = "SELECT * FROM system.local" statement = SimpleStatement(query) future = session.execute_async(statement, trace=True) future.result() self.assertRaises(TraceUnavailable, future.get_query_trace, -1.0) def test_string_coverage(self): """ Ensure str(future) returns without error """ cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() query = "SELECT * FROM system.local" statement = SimpleStatement(query) future = session.execute_async(statement) self.assertIn(query, str(future)) future.result() self.assertIn(query, str(future)) self.assertIn('result', str(future))
37.359259
129
0.674234
from tests.integration import PROTOCOL_VERSION try: import unittest2 as unittest except ImportError: import unittest import cassandra from cassandra.query import SimpleStatement, TraceUnavailable from cassandra.policies import RoundRobinPolicy, ExponentialReconnectionPolicy, RetryPolicy, SimpleConvictionPolicy, HostDistance from cassandra.cluster import Cluster, NoHostAvailable class ClusterTests(unittest.TestCase): def test_basic(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() result = session.execute( """ CREATE KEYSPACE clustertests WITH replication = {'class': 'SimpleStrategy', 'replication_factor': '1'} """) self.assertEqual(None, result) result = session.execute( """ CREATE TABLE clustertests.cf0 ( a text, b text, c text, PRIMARY KEY (a, b) ) """) self.assertEqual(None, result) result = session.execute( """ INSERT INTO clustertests.cf0 (a, b, c) VALUES ('a', 'b', 'c') """) self.assertEqual(None, result) result = session.execute("SELECT * FROM clustertests.cf0") self.assertEqual([('a', 'b', 'c')], result) cluster.shutdown() def test_connect_on_keyspace(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() result = session.execute( """ INSERT INTO test3rf.test (k, v) VALUES (8889, 8889) """) self.assertEqual(None, result) result = session.execute("SELECT * FROM test3rf.test") self.assertEqual([(8889, 8889)], result) session2 = cluster.connect('test3rf') result2 = session2.execute("SELECT * FROM test") self.assertEqual(result, result2) def test_set_keyspace_twice(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() session.execute("USE system") session.execute("USE system") def test_default_connections(self): Cluster( load_balancing_policy=RoundRobinPolicy(), reconnection_policy=ExponentialReconnectionPolicy(1.0, 600.0), default_retry_policy=RetryPolicy(), conviction_policy_factory=SimpleConvictionPolicy, protocol_version=PROTOCOL_VERSION ) def test_connect_to_already_shutdown_cluster(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) cluster.shutdown() self.assertRaises(Exception, cluster.connect) def test_auth_provider_is_callable(self): self.assertRaises(TypeError, Cluster, auth_provider=1, protocol_version=1) c = Cluster(protocol_version=1) self.assertRaises(TypeError, setattr, c, 'auth_provider', 1) def test_v2_auth_provider(self): bad_auth_provider = lambda x: {'username': 'foo', 'password': 'bar'} self.assertRaises(TypeError, Cluster, auth_provider=bad_auth_provider, protocol_version=2) c = Cluster(protocol_version=2) self.assertRaises(TypeError, setattr, c, 'auth_provider', bad_auth_provider) def test_conviction_policy_factory_is_callable(self): self.assertRaises(ValueError, Cluster, conviction_policy_factory=1) def test_connect_to_bad_hosts(self): cluster = Cluster(['127.1.2.9', '127.1.2.10'], protocol_version=PROTOCOL_VERSION) self.assertRaises(NoHostAvailable, cluster.connect) def test_cluster_settings(self): if PROTOCOL_VERSION >= 3: raise unittest.SkipTest("min/max requests and core/max conns aren't used with v3 protocol") cluster = Cluster(protocol_version=PROTOCOL_VERSION) min_requests_per_connection = cluster.get_min_requests_per_connection(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MIN_REQUESTS, min_requests_per_connection) cluster.set_min_requests_per_connection(HostDistance.LOCAL, min_requests_per_connection + 1) self.assertEqual(cluster.get_min_requests_per_connection(HostDistance.LOCAL), min_requests_per_connection + 1) max_requests_per_connection = cluster.get_max_requests_per_connection(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MAX_REQUESTS, max_requests_per_connection) cluster.set_max_requests_per_connection(HostDistance.LOCAL, max_requests_per_connection + 1) self.assertEqual(cluster.get_max_requests_per_connection(HostDistance.LOCAL), max_requests_per_connection + 1) core_connections_per_host = cluster.get_core_connections_per_host(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MIN_CONNECTIONS_PER_LOCAL_HOST, core_connections_per_host) cluster.set_core_connections_per_host(HostDistance.LOCAL, core_connections_per_host + 1) self.assertEqual(cluster.get_core_connections_per_host(HostDistance.LOCAL), core_connections_per_host + 1) max_connections_per_host = cluster.get_max_connections_per_host(HostDistance.LOCAL) self.assertEqual(cassandra.cluster.DEFAULT_MAX_CONNECTIONS_PER_LOCAL_HOST, max_connections_per_host) cluster.set_max_connections_per_host(HostDistance.LOCAL, max_connections_per_host + 1) self.assertEqual(cluster.get_max_connections_per_host(HostDistance.LOCAL), max_connections_per_host + 1) def test_submit_schema_refresh(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) cluster.connect() self.assertNotIn("newkeyspace", cluster.metadata.keyspaces) other_cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = other_cluster.connect() session.execute( """ CREATE KEYSPACE newkeyspace WITH replication = {'class': 'SimpleStrategy', 'replication_factor': '1'} """) future = cluster.submit_schema_refresh() future.result() self.assertIn("newkeyspace", cluster.metadata.keyspaces) def test_trace(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() self.assertRaises(TypeError, session.execute, "SELECT * FROM system.local", trace=True) def check_trace(trace): self.assertIsNot(None, trace.request_type) self.assertIsNot(None, trace.duration) self.assertIsNot(None, trace.started_at) self.assertIsNot(None, trace.coordinator) self.assertIsNot(None, trace.events) query = "SELECT * FROM system.local" statement = SimpleStatement(query) session.execute(statement, trace=True) check_trace(statement.trace) query = "SELECT * FROM system.local" statement = SimpleStatement(query) session.execute(statement) self.assertEqual(None, statement.trace) statement2 = SimpleStatement(query) future = session.execute_async(statement2, trace=True) future.result() check_trace(future.get_query_trace()) statement2 = SimpleStatement(query) future = session.execute_async(statement2) future.result() self.assertEqual(None, future.get_query_trace()) prepared = session.prepare("SELECT * FROM system.local") future = session.execute_async(prepared, parameters=(), trace=True) future.result() check_trace(future.get_query_trace()) def test_trace_timeout(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() query = "SELECT * FROM system.local" statement = SimpleStatement(query) future = session.execute_async(statement, trace=True) future.result() self.assertRaises(TraceUnavailable, future.get_query_trace, -1.0) def test_string_coverage(self): cluster = Cluster(protocol_version=PROTOCOL_VERSION) session = cluster.connect() query = "SELECT * FROM system.local" statement = SimpleStatement(query) future = session.execute_async(statement) self.assertIn(query, str(future)) future.result() self.assertIn(query, str(future)) self.assertIn('result', str(future))
true
true
1c486091f1cf476ba15dd0ac7d22a9dfbee4c1fc
7,319
py
Python
code/src/d00_utils/feat_utils.py
edugm94/temporal-feat-emotion-prediction
6548bbf5f5d8969de97c076ebc9b5462d7b8bdd4
[ "MIT" ]
null
null
null
code/src/d00_utils/feat_utils.py
edugm94/temporal-feat-emotion-prediction
6548bbf5f5d8969de97c076ebc9b5462d7b8bdd4
[ "MIT" ]
null
null
null
code/src/d00_utils/feat_utils.py
edugm94/temporal-feat-emotion-prediction
6548bbf5f5d8969de97c076ebc9b5462d7b8bdd4
[ "MIT" ]
null
null
null
# !/usr/bin/env python # -*- coding: UTF-8 -*- # # Author: Eduardo Gutierrez Maestro # Date: 2021.12.14 # email: eduardo.gutierrez-maestro@oru.se # # Center for Applied Autonomous Sensor Systems (AASS), Cognitive Robotic Systems Labs # University of Orebro, Sweden import numpy as np import pandas as pd from scipy.fft import fft, ifft from csaps import csaps def clean_dataset(labels, discard=0.1): # Obtain an accounting of hbiw many vectors there is for each emotion unique, counts = np.unique(labels, return_counts=True) counting = dict(zip(unique, counts)) # Get the total amount of vectors and the threshold to filter dictionary tot = sum(counting.values()) threshold = tot * discard # Get a dictionary with the emotions that should be cleaned from the initial variables # It is kept a dictionary to check the lenght of the cleaned values at the end emo_del_dict = dict(filter(lambda elem: elem[1] < threshold, counting.items())) # Array that contains the value of the emotions to be cleaned in the "labels" variable emo_del_arr = np.array(list(emo_del_dict.keys())) # Array containing the index that should be deleted from "data" and "label" indx_del_arr = np.where(labels == emo_del_arr)[0] assert indx_del_arr.shape[0] == sum( emo_del_dict.values()), "The amount of vectors to delete does not match! Check it." return indx_del_arr def filter_nrows(feature_, lab_): # 1 Step: Find min number of rows within signals min_nrow = np.inf for arr in feature_: nrow_ = arr.shape[0] min_nrow = nrow_ if nrow_ < min_nrow else min_nrow # 2 Step: Modify each array in the list in case feat_signals_filter = [] lab_signals_filter = [] for arr, lab in zip(feature_, lab_): arr_ = arr[0:min_nrow, :] lab_ = lab[0:min_nrow,] feat_signals_filter.append(arr_) lab_signals_filter.append(lab_) return feat_signals_filter, lab_signals_filter[0] #def extract_ts_features(df, emotion, weda, patient, day, signal): def extract_ts_features(df, signal): #DATASET = "/home/eduardo/phd/projects/opt-physio-feat-extractor/2-emotion-to-vector/out/filter/" #DATASET = "/home/eduardo/phd/projects/physio-feat-extractor/physio-feat-extractor/2-emotion-to-vector/out/filter/" signal2Freq = { "HR": 1, "ACC": 32, "EDA": 4, "TEMP": 4 } FREQ = signal2Freq[signal] Ts = 1 / FREQ WINDOW = 60 # sliding window size OVERLAP = 0.1 #datapath = DATASET + "{}/{}/{}/{}/{}.csv".format(emotion, weda, patient, day, signal) #df = pd.read_csv(datapath, sep='\t') df = df.reset_index() if signal == 'EDA': eda = df['eda'].to_numpy() x = np.arange(0, len(eda), 1) scl = csaps(x, eda, x, smooth=0.5) scr = np.real_if_close(ifft(fft(eda) / fft(scl))) df['scr'] = scr df['scl'] = scl if df.shape[0] == 1: # You should create an Dataframe with -1 to indicate that there is no data available print("Empty DataFrame. Exiting program...") return -1, -1 init_id = df['index'].iloc[0] end_id = df['index'].iloc[-1] # Boundaries to control cases where EMA is rigth at beginning and end of Dataframe init_bound_ind = int(init_id + WINDOW * FREQ / 2) end_bound_ind = int(end_id - WINDOW * FREQ / 2) # Get id that correspond to a label; Filtering: checking boundaries idx = df.index[(df['label'] != -1) & (init_bound_ind < df['index']) & (df['index'] < end_bound_ind)] idxs = np.asarray(idx) idxs_aux = idxs[0:len(idxs) - 1] init_ema_aux = (idxs[1:len(idxs)] - idxs_aux).reshape(-1, 1) init_ema_id = np.where(np.any(init_ema_aux > 1, axis=1))[0] + 1 init_ema_indices = idxs[init_ema_id].tolist() counter = 1 w_central = idxs[0] # It is chosen first element: filtered list id_ = 0 offset = (1 - OVERLAP) * WINDOW * FREQ ts_df = pd.DataFrame(columns=['id', 'time', 'kind', 'value']) ts_df_label = pd.DataFrame(columns=['id', 'label']) while id_ <= len(idxs): w_left = w_central - (WINDOW * FREQ / 2 - 1) w_right = w_central + (WINDOW * FREQ / 2 - 1) df_window = df.loc[df['index'].between(w_left, w_right)] label_ = int(df['label'][df['index'].iloc[w_central]]) if signal == 'ACC': x_ = df_window['x'].to_numpy() y_ = df_window['y'].to_numpy() z_ = df_window['z'].to_numpy() n_ = df_window['n'].to_numpy() ts_ = np.arange(len(x_)) data_x = {'id': counter, 'time': ts_, 'kind': 'acc_x', 'value': x_} data_y = {'id': counter, 'time': ts_, 'kind': 'acc_y', 'value': y_} data_z = {'id': counter, 'time': ts_, 'kind': 'acc_z', 'value': z_} data_n = {'id': counter, 'time': ts_, 'kind': 'acc_n', 'value': n_} aux_x = pd.DataFrame(data=data_x) aux_y = pd.DataFrame(data=data_y) aux_z = pd.DataFrame(data=data_z) aux_n = pd.DataFrame(data=data_n) aux_ = pd.concat([aux_x, aux_y, aux_z, aux_n], axis=0) elif signal == "EDA": eda_ = df_window[signal.lower()].to_numpy() ts_ = np.arange(len(eda_)) data_eda = {'id': counter, 'time': ts_, 'kind': "eda", 'value': eda_} aux_eda = pd.DataFrame(data=data_eda) scl_ = df_window['scl'].to_numpy() ts_ = np.arange(len(scl_)) data_scl = {'id': counter, 'time': ts_, 'kind': "scl", 'value': scl_} aux_scl = pd.DataFrame(data=data_scl) scr_ = df_window['scr'].to_numpy() ts_ = np.arange(len(scr_)) data_scr = {'id': counter, 'time': ts_, 'kind': "scr", 'value': scr_} aux_scr = pd.DataFrame(data=data_scr) aux_ = pd.concat([aux_eda, aux_scl, aux_scr], axis=0) elif signal == "TEMP": data_ = df_window[signal.lower()].to_numpy() ts_ = np.arange(len(data_)) data = {'id': counter, 'time': ts_, 'kind': "temp", 'value': data_} aux_ = pd.DataFrame(data=data) else: data_ = df_window[signal.lower()].to_numpy() ts_ = np.arange(len(data_)) data = {'id': counter, 'time': ts_, 'kind': "hr", 'value': data_} aux_ = pd.DataFrame(data=data) data_label = {'id': [counter], "label": [label_]} aux_label = pd.DataFrame(data=data_label) ts_df = pd.concat([ts_df, aux_], axis=0) ts_df_label = pd.concat([ts_df_label, aux_label], axis=0) # Logic code to move sliding window w_central_aux = int(w_central + offset) if w_central_aux not in idxs: if not init_ema_indices: # if this list is empty it means that you arrived to the end break # Skip to next starting EMA indicated by init_ema_indices w_central = init_ema_indices.pop(0) # modify variable id_ id_ = np.where(idxs == w_central)[0][0] else: # id_ = indice que ocupa w_central_aux en la lista idxs w_central = w_central_aux id_ = np.where(idxs == w_central_aux)[0][0] counter += 1 return ts_df, ts_df_label
38.521053
119
0.599399
import numpy as np import pandas as pd from scipy.fft import fft, ifft from csaps import csaps def clean_dataset(labels, discard=0.1): unique, counts = np.unique(labels, return_counts=True) counting = dict(zip(unique, counts)) tot = sum(counting.values()) threshold = tot * discard emo_del_dict = dict(filter(lambda elem: elem[1] < threshold, counting.items())) emo_del_arr = np.array(list(emo_del_dict.keys())) indx_del_arr = np.where(labels == emo_del_arr)[0] assert indx_del_arr.shape[0] == sum( emo_del_dict.values()), "The amount of vectors to delete does not match! Check it." return indx_del_arr def filter_nrows(feature_, lab_): min_nrow = np.inf for arr in feature_: nrow_ = arr.shape[0] min_nrow = nrow_ if nrow_ < min_nrow else min_nrow feat_signals_filter = [] lab_signals_filter = [] for arr, lab in zip(feature_, lab_): arr_ = arr[0:min_nrow, :] lab_ = lab[0:min_nrow,] feat_signals_filter.append(arr_) lab_signals_filter.append(lab_) return feat_signals_filter, lab_signals_filter[0] def extract_ts_features(df, signal): signal2Freq = { "HR": 1, "ACC": 32, "EDA": 4, "TEMP": 4 } FREQ = signal2Freq[signal] Ts = 1 / FREQ WINDOW = 60 OVERLAP = 0.1 df = df.reset_index() if signal == 'EDA': eda = df['eda'].to_numpy() x = np.arange(0, len(eda), 1) scl = csaps(x, eda, x, smooth=0.5) scr = np.real_if_close(ifft(fft(eda) / fft(scl))) df['scr'] = scr df['scl'] = scl if df.shape[0] == 1: print("Empty DataFrame. Exiting program...") return -1, -1 init_id = df['index'].iloc[0] end_id = df['index'].iloc[-1] init_bound_ind = int(init_id + WINDOW * FREQ / 2) end_bound_ind = int(end_id - WINDOW * FREQ / 2) idx = df.index[(df['label'] != -1) & (init_bound_ind < df['index']) & (df['index'] < end_bound_ind)] idxs = np.asarray(idx) idxs_aux = idxs[0:len(idxs) - 1] init_ema_aux = (idxs[1:len(idxs)] - idxs_aux).reshape(-1, 1) init_ema_id = np.where(np.any(init_ema_aux > 1, axis=1))[0] + 1 init_ema_indices = idxs[init_ema_id].tolist() counter = 1 w_central = idxs[0] id_ = 0 offset = (1 - OVERLAP) * WINDOW * FREQ ts_df = pd.DataFrame(columns=['id', 'time', 'kind', 'value']) ts_df_label = pd.DataFrame(columns=['id', 'label']) while id_ <= len(idxs): w_left = w_central - (WINDOW * FREQ / 2 - 1) w_right = w_central + (WINDOW * FREQ / 2 - 1) df_window = df.loc[df['index'].between(w_left, w_right)] label_ = int(df['label'][df['index'].iloc[w_central]]) if signal == 'ACC': x_ = df_window['x'].to_numpy() y_ = df_window['y'].to_numpy() z_ = df_window['z'].to_numpy() n_ = df_window['n'].to_numpy() ts_ = np.arange(len(x_)) data_x = {'id': counter, 'time': ts_, 'kind': 'acc_x', 'value': x_} data_y = {'id': counter, 'time': ts_, 'kind': 'acc_y', 'value': y_} data_z = {'id': counter, 'time': ts_, 'kind': 'acc_z', 'value': z_} data_n = {'id': counter, 'time': ts_, 'kind': 'acc_n', 'value': n_} aux_x = pd.DataFrame(data=data_x) aux_y = pd.DataFrame(data=data_y) aux_z = pd.DataFrame(data=data_z) aux_n = pd.DataFrame(data=data_n) aux_ = pd.concat([aux_x, aux_y, aux_z, aux_n], axis=0) elif signal == "EDA": eda_ = df_window[signal.lower()].to_numpy() ts_ = np.arange(len(eda_)) data_eda = {'id': counter, 'time': ts_, 'kind': "eda", 'value': eda_} aux_eda = pd.DataFrame(data=data_eda) scl_ = df_window['scl'].to_numpy() ts_ = np.arange(len(scl_)) data_scl = {'id': counter, 'time': ts_, 'kind': "scl", 'value': scl_} aux_scl = pd.DataFrame(data=data_scl) scr_ = df_window['scr'].to_numpy() ts_ = np.arange(len(scr_)) data_scr = {'id': counter, 'time': ts_, 'kind': "scr", 'value': scr_} aux_scr = pd.DataFrame(data=data_scr) aux_ = pd.concat([aux_eda, aux_scl, aux_scr], axis=0) elif signal == "TEMP": data_ = df_window[signal.lower()].to_numpy() ts_ = np.arange(len(data_)) data = {'id': counter, 'time': ts_, 'kind': "temp", 'value': data_} aux_ = pd.DataFrame(data=data) else: data_ = df_window[signal.lower()].to_numpy() ts_ = np.arange(len(data_)) data = {'id': counter, 'time': ts_, 'kind': "hr", 'value': data_} aux_ = pd.DataFrame(data=data) data_label = {'id': [counter], "label": [label_]} aux_label = pd.DataFrame(data=data_label) ts_df = pd.concat([ts_df, aux_], axis=0) ts_df_label = pd.concat([ts_df_label, aux_label], axis=0) w_central_aux = int(w_central + offset) if w_central_aux not in idxs: if not init_ema_indices: break w_central = init_ema_indices.pop(0) id_ = np.where(idxs == w_central)[0][0] else: w_central = w_central_aux id_ = np.where(idxs == w_central_aux)[0][0] counter += 1 return ts_df, ts_df_label
true
true
1c4860b88312afac4669bab44eca6d6d09937ccf
1,704
py
Python
python/open3d/ml/torch/pipelines.py
Dudulle/Open3D
ffed2d1bee6d45b6acc4b7ae7133752e50d6ecab
[ "MIT" ]
28
2021-03-02T09:51:12.000Z
2022-03-17T09:27:46.000Z
python/open3d/ml/torch/pipelines.py
Dudulle/Open3D
ffed2d1bee6d45b6acc4b7ae7133752e50d6ecab
[ "MIT" ]
27
2021-03-08T06:56:35.000Z
2022-03-25T14:00:32.000Z
python/open3d/ml/torch/pipelines.py
Dudulle/Open3D
ffed2d1bee6d45b6acc4b7ae7133752e50d6ecab
[ "MIT" ]
7
2021-08-24T02:20:13.000Z
2021-12-31T09:45:02.000Z
# ---------------------------------------------------------------------------- # - Open3D: www.open3d.org - # ---------------------------------------------------------------------------- # The MIT License (MIT) # # Copyright (c) 2020 www.open3d.org # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. # ---------------------------------------------------------------------------- """3D ML pipelines for PyTorch.""" import os as _os from open3d import _build_config if _build_config['BUNDLE_OPEN3D_ML']: if 'OPEN3D_ML_ROOT' in _os.environ: from ml3d.torch.pipelines import * else: from open3d._ml3d.torch.pipelines import *
47.333333
79
0.632042
import os as _os from open3d import _build_config if _build_config['BUNDLE_OPEN3D_ML']: if 'OPEN3D_ML_ROOT' in _os.environ: from ml3d.torch.pipelines import * else: from open3d._ml3d.torch.pipelines import *
true
true
1c4861abce1abcc4cf3861647e4c50dd84b20861
5,951
py
Python
util.py
seanliu96/R-Net
8462330451079a2ff67cd431fe30a57a6ca3d802
[ "MIT" ]
null
null
null
util.py
seanliu96/R-Net
8462330451079a2ff67cd431fe30a57a6ca3d802
[ "MIT" ]
null
null
null
util.py
seanliu96/R-Net
8462330451079a2ff67cd431fe30a57a6ca3d802
[ "MIT" ]
null
null
null
import tensorflow as tf import numpy as np import re from collections import Counter import string def get_record_parser(config, is_test=False): def parse(example): para_limit = config.test_para_limit if is_test else config.para_limit ques_limit = config.test_ques_limit if is_test else config.ques_limit char_limit = config.char_limit features = tf.parse_single_example(example, features={ "context_idxs": tf.FixedLenFeature([], tf.string), "ques_idxs": tf.FixedLenFeature([], tf.string), "context_char_idxs": tf.FixedLenFeature([], tf.string), "ques_char_idxs": tf.FixedLenFeature([], tf.string), "y1": tf.FixedLenFeature([], tf.string), "y2": tf.FixedLenFeature([], tf.string), "id": tf.FixedLenFeature([], tf.int64) }) context_idxs = tf.reshape(tf.decode_raw( features["context_idxs"], tf.int32), [para_limit]) ques_idxs = tf.reshape(tf.decode_raw( features["ques_idxs"], tf.int32), [ques_limit]) context_char_idxs = tf.reshape(tf.decode_raw( features["context_char_idxs"], tf.int32), [para_limit, char_limit]) ques_char_idxs = tf.reshape(tf.decode_raw( features["ques_char_idxs"], tf.int32), [ques_limit, char_limit]) y1 = tf.reshape(tf.decode_raw( features["y1"], tf.float32), [para_limit]) y2 = tf.reshape(tf.decode_raw( features["y2"], tf.float32), [para_limit]) qa_id = features["id"] return context_idxs, ques_idxs, context_char_idxs, ques_char_idxs, y1, y2, qa_id return parse def get_batch_dataset(record_file, parser, config): """ Read a file and construct batches """ num_threads = tf.constant(config.num_threads, dtype=tf.int32) dataset = tf.data.TFRecordDataset(record_file).map( parser, num_parallel_calls=num_threads).shuffle(config.capacity).repeat() if config.is_bucket: buckets = [tf.constant(num) for num in range(*config.bucket_range)] def key_func(context_idxs, ques_idxs, context_char_idxs, ques_char_idxs, y1, y2, qa_id): c_len = tf.reduce_sum( tf.cast(tf.cast(context_idxs, tf.bool), tf.int32)) buckets_min = [np.iinfo(np.int32).min] + buckets buckets_max = buckets + [np.iinfo(np.int32).max] conditions_c = tf.logical_and( tf.less(buckets_min, c_len), tf.less_equal(c_len, buckets_max)) bucket_id = tf.reduce_min(tf.where(conditions_c)) return bucket_id def reduce_func(key, elements): return elements.batch(config.batch_size) dataset = dataset.apply(tf.contrib.data.group_by_window( key_func, reduce_func, window_size=5 * config.batch_size)).shuffle(len(buckets) * 25) else: dataset = dataset.batch(config.batch_size) return dataset def get_dataset(record_file, parser, config): num_threads = tf.constant(config.num_threads, dtype=tf.int32) dataset = tf.data.TFRecordDataset(record_file).map( parser, num_parallel_calls=num_threads).repeat().batch(config.batch_size) return dataset def convert_tokens(eval_file, qa_id, pp1, pp2): answer_dict = {} remapped_dict = {} for qid, p1, p2 in zip(qa_id, pp1, pp2): context = eval_file[str(qid)]["context"] spans = eval_file[str(qid)]["spans"] uuid = eval_file[str(qid)]["uuid"] start_idx = spans[p1][0] end_idx = spans[p2][1] answer_dict[str(qid)] = context[start_idx: end_idx] remapped_dict[uuid] = context[start_idx: end_idx] return answer_dict, remapped_dict def evaluate(eval_file, answer_dict): f1 = exact_match = total = 0 for key, value in answer_dict.items(): total += 1 ground_truths = eval_file[key]["answers"] prediction = value exact_match += metric_max_over_ground_truths( exact_match_score, prediction, ground_truths) f1 += metric_max_over_ground_truths(f1_score, prediction, ground_truths) exact_match = 100.0 * exact_match / total f1 = 100.0 * f1 / total return {'exact_match': exact_match, 'f1': f1} def normalize_answer(s): def remove_articles(text): return re.sub(r'\b(a|an|the)\b', ' ', text) def white_space_fix(text): return ' '.join(text.split()) def remove_punc(text): exclude = set(string.punctuation) return ''.join(ch for ch in text if ch not in exclude) def lower(text): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(s)))) def f1_score(prediction, ground_truth): prediction_tokens = normalize_answer(prediction).split() ground_truth_tokens = normalize_answer(ground_truth).split() common = Counter(prediction_tokens) & Counter(ground_truth_tokens) num_same = sum(common.values()) if num_same == 0: return 0 precision = 1.0 * num_same / len(prediction_tokens) recall = 1.0 * num_same / len(ground_truth_tokens) f1 = (2 * precision * recall) / (precision + recall) return f1 def exact_match_score(prediction, ground_truth): return (normalize_answer(prediction) == normalize_answer(ground_truth)) def metric_max_over_ground_truths(metric_fn, prediction, ground_truths): scores_for_ground_truths = [] for ground_truth in ground_truths: score = metric_fn(prediction, ground_truth) scores_for_ground_truths.append(score) return max(scores_for_ground_truths)
40.482993
102
0.623593
import tensorflow as tf import numpy as np import re from collections import Counter import string def get_record_parser(config, is_test=False): def parse(example): para_limit = config.test_para_limit if is_test else config.para_limit ques_limit = config.test_ques_limit if is_test else config.ques_limit char_limit = config.char_limit features = tf.parse_single_example(example, features={ "context_idxs": tf.FixedLenFeature([], tf.string), "ques_idxs": tf.FixedLenFeature([], tf.string), "context_char_idxs": tf.FixedLenFeature([], tf.string), "ques_char_idxs": tf.FixedLenFeature([], tf.string), "y1": tf.FixedLenFeature([], tf.string), "y2": tf.FixedLenFeature([], tf.string), "id": tf.FixedLenFeature([], tf.int64) }) context_idxs = tf.reshape(tf.decode_raw( features["context_idxs"], tf.int32), [para_limit]) ques_idxs = tf.reshape(tf.decode_raw( features["ques_idxs"], tf.int32), [ques_limit]) context_char_idxs = tf.reshape(tf.decode_raw( features["context_char_idxs"], tf.int32), [para_limit, char_limit]) ques_char_idxs = tf.reshape(tf.decode_raw( features["ques_char_idxs"], tf.int32), [ques_limit, char_limit]) y1 = tf.reshape(tf.decode_raw( features["y1"], tf.float32), [para_limit]) y2 = tf.reshape(tf.decode_raw( features["y2"], tf.float32), [para_limit]) qa_id = features["id"] return context_idxs, ques_idxs, context_char_idxs, ques_char_idxs, y1, y2, qa_id return parse def get_batch_dataset(record_file, parser, config): num_threads = tf.constant(config.num_threads, dtype=tf.int32) dataset = tf.data.TFRecordDataset(record_file).map( parser, num_parallel_calls=num_threads).shuffle(config.capacity).repeat() if config.is_bucket: buckets = [tf.constant(num) for num in range(*config.bucket_range)] def key_func(context_idxs, ques_idxs, context_char_idxs, ques_char_idxs, y1, y2, qa_id): c_len = tf.reduce_sum( tf.cast(tf.cast(context_idxs, tf.bool), tf.int32)) buckets_min = [np.iinfo(np.int32).min] + buckets buckets_max = buckets + [np.iinfo(np.int32).max] conditions_c = tf.logical_and( tf.less(buckets_min, c_len), tf.less_equal(c_len, buckets_max)) bucket_id = tf.reduce_min(tf.where(conditions_c)) return bucket_id def reduce_func(key, elements): return elements.batch(config.batch_size) dataset = dataset.apply(tf.contrib.data.group_by_window( key_func, reduce_func, window_size=5 * config.batch_size)).shuffle(len(buckets) * 25) else: dataset = dataset.batch(config.batch_size) return dataset def get_dataset(record_file, parser, config): num_threads = tf.constant(config.num_threads, dtype=tf.int32) dataset = tf.data.TFRecordDataset(record_file).map( parser, num_parallel_calls=num_threads).repeat().batch(config.batch_size) return dataset def convert_tokens(eval_file, qa_id, pp1, pp2): answer_dict = {} remapped_dict = {} for qid, p1, p2 in zip(qa_id, pp1, pp2): context = eval_file[str(qid)]["context"] spans = eval_file[str(qid)]["spans"] uuid = eval_file[str(qid)]["uuid"] start_idx = spans[p1][0] end_idx = spans[p2][1] answer_dict[str(qid)] = context[start_idx: end_idx] remapped_dict[uuid] = context[start_idx: end_idx] return answer_dict, remapped_dict def evaluate(eval_file, answer_dict): f1 = exact_match = total = 0 for key, value in answer_dict.items(): total += 1 ground_truths = eval_file[key]["answers"] prediction = value exact_match += metric_max_over_ground_truths( exact_match_score, prediction, ground_truths) f1 += metric_max_over_ground_truths(f1_score, prediction, ground_truths) exact_match = 100.0 * exact_match / total f1 = 100.0 * f1 / total return {'exact_match': exact_match, 'f1': f1} def normalize_answer(s): def remove_articles(text): return re.sub(r'\b(a|an|the)\b', ' ', text) def white_space_fix(text): return ' '.join(text.split()) def remove_punc(text): exclude = set(string.punctuation) return ''.join(ch for ch in text if ch not in exclude) def lower(text): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(s)))) def f1_score(prediction, ground_truth): prediction_tokens = normalize_answer(prediction).split() ground_truth_tokens = normalize_answer(ground_truth).split() common = Counter(prediction_tokens) & Counter(ground_truth_tokens) num_same = sum(common.values()) if num_same == 0: return 0 precision = 1.0 * num_same / len(prediction_tokens) recall = 1.0 * num_same / len(ground_truth_tokens) f1 = (2 * precision * recall) / (precision + recall) return f1 def exact_match_score(prediction, ground_truth): return (normalize_answer(prediction) == normalize_answer(ground_truth)) def metric_max_over_ground_truths(metric_fn, prediction, ground_truths): scores_for_ground_truths = [] for ground_truth in ground_truths: score = metric_fn(prediction, ground_truth) scores_for_ground_truths.append(score) return max(scores_for_ground_truths)
true
true
1c48620c64fa51850f5fd4bc16ab11b4b2f1dfac
4,735
py
Python
example.py
lbenassi/InstagramBot
49f8bad5de8d5df719f102c66acb6779b677bc5f
[ "MIT" ]
1
2019-08-05T23:02:58.000Z
2019-08-05T23:02:58.000Z
example.py
lbenassi/InstagramBot
49f8bad5de8d5df719f102c66acb6779b677bc5f
[ "MIT" ]
null
null
null
example.py
lbenassi/InstagramBot
49f8bad5de8d5df719f102c66acb6779b677bc5f
[ "MIT" ]
null
null
null
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import time from src import InstaBot from src.check_status import check_status from src.feed_scanner import feed_scanner from src.follow_protocol import follow_protocol from src.unfollow_protocol import unfollow_protocol bot = InstaBot( login="", password="", like_per_day=600, comments_per_day=200, tag_list=['follow4follow', 'f4f', 'cute', 'l:212999109','party','time','luxary','happy','birthday','robatkarim','eshq','dostdashtani','boro','dawshi','nab','love','party','smile','style','eslamshahr','parand','like','dawshi','lovely','instagramers','socialenvy','اینستاگرام','ایران','فالوور','لایک','لاکچری','بام تهران','کرج','محرم','تاسوعا','عاشورا','مشهد','عکس','عکاسی','ویدیو','فالوبک','photographer','Desine','karaj','teh','tehran','mashhad','robatkarim','parand','پرند','park','خمینی','esfahan','arak','parkmoalem','پارک معلم','قلعه حسن خان','تجریش','دربند'], tag_blacklist=['rain', 'thunderstorm'], user_blacklist={}, max_like_for_one_tag=50, follow_per_day=700, follow_time=1 * 12, unfollow_per_day=300, unfollow_break_min=15, unfollow_break_max=30, log_mod=0, proxy='', # List of list of words, each of which will be used to generate comment # For example: "This shot feels wow!" comment_list=[["this", "the", "your"], ["photo", "picture", "pic", "shot", "snapshot"], ["is", "looks", "feels", "is really"], ["great", "super", "good", "very good", "good", "wow", "WOW", "cool", "GREAT","magnificent", "magical", "very cool", "stylish", "beautiful", "so beautiful", "so stylish", "so professional", "lovely", "so lovely", "very lovely", "glorious","so glorious", "very glorious", "jazab", "excellent", "amazing"], [".", "..", "...", "!", "!!", "!!!"]], # Use unwanted_username_list to block usernames containing a string ## Will do partial matches; i.e. 'mozart' will block 'legend_mozart' ### 'free_followers' will be blocked because it contains 'free' unwanted_username_list=[ 'second', 'stuff', 'art', 'project', 'love', 'life', 'food', 'blog', 'free', 'keren', 'photo', 'graphy', 'indo', 'travel', 'art', 'shop', 'store', 'sex', 'toko', 'jual', 'online', 'murah', 'jam', 'kaos', 'case', 'baju', 'fashion', 'corp', 'tas', 'butik', 'grosir', 'karpet', 'sosis', 'salon', 'skin', 'care', 'cloth', 'tech', 'rental', 'kamera', 'beauty', 'express', 'kredit', 'collection', 'impor', 'preloved', 'follow', 'follower', 'gain', '.id', '_id', 'bags' ], unfollow_whitelist=['example_user_1', 'example_user_2']) while True: #print("# MODE 0 = ORIGINAL MODE BY LEVPASHA") #print("## MODE 1 = MODIFIED MODE BY KEMONG") #print("### MODE 2 = ORIGINAL MODE + UNFOLLOW WHO DON'T FOLLOW BACK") #print("#### MODE 3 = MODIFIED MODE : UNFOLLOW USERS WHO DON'T FOLLOW YOU BASED ON RECENT FEED") #print("##### MODE 4 = MODIFIED MODE : FOLLOW USERS BASED ON RECENT FEED ONLY") #print("###### MODE 5 = MODIFIED MODE : JUST UNFOLLOW EVERYBODY, EITHER YOUR FOLLOWER OR NOT") ################################ ## WARNING ### ################################ # DON'T USE MODE 5 FOR A LONG PERIOD. YOU RISK YOUR ACCOUNT FROM GETTING BANNED ## USE MODE 5 IN BURST MODE, USE IT TO UNFOLLOW PEOPLE AS MANY AS YOU WANT IN SHORT TIME PERIOD mode = 0 #print("You choose mode : %i" %(mode)) #print("CTRL + C to cancel this operation or wait 30 seconds to start") #time.sleep(30) if mode == 0: bot.new_auto_mod() elif mode == 1: check_status(bot) while bot.self_following - bot.self_follower > 200: unfollow_protocol(bot) time.sleep(10 * 60) check_status(bot) while bot.self_following - bot.self_follower < 400: while len(bot.user_info_list) < 50: feed_scanner(bot) time.sleep(5 * 60) follow_protocol(bot) time.sleep(10 * 60) check_status(bot) elif mode == 2: bot.bot_mode = 1 bot.new_auto_mod() elif mode == 3: unfollow_protocol(bot) time.sleep(10 * 60) elif mode == 4: feed_scanner(bot) time.sleep(60) follow_protocol(bot) time.sleep(10 * 60) elif mode == 5: bot.bot_mode = 2 unfollow_protocol(bot) else: print("Wrong mode!")
42.276786
569
0.563675
import os import time from src import InstaBot from src.check_status import check_status from src.feed_scanner import feed_scanner from src.follow_protocol import follow_protocol from src.unfollow_protocol import unfollow_protocol bot = InstaBot( login="", password="", like_per_day=600, comments_per_day=200, tag_list=['follow4follow', 'f4f', 'cute', 'l:212999109','party','time','luxary','happy','birthday','robatkarim','eshq','dostdashtani','boro','dawshi','nab','love','party','smile','style','eslamshahr','parand','like','dawshi','lovely','instagramers','socialenvy','اینستاگرام','ایران','فالوور','لایک','لاکچری','بام تهران','کرج','محرم','تاسوعا','عاشورا','مشهد','عکس','عکاسی','ویدیو','فالوبک','photographer','Desine','karaj','teh','tehran','mashhad','robatkarim','parand','پرند','park','خمینی','esfahan','arak','parkmoalem','پارک معلم','قلعه حسن خان','تجریش','دربند'], tag_blacklist=['rain', 'thunderstorm'], user_blacklist={}, max_like_for_one_tag=50, follow_per_day=700, follow_time=1 * 12, unfollow_per_day=300, unfollow_break_min=15, unfollow_break_max=30, log_mod=0, proxy='', comment_list=[["this", "the", "your"], ["photo", "picture", "pic", "shot", "snapshot"], ["is", "looks", "feels", "is really"], ["great", "super", "good", "very good", "good", "wow", "WOW", "cool", "GREAT","magnificent", "magical", "very cool", "stylish", "beautiful", "so beautiful", "so stylish", "so professional", "lovely", "so lovely", "very lovely", "glorious","so glorious", "very glorious", "jazab", "excellent", "amazing"], [".", "..", "...", "!", "!!", "!!!"]], 'store', 'sex', 'toko', 'jual', 'online', 'murah', 'jam', 'kaos', 'case', 'baju', 'fashion', 'corp', 'tas', 'butik', 'grosir', 'karpet', 'sosis', 'salon', 'skin', 'care', 'cloth', 'tech', 'rental', 'kamera', 'beauty', 'express', 'kredit', 'collection', 'impor', 'preloved', 'follow', 'follower', 'gain', '.id', '_id', 'bags' ], unfollow_whitelist=['example_user_1', 'example_user_2']) while True: #print("#### MODE 3 = MODIFIED MODE : UNFOLLOW USERS WHO DON'T FOLLOW YOU BASED ON RECENT FEED") protocol(bot) time.sleep(10 * 60) elif mode == 5: bot.bot_mode = 2 unfollow_protocol(bot) else: print("Wrong mode!")
true
true
1c4862fb07b2c9db29a9081ca5087a83b0ba2309
1,396
py
Python
python2_guiding_test.py
ammumaddy/dhivya-railway
152a64e16ba07d62aa9aa159f503ed0b1a09d5b6
[ "MIT" ]
97
2015-01-02T10:58:05.000Z
2022-03-11T14:00:52.000Z
python2_guiding_test.py
ammumaddy/dhivya-railway
152a64e16ba07d62aa9aa159f503ed0b1a09d5b6
[ "MIT" ]
3
2020-02-14T15:55:21.000Z
2020-02-19T17:33:05.000Z
python2_guiding_test.py
ammumaddy/dhivya-railway
152a64e16ba07d62aa9aa159f503ed0b1a09d5b6
[ "MIT" ]
58
2015-05-28T02:09:51.000Z
2022-03-20T16:37:40.000Z
""" Equivalent of 'guiding_test.py' except for Python2.x, which comes as standard on many systems. Run it with: python python2_guiding_test.py """ import json import subprocess import unittest import os import urllib2, urllib url = "http://127.0.0.1:8083" interpreter = "python" reservation_script = os.path.join("python", "reserve.py") class TrainReservationTest(unittest.TestCase): def test_reserve_seats_via_POST(self): form_data = {"train_id": "express_2000", "seat_count": 4} data = urllib.urlencode(form_data) response = urllib2.urlopen(url + "/reserve", data=data).read() reservation = json.loads(response) assert "express_2000" == reservation["train_id"] assert 4 == len(reservation["seats"]) assert "1A" == reservation["seats"][0] assert "75bcd15" == reservation["booking_reference"] def test_reserve_seats_via_cmd(self): response = subprocess.check_output([interpreter, reservation_script, "express2000", "4"], stderr=subprocess.STDOUT, universal_newlines = True) reservation = json.loads(response) assert "express_2000" == reservation["train_id"] assert 4 == len(reservation["seats"]) assert "1A" == reservation["seats"][0] assert "75bcd15" == reservation["booking_reference"] if __name__ == "__main__": unittest.main()
30.347826
150
0.670487
import json import subprocess import unittest import os import urllib2, urllib url = "http://127.0.0.1:8083" interpreter = "python" reservation_script = os.path.join("python", "reserve.py") class TrainReservationTest(unittest.TestCase): def test_reserve_seats_via_POST(self): form_data = {"train_id": "express_2000", "seat_count": 4} data = urllib.urlencode(form_data) response = urllib2.urlopen(url + "/reserve", data=data).read() reservation = json.loads(response) assert "express_2000" == reservation["train_id"] assert 4 == len(reservation["seats"]) assert "1A" == reservation["seats"][0] assert "75bcd15" == reservation["booking_reference"] def test_reserve_seats_via_cmd(self): response = subprocess.check_output([interpreter, reservation_script, "express2000", "4"], stderr=subprocess.STDOUT, universal_newlines = True) reservation = json.loads(response) assert "express_2000" == reservation["train_id"] assert 4 == len(reservation["seats"]) assert "1A" == reservation["seats"][0] assert "75bcd15" == reservation["booking_reference"] if __name__ == "__main__": unittest.main()
true
true
1c4863108b4acd15cabb6c18697226bdc4fae51a
7,174
py
Python
causallib/contrib/tests/test_shared_sparsity_selection.py
liranszlak/causallib
2636149f6b1e307672aff638a53f8eaf2be56bc9
[ "Apache-2.0" ]
350
2019-06-19T15:56:19.000Z
2022-03-28T23:47:46.000Z
causallib/contrib/tests/test_shared_sparsity_selection.py
liranszlak/causallib
2636149f6b1e307672aff638a53f8eaf2be56bc9
[ "Apache-2.0" ]
13
2019-08-14T22:04:21.000Z
2022-03-14T07:44:12.000Z
causallib/contrib/tests/test_shared_sparsity_selection.py
liranszlak/causallib
2636149f6b1e307672aff638a53f8eaf2be56bc9
[ "Apache-2.0" ]
48
2019-11-02T16:40:56.000Z
2022-02-09T12:55:12.000Z
import numpy as np import pandas as pd from sklearn.datasets import make_classification from sklearn.preprocessing import StandardScaler from sklearn.exceptions import ConvergenceWarning from causallib.contrib.shared_sparsity_selection import SharedSparsityConfounderSelection from causallib.tests.test_confounder_selection import _TestConfounderSelection class TestSharedSparsitySelection(_TestConfounderSelection): def make_xay(self, n_confounders_a, n_max_confounders_y, n_samples, xay_cols=10, seed=None): # rng = np.random.default_rng(seed) if seed: np.random.seed(seed) X, a = make_classification( n_samples=n_samples, n_features=xay_cols + 1, n_informative=int(min(n_confounders_a, xay_cols)), n_redundant=0, n_repeated=0, class_sep=10.0, n_clusters_per_class=1, shuffle=False, # random_state=seed ) y_confounder_indicator = np.zeros(X.shape[1], dtype=bool) y_confounder_indicator[:int(min(n_max_confounders_y, xay_cols))] = 1 np.random.shuffle(y_confounder_indicator) y = X[:, y_confounder_indicator] @ np.random.normal(size=y_confounder_indicator.sum()) X = StandardScaler().fit_transform(X) X = pd.DataFrame(X, columns=["x_" + str(i) for i in range(X.shape[1])]) a = pd.Series(a) y = pd.Series(y) return X, a, y def test_covariate_subset(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) true_subset_confounders = ['x_0', 'x_2'] # Matches random seed: 6 covariates_subset = ['x_0', 'x_2', f'x_{X.shape[1] - 1}', f'x_{X.shape[1] - 3}'] sss = SharedSparsityConfounderSelection(covariates=covariates_subset) sss = self.ensure_covariate_subset(sss, X, a, y, true_subset_confounders) np.testing.assert_array_equal(covariates_subset, sss.covariates) self.assertEqual(len(covariates_subset), sss.selector_.theta_.shape[0]) self.assertEqual(2, sss.selector_.theta_.shape[1]) # Two treatments self.assertEqual(len(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 0]) > 0)) self.assertEqual(len(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 1]) > 0)) def test_covariate_subset_binary(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) true_subset_confounders = ['x_0', 'x_2'] # Matches random seed: 6 covariates_subset = ['x_0', 'x_2', f'x_{X.shape[1] - 1}', f'x_{X.shape[1] - 3}'] # Convert to binary: true_subset_confounders = X.columns.isin(true_subset_confounders) covariates_subset = X.columns.isin(covariates_subset) sss = SharedSparsityConfounderSelection(covariates=covariates_subset) sss = self.ensure_covariate_subset_binary(sss, X, a, y, true_subset_confounders) np.testing.assert_array_equal(covariates_subset, sss.covariates) self.assertEqual(covariates_subset.sum(), sss.selector_.theta_.shape[0]) self.assertEqual(2, sss.selector_.theta_.shape[1]) # Two treatments self.assertEqual(sum(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 0]) > 0)) self.assertEqual(sum(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 1]) > 0)) def test_alphas(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) alphas = [0, 1] for alpha in alphas: sss = SharedSparsityConfounderSelection(mcp_alpha=alpha) sss.fit(X, a, y) Xt = sss.transform(X) self.assertSetEqual(set(Xt.columns), {'x_0', 'x_2'}) with self.assertRaises(AssertionError): sss = SharedSparsityConfounderSelection(mcp_alpha=-1) sss.fit(X, a, y) with self.subTest("shrinkage"): strong = SharedSparsityConfounderSelection(mcp_alpha=0.1).fit(X, a, y).selector_.theta_ weak = SharedSparsityConfounderSelection(mcp_alpha=100).fit(X, a, y).selector_.theta_ self.assertLess(np.linalg.norm(strong), np.linalg.norm(weak)) def test_lambdas(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) with self.subTest("Automatic (default) lambda"): sss = SharedSparsityConfounderSelection(mcp_lambda="auto") sss.fit(X, a, y) expected = 0.2 * np.sqrt(2 * np.log(X.shape[1]) / (X.shape[0] / 2)) self.assertAlmostEqual(sss.selector_.lmda_, expected) with self.subTest("Pre-specified lambda"): lmda = 2.1 sss = SharedSparsityConfounderSelection(mcp_lambda=lmda) sss.fit(X, a, y) self.assertEqual(sss.selector_.lmda_, lmda) with self.subTest("Illegal lambda"): with self.assertRaises(AssertionError): sss = SharedSparsityConfounderSelection(mcp_lambda=-1) sss.fit(X, a, y) with self.subTest("shrinkage"): weak = SharedSparsityConfounderSelection(mcp_lambda=0.1).fit(X, a, y).selector_.theta_ strong = SharedSparsityConfounderSelection(mcp_lambda=1).fit(X, a, y).selector_.theta_ self.assertLess(np.linalg.norm(strong), np.linalg.norm(weak)) def test_max_iter(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) with self.subTest("Force convergence warning"): sss = SharedSparsityConfounderSelection(max_iter=2) with self.assertWarns(ConvergenceWarning): sss.fit(X, a, y) # with self.subTest("Convergence happens in less than max_iter"): # import timeit # n_repeats = 50 # times = [] # for max_iter in [10000, 100000]: # # Algorithm will converge long before exceeding `max_iter` and so time should remain similar # sss = SharedSparsityConfounderSelection(max_iter=max_iter) # avg_time = timeit.timeit(lambda: sss.fit(X, a, y), number=n_repeats) # times.append(avg_time) # self.assertAlmostEqual(times[0], times[1], places=1) def test_final_selection(self): """Test against current implementation to allow for refactoring""" X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) sss = SharedSparsityConfounderSelection() sss.fit(X, a, y) Xt = sss.transform(X) self.assertSetEqual(set(Xt.columns), {'x_0', 'x_2'}) def test_importance_getter(self): from causallib.preprocessing.confounder_selection import _get_feature_importances X, a, y = self.make_xay(2, 2, xay_cols=2, n_samples=100, seed=1) sss = SharedSparsityConfounderSelection() sss.fit(X, a, y) importance = _get_feature_importances(sss, sss.importance_getter) expected = np.array([[0.0, 0.0], [5.86299046, 5.94375083], [0.0, 0.0] ]) np.testing.assert_array_almost_equal(expected.transpose(), importance) np.testing.assert_array_almost_equal(sss.selector_.theta_.transpose(), importance)
47.197368
110
0.64678
import numpy as np import pandas as pd from sklearn.datasets import make_classification from sklearn.preprocessing import StandardScaler from sklearn.exceptions import ConvergenceWarning from causallib.contrib.shared_sparsity_selection import SharedSparsityConfounderSelection from causallib.tests.test_confounder_selection import _TestConfounderSelection class TestSharedSparsitySelection(_TestConfounderSelection): def make_xay(self, n_confounders_a, n_max_confounders_y, n_samples, xay_cols=10, seed=None): if seed: np.random.seed(seed) X, a = make_classification( n_samples=n_samples, n_features=xay_cols + 1, n_informative=int(min(n_confounders_a, xay_cols)), n_redundant=0, n_repeated=0, class_sep=10.0, n_clusters_per_class=1, shuffle=False, ) y_confounder_indicator = np.zeros(X.shape[1], dtype=bool) y_confounder_indicator[:int(min(n_max_confounders_y, xay_cols))] = 1 np.random.shuffle(y_confounder_indicator) y = X[:, y_confounder_indicator] @ np.random.normal(size=y_confounder_indicator.sum()) X = StandardScaler().fit_transform(X) X = pd.DataFrame(X, columns=["x_" + str(i) for i in range(X.shape[1])]) a = pd.Series(a) y = pd.Series(y) return X, a, y def test_covariate_subset(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) true_subset_confounders = ['x_0', 'x_2'] covariates_subset = ['x_0', 'x_2', f'x_{X.shape[1] - 1}', f'x_{X.shape[1] - 3}'] sss = SharedSparsityConfounderSelection(covariates=covariates_subset) sss = self.ensure_covariate_subset(sss, X, a, y, true_subset_confounders) np.testing.assert_array_equal(covariates_subset, sss.covariates) self.assertEqual(len(covariates_subset), sss.selector_.theta_.shape[0]) self.assertEqual(2, sss.selector_.theta_.shape[1]) self.assertEqual(len(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 0]) > 0)) self.assertEqual(len(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 1]) > 0)) def test_covariate_subset_binary(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) true_subset_confounders = ['x_0', 'x_2'] covariates_subset = ['x_0', 'x_2', f'x_{X.shape[1] - 1}', f'x_{X.shape[1] - 3}'] true_subset_confounders = X.columns.isin(true_subset_confounders) covariates_subset = X.columns.isin(covariates_subset) sss = SharedSparsityConfounderSelection(covariates=covariates_subset) sss = self.ensure_covariate_subset_binary(sss, X, a, y, true_subset_confounders) np.testing.assert_array_equal(covariates_subset, sss.covariates) self.assertEqual(covariates_subset.sum(), sss.selector_.theta_.shape[0]) self.assertEqual(2, sss.selector_.theta_.shape[1]) self.assertEqual(sum(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 0]) > 0)) self.assertEqual(sum(true_subset_confounders), np.sum(np.abs(sss.selector_.theta_[:, 1]) > 0)) def test_alphas(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) alphas = [0, 1] for alpha in alphas: sss = SharedSparsityConfounderSelection(mcp_alpha=alpha) sss.fit(X, a, y) Xt = sss.transform(X) self.assertSetEqual(set(Xt.columns), {'x_0', 'x_2'}) with self.assertRaises(AssertionError): sss = SharedSparsityConfounderSelection(mcp_alpha=-1) sss.fit(X, a, y) with self.subTest("shrinkage"): strong = SharedSparsityConfounderSelection(mcp_alpha=0.1).fit(X, a, y).selector_.theta_ weak = SharedSparsityConfounderSelection(mcp_alpha=100).fit(X, a, y).selector_.theta_ self.assertLess(np.linalg.norm(strong), np.linalg.norm(weak)) def test_lambdas(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) with self.subTest("Automatic (default) lambda"): sss = SharedSparsityConfounderSelection(mcp_lambda="auto") sss.fit(X, a, y) expected = 0.2 * np.sqrt(2 * np.log(X.shape[1]) / (X.shape[0] / 2)) self.assertAlmostEqual(sss.selector_.lmda_, expected) with self.subTest("Pre-specified lambda"): lmda = 2.1 sss = SharedSparsityConfounderSelection(mcp_lambda=lmda) sss.fit(X, a, y) self.assertEqual(sss.selector_.lmda_, lmda) with self.subTest("Illegal lambda"): with self.assertRaises(AssertionError): sss = SharedSparsityConfounderSelection(mcp_lambda=-1) sss.fit(X, a, y) with self.subTest("shrinkage"): weak = SharedSparsityConfounderSelection(mcp_lambda=0.1).fit(X, a, y).selector_.theta_ strong = SharedSparsityConfounderSelection(mcp_lambda=1).fit(X, a, y).selector_.theta_ self.assertLess(np.linalg.norm(strong), np.linalg.norm(weak)) def test_max_iter(self): X, a, y = self.make_xay(6, 4, n_samples=100, seed=1) with self.subTest("Force convergence warning"): sss = SharedSparsityConfounderSelection(max_iter=2) with self.assertWarns(ConvergenceWarning): sss.fit(X, a, y) self.make_xay(6, 4, n_samples=100, seed=1) sss = SharedSparsityConfounderSelection() sss.fit(X, a, y) Xt = sss.transform(X) self.assertSetEqual(set(Xt.columns), {'x_0', 'x_2'}) def test_importance_getter(self): from causallib.preprocessing.confounder_selection import _get_feature_importances X, a, y = self.make_xay(2, 2, xay_cols=2, n_samples=100, seed=1) sss = SharedSparsityConfounderSelection() sss.fit(X, a, y) importance = _get_feature_importances(sss, sss.importance_getter) expected = np.array([[0.0, 0.0], [5.86299046, 5.94375083], [0.0, 0.0] ]) np.testing.assert_array_almost_equal(expected.transpose(), importance) np.testing.assert_array_almost_equal(sss.selector_.theta_.transpose(), importance)
true
true
1c4864c7568edd42683a2109677a37b005cc8076
48
py
Python
sortedm2m/__init__.py
Freston5/daysiweb
95751b467f0e76c3cb60bb09693c59af9c74ded2
[ "MIT" ]
null
null
null
sortedm2m/__init__.py
Freston5/daysiweb
95751b467f0e76c3cb60bb09693c59af9c74ded2
[ "MIT" ]
null
null
null
sortedm2m/__init__.py
Freston5/daysiweb
95751b467f0e76c3cb60bb09693c59af9c74ded2
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- __version__ = '1.3.3'
9.6
23
0.5
__version__ = '1.3.3'
true
true
1c4865911cd82746699b01afdfe934853aeba6b9
3,337
py
Python
selenium_pipeline/hyatt_hotels_fetch_addresses.py
Praneethvvs/CircleCi_FastApi
0aec14fcffcfe7053cf7db688728347feea26f70
[ "MIT" ]
null
null
null
selenium_pipeline/hyatt_hotels_fetch_addresses.py
Praneethvvs/CircleCi_FastApi
0aec14fcffcfe7053cf7db688728347feea26f70
[ "MIT" ]
null
null
null
selenium_pipeline/hyatt_hotels_fetch_addresses.py
Praneethvvs/CircleCi_FastApi
0aec14fcffcfe7053cf7db688728347feea26f70
[ "MIT" ]
null
null
null
import time import pandas as pd from selenium import webdriver from selenium.webdriver.common.by import By from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.support import expected_conditions as EC from selenium.webdriver.common.desired_capabilities import DesiredCapabilities import traceback import itertools from selenium.webdriver.common.keys import Keys DRIVER_PATH = r"C:\Program Files (x86)\chromedriver.exe" class Address_Scraping(): def __init__(self): self.chrome_driver = webdriver.Chrome(DRIVER_PATH) def get_hyperlinks(self): self.chrome_driver.get("https://www.hyatt.com/explore-hotels") try: # The WebDriverWait method waits until it locates the presence of the element" WebDriverWait(self.chrome_driver, 20).until( EC.presence_of_element_located((By.CLASS_NAME, "countries.b-ph0"))) us_add = self.chrome_driver.find_element_by_xpath( "//ul[@class='countries b-ph0']//li[@data-js-country='United States']") links = us_add.find_elements_by_tag_name('a') hyperlinks = [link_field.get_attribute("href") for link_field in links] return hyperlinks except: print("error") traceback.print_exc() time.sleep(2) # chrome_driver.quit() def fetch_addresses_to_df(self): links_list = self.get_hyperlinks() # assert links_list != [] results_list = [] error_links_list = [] for index, link in enumerate(links_list, start=1): if index == 5: break try: print("passing through link ------------>", link) self.chrome_driver.get(link) address_div = self.chrome_driver.find_elements_by_xpath( "//div[@class='site-info-container b-mt2 b-mb2 b-mt0@sm b-mb0@sm']//a[@class='site-info-address b-d-inline-block b-d-flex@lg b-d-inline-block@xl b-mb2@sm b-mb1@md b-mr2']//span[@class='b-d-inline-block']") phone_num_div = self.chrome_driver.find_element_by_xpath( "//div[@class='site-info-container b-mt2 b-mb2 b-mt0@sm b-mb0@sm']//a[@class='site-info-phone b-d-inline-block b-d-block@lg b-mb1@sm b-mr2']//span[@class='hover-border b-d-none b-d-inline@lg']") address = "".join(map(lambda x: x.text, address_div)) phone_number = ", " + phone_num_div.text # self.chrome_driver.find_element_by_partial_link_text("Hoover, Alabama, United States, 35244").click() # time.sleep(3) # self.chrome_driver.close() # get_url = self.chrome_driver.current_url # print(get_url) # exit() combined_output = "".join([address, phone_number]) results_list.append(combined_output.split(",")) except: traceback.print_exc() error_links_list.append(link) final_df = pd.DataFrame(results_list, columns=["street", "city", "state", "country", "zip", "phone_number"], index=None) final_df.to_excel("hyatt_hotels.xlsx", index=False) if __name__ == "__main__": Address_Scraping().fetch_addresses_to_df()
40.695122
225
0.620617
import time import pandas as pd from selenium import webdriver from selenium.webdriver.common.by import By from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.support import expected_conditions as EC from selenium.webdriver.common.desired_capabilities import DesiredCapabilities import traceback import itertools from selenium.webdriver.common.keys import Keys DRIVER_PATH = r"C:\Program Files (x86)\chromedriver.exe" class Address_Scraping(): def __init__(self): self.chrome_driver = webdriver.Chrome(DRIVER_PATH) def get_hyperlinks(self): self.chrome_driver.get("https://www.hyatt.com/explore-hotels") try: WebDriverWait(self.chrome_driver, 20).until( EC.presence_of_element_located((By.CLASS_NAME, "countries.b-ph0"))) us_add = self.chrome_driver.find_element_by_xpath( "//ul[@class='countries b-ph0']//li[@data-js-country='United States']") links = us_add.find_elements_by_tag_name('a') hyperlinks = [link_field.get_attribute("href") for link_field in links] return hyperlinks except: print("error") traceback.print_exc() time.sleep(2) # chrome_driver.quit() def fetch_addresses_to_df(self): links_list = self.get_hyperlinks() # assert links_list != [] results_list = [] error_links_list = [] for index, link in enumerate(links_list, start=1): if index == 5: break try: print("passing through link ------------>", link) self.chrome_driver.get(link) address_div = self.chrome_driver.find_elements_by_xpath( "//div[@class='site-info-container b-mt2 b-mb2 b-mt0@sm b-mb0@sm']//a[@class='site-info-address b-d-inline-block b-d-flex@lg b-d-inline-block@xl b-mb2@sm b-mb1@md b-mr2']//span[@class='b-d-inline-block']") phone_num_div = self.chrome_driver.find_element_by_xpath( "//div[@class='site-info-container b-mt2 b-mb2 b-mt0@sm b-mb0@sm']//a[@class='site-info-phone b-d-inline-block b-d-block@lg b-mb1@sm b-mr2']//span[@class='hover-border b-d-none b-d-inline@lg']") address = "".join(map(lambda x: x.text, address_div)) phone_number = ", " + phone_num_div.text # self.chrome_driver.find_element_by_partial_link_text("Hoover, Alabama, United States, 35244").click() # time.sleep(3) # self.chrome_driver.close() # get_url = self.chrome_driver.current_url # print(get_url) # exit() combined_output = "".join([address, phone_number]) results_list.append(combined_output.split(",")) except: traceback.print_exc() error_links_list.append(link) final_df = pd.DataFrame(results_list, columns=["street", "city", "state", "country", "zip", "phone_number"], index=None) final_df.to_excel("hyatt_hotels.xlsx", index=False) if __name__ == "__main__": Address_Scraping().fetch_addresses_to_df()
true
true
1c4866945a2ab251b49fcecdc203276c56d51585
379
py
Python
Day28/Solution.py
MarceloKabbalah/30-Days-Of-Code
094de037347b00105c5385add9de7bf605277e16
[ "MIT" ]
null
null
null
Day28/Solution.py
MarceloKabbalah/30-Days-Of-Code
094de037347b00105c5385add9de7bf605277e16
[ "MIT" ]
null
null
null
Day28/Solution.py
MarceloKabbalah/30-Days-Of-Code
094de037347b00105c5385add9de7bf605277e16
[ "MIT" ]
null
null
null
#!/bin/python # compatible with python3 import sys import re N = int(input().strip()) names = [] for a0 in range(N): firstName,emailID = input().strip().split(' ') firstName,emailID = [str(firstName),str(emailID)] match = re.search(r'[\w\.-]+@gmail.com', emailID) if match: names.append(firstName) names.sort() for name in names: print( name )
19.947368
53
0.62533
import sys import re N = int(input().strip()) names = [] for a0 in range(N): firstName,emailID = input().strip().split(' ') firstName,emailID = [str(firstName),str(emailID)] match = re.search(r'[\w\.-]+@gmail.com', emailID) if match: names.append(firstName) names.sort() for name in names: print( name )
true
true
1c486728773d0fecc3487bf43c095f48ffa913bc
1,645
py
Python
detailsScrape/oilymoistd/oilymoistd17.py
Asyikin98/SkinFerm
72fd1ad6339c96adf5ec154bde566de9eb1472c3
[ "MIT" ]
null
null
null
detailsScrape/oilymoistd/oilymoistd17.py
Asyikin98/SkinFerm
72fd1ad6339c96adf5ec154bde566de9eb1472c3
[ "MIT" ]
2
2021-02-03T01:55:13.000Z
2021-04-30T12:46:33.000Z
detailsScrape/oilymoistd/oilymoistd17.py
Asyikin98/SkinFerm
72fd1ad6339c96adf5ec154bde566de9eb1472c3
[ "MIT" ]
null
null
null
import urllib.request import random from bs4 import BeautifulSoup from requests import get import mysql.connector conn = mysql.connector.connect(user="root", passwd="",host="localhost", database="product") cursor = conn.cursor() sql = """INSERT INTO oilymoistd (about, rate, top, comment, dari) VALUES (%s, %s, %s, %s, %s)""" def crawl_url(pageUrl, moistoilyd_arr): url = 'https://www.skinstore.com/high-expectations-cannabis-facial-oil-32-cannabis-sativa-seed-oil-1-oz-30ml/12289881.html' page = get(url) soup = BeautifulSoup(page.text, 'html.parser') type(soup) #######################################################for product 1############################################################################ moist = soup.find_all('div', class_='primary-wrap column-row') try: for moistd in moist : about = moistd.find("div",{"class":"productDescription_synopsisContent"}).get_text().strip() rate = moistd.find("span",{"class":"visually-hidden productReviews_aggregateRating_hiddenLabel"}).get_text().strip() top = moistd.find("h2",{"class":"productReviews_topReviewsTitle"}).get_text().strip() comment = moistd.find("p",{"class":"productReviews_topReviewsExcerpt"}).get_text().strip() dari = moistd.find("div",{"class":"productReviews_footerDateAndName"}).get_text().strip() moistoilyd_arr.append((about, rate, top, comment, dari)) finally: return moistoilyd_arr moistoilyd_arr = crawl_url("", []) print(len(moistoilyd_arr)) cursor.executemany(sql, moistoilyd_arr) conn.commit() cursor.close() conn.close()
36.555556
148
0.630395
import urllib.request import random from bs4 import BeautifulSoup from requests import get import mysql.connector conn = mysql.connector.connect(user="root", passwd="",host="localhost", database="product") cursor = conn.cursor() sql = """INSERT INTO oilymoistd (about, rate, top, comment, dari) VALUES (%s, %s, %s, %s, %s)""" def crawl_url(pageUrl, moistoilyd_arr): url = 'https://www.skinstore.com/high-expectations-cannabis-facial-oil-32-cannabis-sativa-seed-oil-1-oz-30ml/12289881.html' page = get(url) soup = BeautifulSoup(page.text, 'html.parser') type(soup)
true
true
1c486733754ce24861ac7025b7b44eb64a9b0479
742
py
Python
api/endpoints/fruit/get.py
DarkbordermanTemplate/fastapi-redis-sqlalchemy
80fbdc419b19592b08bc2227c9d7c2925b7b91e2
[ "BSD-2-Clause" ]
5
2021-02-08T06:37:48.000Z
2021-09-12T14:55:34.000Z
api/endpoints/fruit/get.py
DarkbordermanTemplate/fastapi-redis-sqlalchemy
80fbdc419b19592b08bc2227c9d7c2925b7b91e2
[ "BSD-2-Clause" ]
null
null
null
api/endpoints/fruit/get.py
DarkbordermanTemplate/fastapi-redis-sqlalchemy
80fbdc419b19592b08bc2227c9d7c2925b7b91e2
[ "BSD-2-Clause" ]
null
null
null
from cache import REDIS from common.enums import EnumResponse from fastapi.responses import JSONResponse from loguru import logger DOC = { 200: { "description": "API response successfully", "content": {"application/json": {"example": {"name": "apple"}}}, }, 400: EnumResponse.BAD_REQUEST.value.doc, 500: EnumResponse.INTERNAL_SERVER_ERROR.value.doc, } def get(name: str): try: if REDIS.get(name) is None: return EnumResponse.BAD_REQUEST.value.response return JSONResponse({"name": name, "count": int(REDIS.get(name).decode())}, 200) # type: ignore except Exception as error: logger.warning(error) return EnumResponse.INTERNAL_SERVER_ERROR.value.response
30.916667
104
0.679245
from cache import REDIS from common.enums import EnumResponse from fastapi.responses import JSONResponse from loguru import logger DOC = { 200: { "description": "API response successfully", "content": {"application/json": {"example": {"name": "apple"}}}, }, 400: EnumResponse.BAD_REQUEST.value.doc, 500: EnumResponse.INTERNAL_SERVER_ERROR.value.doc, } def get(name: str): try: if REDIS.get(name) is None: return EnumResponse.BAD_REQUEST.value.response return JSONResponse({"name": name, "count": int(REDIS.get(name).decode())}, 200) except Exception as error: logger.warning(error) return EnumResponse.INTERNAL_SERVER_ERROR.value.response
true
true
1c48680130184ca429dd07e0772847c963db0ed3
368
py
Python
LVM-Tool/function.py
Shashwatsingh22/Linux-Automated-Tools
2e9c0f064ac70571a1a59e30f69e24d8ae05616a
[ "MIT" ]
null
null
null
LVM-Tool/function.py
Shashwatsingh22/Linux-Automated-Tools
2e9c0f064ac70571a1a59e30f69e24d8ae05616a
[ "MIT" ]
null
null
null
LVM-Tool/function.py
Shashwatsingh22/Linux-Automated-Tools
2e9c0f064ac70571a1a59e30f69e24d8ae05616a
[ "MIT" ]
null
null
null
from pyfiglet import Figlet def render(text,style,num): f=Figlet(font=style) print('\n') print(f.renderText(text)) def sh_menu(): print("""\t\t\t Press 1: Create. Press 2: Complete Detail. Press 3: Specific. Press 4: Exit. \n Enter the Choice: """,end=" ")
24.533333
44
0.486413
from pyfiglet import Figlet def render(text,style,num): f=Figlet(font=style) print('\n') print(f.renderText(text)) def sh_menu(): print("""\t\t\t Press 1: Create. Press 2: Complete Detail. Press 3: Specific. Press 4: Exit. \n Enter the Choice: """,end=" ")
true
true
1c4868035b91b83cc30e54e14d221dc6f5c6ac0e
2,474
py
Python
src/appier/test/exceptions.py
veryprofessionaldodo/appier
1a0c146753428a3d1a8c484467766ee871047757
[ "Apache-2.0" ]
null
null
null
src/appier/test/exceptions.py
veryprofessionaldodo/appier
1a0c146753428a3d1a8c484467766ee871047757
[ "Apache-2.0" ]
null
null
null
src/appier/test/exceptions.py
veryprofessionaldodo/appier
1a0c146753428a3d1a8c484467766ee871047757
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/python # -*- coding: utf-8 -*- # Hive Appier Framework # Copyright (c) 2008-2020 Hive Solutions Lda. # # This file is part of Hive Appier Framework. # # Hive Appier Framework is free software: you can redistribute it and/or modify # it under the terms of the Apache License as published by the Apache # Foundation, either version 2.0 of the License, or (at your option) any # later version. # # Hive Appier Framework is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # Apache License for more details. # # You should have received a copy of the Apache License along with # Hive Appier Framework. If not, see <http://www.apache.org/licenses/>. __author__ = "João Magalhães <joamag@hive.pt>" """ The author(s) of the module """ __version__ = "1.0.0" """ The version of the module """ __revision__ = "$LastChangedRevision$" """ The revision number of the module """ __date__ = "$LastChangedDate$" """ The last change date of the module """ __copyright__ = "Copyright (c) 2008-2020 Hive Solutions Lda." """ The copyright for the module """ __license__ = "Apache License, Version 2.0" """ The license for the module """ import unittest import appier class ExceptionsTest(unittest.TestCase): def test_encoding(self): exception = appier.AppierException(message = "Olá Mundo") self.assertEqual(str(exception), "Olá Mundo") message_u = appier.legacy.u("Olá Mundo") exception = appier.AppierException(message = message_u) self.assertEqual(str(exception), "Olá Mundo") self.assertEqual(appier.legacy.UNICODE(exception), appier.legacy.u("Olá Mundo")) def test_validation(self): errors = dict(name = ["is empty"]) error = appier.ValidationError(errors, object) errors_s = error.errors_s() self.assertEqual(errors_s, "name => is empty") errors = dict(name = ["Olá Mundo"]) error = appier.ValidationError(errors, object) errors_s = error.errors_s() self.assertEqual(errors_s, appier.legacy.u("name => Olá Mundo")) errors = dict(name = [appier.legacy.u("Olá Mundo")]) error = appier.ValidationError(errors, object) errors_s = error.errors_s() self.assertEqual(errors_s, appier.legacy.u("name => Olá Mundo"))
33.890411
89
0.67017
__author__ = "João Magalhães <joamag@hive.pt>" __version__ = "1.0.0" __revision__ = "$LastChangedRevision$" __date__ = "$LastChangedDate$" __copyright__ = "Copyright (c) 2008-2020 Hive Solutions Lda." __license__ = "Apache License, Version 2.0" import unittest import appier class ExceptionsTest(unittest.TestCase): def test_encoding(self): exception = appier.AppierException(message = "Olá Mundo") self.assertEqual(str(exception), "Olá Mundo") message_u = appier.legacy.u("Olá Mundo") exception = appier.AppierException(message = message_u) self.assertEqual(str(exception), "Olá Mundo") self.assertEqual(appier.legacy.UNICODE(exception), appier.legacy.u("Olá Mundo")) def test_validation(self): errors = dict(name = ["is empty"]) error = appier.ValidationError(errors, object) errors_s = error.errors_s() self.assertEqual(errors_s, "name => is empty") errors = dict(name = ["Olá Mundo"]) error = appier.ValidationError(errors, object) errors_s = error.errors_s() self.assertEqual(errors_s, appier.legacy.u("name => Olá Mundo")) errors = dict(name = [appier.legacy.u("Olá Mundo")]) error = appier.ValidationError(errors, object) errors_s = error.errors_s() self.assertEqual(errors_s, appier.legacy.u("name => Olá Mundo"))
true
true
1c48683aa3013e98712b3e7bf3aafb554f2f1671
955
py
Python
cybox/objects/uri_object.py
siemens/python-cybox
b692a98c8a62bd696e2a0dda802ada7359853482
[ "BSD-3-Clause" ]
null
null
null
cybox/objects/uri_object.py
siemens/python-cybox
b692a98c8a62bd696e2a0dda802ada7359853482
[ "BSD-3-Clause" ]
null
null
null
cybox/objects/uri_object.py
siemens/python-cybox
b692a98c8a62bd696e2a0dda802ada7359853482
[ "BSD-3-Clause" ]
1
2019-04-16T18:37:32.000Z
2019-04-16T18:37:32.000Z
# Copyright (c) 2014, The MITRE Corporation. All rights reserved. # See LICENSE.txt for complete terms. import cybox import cybox.bindings.uri_object as uri_binding from cybox.common import ObjectProperties, AnyURI class URI(ObjectProperties): _binding = uri_binding _binding_class = uri_binding.URIObjectType _namespace = 'http://cybox.mitre.org/objects#URIObject-2' _XSI_NS = 'URIObj' _XSI_TYPE = "URIObjectType" TYPE_URL = "URL" TYPE_GENERAL = "General URN" TYPE_DOMAIN = "Domain Name" TYPES = (TYPE_URL, TYPE_GENERAL, TYPE_DOMAIN) value = cybox.TypedField("Value", AnyURI) type_ = cybox.TypedField("type_", key_name="type") def __init__(self, value=None, type_=None): super(URI, self).__init__() self.value = value self.type_ = type_ def __str__(self): return self.__unicode__().encode("utf-8") def __unicode__(self): return unicode(self.value)
26.527778
65
0.690052
import cybox import cybox.bindings.uri_object as uri_binding from cybox.common import ObjectProperties, AnyURI class URI(ObjectProperties): _binding = uri_binding _binding_class = uri_binding.URIObjectType _namespace = 'http://cybox.mitre.org/objects#URIObject-2' _XSI_NS = 'URIObj' _XSI_TYPE = "URIObjectType" TYPE_URL = "URL" TYPE_GENERAL = "General URN" TYPE_DOMAIN = "Domain Name" TYPES = (TYPE_URL, TYPE_GENERAL, TYPE_DOMAIN) value = cybox.TypedField("Value", AnyURI) type_ = cybox.TypedField("type_", key_name="type") def __init__(self, value=None, type_=None): super(URI, self).__init__() self.value = value self.type_ = type_ def __str__(self): return self.__unicode__().encode("utf-8") def __unicode__(self): return unicode(self.value)
true
true
1c486ab1f4c57339efdceb4e5602b8c3f5c54e15
1,317
py
Python
FastAPI/app/main.py
bing9/raspberrypi_projects
5ca1b8101517f856af3f86a49518a89c1d8e29f9
[ "MIT" ]
null
null
null
FastAPI/app/main.py
bing9/raspberrypi_projects
5ca1b8101517f856af3f86a49518a89c1d8e29f9
[ "MIT" ]
null
null
null
FastAPI/app/main.py
bing9/raspberrypi_projects
5ca1b8101517f856af3f86a49518a89c1d8e29f9
[ "MIT" ]
null
null
null
from fastapi import FastAPI, UploadFile # from typing import Optional # from pydantic import BaseModel from subprocess import Popen #check_output # from starlette.responses import # from dotenv import load_dotenv import os app = FastAPI() @app.get('/') def index(): return 'My Personal Server' # @app.get('/apple-touch-icon-120x120-precomposed.png') # def image_png(): # with open('./FastAPI/app/apple-touch-icon-120x120-precomposed.png', 'r') as file: # img = file.read() # return UploadFile(filename="apple-touch-icon-120x120-precomposed.png", file=img, content_type="image/png") @app.get('/start_kodi') def start_kodi(): # load_dotenv() # return check_output(['sshpass', '-p', os.environ['SSHPASS'], 'ssh', '-p', '1990', 'jetson@192.168.0.170', 'kd']) try: Popen(["startx", "kodi"]) return {'successfully launched kodi'} except: return {'failed to launch kodi'} @app.get('/start_chrome') def start_chrome(): # load_dotenv() # return check_output(['sshpass', '-p', os.environ['SSHPASS'], 'ssh', '-p', '1990', 'jetson@192.168.0.170', 'kd']) try: Popen(["startx", "chromium-browser"]) return {'successfully launched chrome'} except: return {'failed to launch chrome'}
31.357143
119
0.631739
from fastapi import FastAPI, UploadFile from subprocess import Popen import os app = FastAPI() @app.get('/') def index(): return 'My Personal Server' @app.get('/start_kodi') def start_kodi(): try: Popen(["startx", "kodi"]) return {'successfully launched kodi'} except: return {'failed to launch kodi'} @app.get('/start_chrome') def start_chrome(): try: Popen(["startx", "chromium-browser"]) return {'successfully launched chrome'} except: return {'failed to launch chrome'}
true
true
1c486bab8c2b70b4200b22d717bec1830d4b9e0e
1,297
py
Python
solum/objects/assembly.py
devdattakulkarni/test-solum
4e9ddb82d217116aa2c30a6f2581080cbdfae325
[ "Apache-2.0" ]
null
null
null
solum/objects/assembly.py
devdattakulkarni/test-solum
4e9ddb82d217116aa2c30a6f2581080cbdfae325
[ "Apache-2.0" ]
null
null
null
solum/objects/assembly.py
devdattakulkarni/test-solum
4e9ddb82d217116aa2c30a6f2581080cbdfae325
[ "Apache-2.0" ]
null
null
null
# Copyright 2014 - Rackspace # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from solum.objects import base class Assembly(base.CrudMixin): # Version 1.0: Initial version VERSION = '1.0' class AssemblyList(list, base.CrudListMixin): """List of Assemblies.""" class States(object): QUEUED = 'QUEUED' UNIT_TESTING = 'UNIT_TESTING' UNIT_TESTING_FAILED = 'UNIT_TESTING_FAILED' UNIT_TESTING_PASSED = 'UNIT_TESTING_PASSED' BUILDING = 'BUILDING' BUILT = 'BUILT' DEPLOYING = 'DEPLOYING' ERROR = 'ERROR' READY = 'READY' DELETING = 'DELETING' ERROR_STACK_DELETE_FAILED = 'ERROR_STACK_DELETE_FAILED' ERROR_STACK_CREATE_FAILED = 'ERROR_STACK_CREATE_FAILED' ERROR_CODE_DEPLOYMENT = 'ERROR_CODE_DEPLOYMENT' STARTING_APP = 'STARTING_APP'
30.880952
75
0.733231
from solum.objects import base class Assembly(base.CrudMixin): VERSION = '1.0' class AssemblyList(list, base.CrudListMixin): class States(object): QUEUED = 'QUEUED' UNIT_TESTING = 'UNIT_TESTING' UNIT_TESTING_FAILED = 'UNIT_TESTING_FAILED' UNIT_TESTING_PASSED = 'UNIT_TESTING_PASSED' BUILDING = 'BUILDING' BUILT = 'BUILT' DEPLOYING = 'DEPLOYING' ERROR = 'ERROR' READY = 'READY' DELETING = 'DELETING' ERROR_STACK_DELETE_FAILED = 'ERROR_STACK_DELETE_FAILED' ERROR_STACK_CREATE_FAILED = 'ERROR_STACK_CREATE_FAILED' ERROR_CODE_DEPLOYMENT = 'ERROR_CODE_DEPLOYMENT' STARTING_APP = 'STARTING_APP'
true
true
1c486cbfbc9842418bee088aeb5300aed2824063
29,399
py
Python
core/domain/story_services.py
jlau323/oppia
37438a2c9bf7e66892fb9a6a93a1fe4ca7a82691
[ "Apache-2.0" ]
2
2021-04-08T01:06:08.000Z
2021-06-02T08:20:13.000Z
core/domain/story_services.py
jlau323/oppia
37438a2c9bf7e66892fb9a6a93a1fe4ca7a82691
[ "Apache-2.0" ]
null
null
null
core/domain/story_services.py
jlau323/oppia
37438a2c9bf7e66892fb9a6a93a1fe4ca7a82691
[ "Apache-2.0" ]
1
2020-12-11T06:56:31.000Z
2020-12-11T06:56:31.000Z
# Copyright 2018 The Oppia Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS-IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Commands that can be used to operate on stories. All functions here should be agnostic of how StoryModel objects are stored in the database. In particular, the various query methods should delegate to the Story model class. This will enable the story storage model to be changed without affecting this module and others above it. """ from __future__ import absolute_import # pylint: disable=import-only-modules from __future__ import unicode_literals # pylint: disable=import-only-modules import logging from constants import constants from core.domain import android_validation_constants from core.domain import caching_services from core.domain import exp_fetchers from core.domain import opportunity_services from core.domain import rights_manager from core.domain import story_domain from core.domain import story_fetchers from core.domain import suggestion_services from core.domain import topic_fetchers from core.platform import models import feconf import utils (exp_models, story_models, user_models,) = models.Registry.import_models( [models.NAMES.exploration, models.NAMES.story, models.NAMES.user]) def get_new_story_id(): """Returns a new story id. Returns: str. A new story id. """ return story_models.StoryModel.get_new_id('') def _create_story(committer_id, story, commit_message, commit_cmds): """Creates a new story. Args: committer_id: str. ID of the committer. story: Story. The story domain object. commit_message: str. A description of changes made to the story. commit_cmds: list(StoryChange). A list of change commands made to the given story. """ story.validate() model = story_models.StoryModel( id=story.id, description=story.description, title=story.title, thumbnail_bg_color=story.thumbnail_bg_color, thumbnail_filename=story.thumbnail_filename, language_code=story.language_code, story_contents_schema_version=story.story_contents_schema_version, notes=story.notes, story_contents=story.story_contents.to_dict(), corresponding_topic_id=story.corresponding_topic_id, url_fragment=story.url_fragment, meta_tag_content=story.meta_tag_content ) commit_cmd_dicts = [commit_cmd.to_dict() for commit_cmd in commit_cmds] model.commit(committer_id, commit_message, commit_cmd_dicts) story.version += 1 create_story_summary(story.id) def save_new_story(committer_id, story): """Saves a new story. Args: committer_id: str. ID of the committer. story: Story. Story to be saved. """ commit_message = ( 'New story created with title \'%s\'.' % story.title) _create_story( committer_id, story, commit_message, [story_domain.StoryChange({ 'cmd': story_domain.CMD_CREATE_NEW, 'title': story.title })]) # Repository SAVE and DELETE methods. def apply_change_list(story_id, change_list): """Applies a changelist to a story and returns the result. Args: story_id: str. ID of the given story. change_list: list(StoryChange). A change list to be applied to the given story. Returns: Story, list(str), list(str). The resulting story domain object, the exploration IDs removed from story and the exploration IDs added to the story. """ story = story_fetchers.get_story_by_id(story_id) exp_ids_in_old_story = story.story_contents.get_all_linked_exp_ids() try: for change in change_list: if not isinstance(change, story_domain.StoryChange): raise Exception('Expected change to be of type StoryChange') if change.cmd == story_domain.CMD_ADD_STORY_NODE: story.add_node(change.node_id, change.title) elif change.cmd == story_domain.CMD_DELETE_STORY_NODE: story.delete_node(change.node_id) elif (change.cmd == story_domain.CMD_UPDATE_STORY_NODE_OUTLINE_STATUS): if change.new_value: story.mark_node_outline_as_finalized(change.node_id) else: story.mark_node_outline_as_unfinalized(change.node_id) elif change.cmd == story_domain.CMD_UPDATE_STORY_NODE_PROPERTY: if (change.property_name == story_domain.STORY_NODE_PROPERTY_OUTLINE): story.update_node_outline(change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_TITLE): story.update_node_title(change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_DESCRIPTION): story.update_node_description( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_THUMBNAIL_FILENAME): story.update_node_thumbnail_filename( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_THUMBNAIL_BG_COLOR): story.update_node_thumbnail_bg_color( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_ACQUIRED_SKILL_IDS): story.update_node_acquired_skill_ids( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_PREREQUISITE_SKILL_IDS): story.update_node_prerequisite_skill_ids( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_DESTINATION_NODE_IDS): story.update_node_destination_node_ids( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_EXPLORATION_ID): story.update_node_exploration_id( change.node_id, change.new_value) elif change.cmd == story_domain.CMD_UPDATE_STORY_PROPERTY: if (change.property_name == story_domain.STORY_PROPERTY_TITLE): story.update_title(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_THUMBNAIL_FILENAME): story.update_thumbnail_filename(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_THUMBNAIL_BG_COLOR): story.update_thumbnail_bg_color(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_DESCRIPTION): story.update_description(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_NOTES): story.update_notes(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_LANGUAGE_CODE): story.update_language_code(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_URL_FRAGMENT): story.update_url_fragment(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_META_TAG_CONTENT): story.update_meta_tag_content(change.new_value) elif change.cmd == story_domain.CMD_UPDATE_STORY_CONTENTS_PROPERTY: if (change.property_name == story_domain.INITIAL_NODE_ID): story.update_initial_node(change.new_value) if change.property_name == story_domain.NODE: story.rearrange_node_in_story( change.old_value, change.new_value) elif ( change.cmd == story_domain.CMD_MIGRATE_SCHEMA_TO_LATEST_VERSION): # Loading the story model from the datastore into a # Story domain object automatically converts it to use the # latest schema version. As a result, simply resaving the # story is sufficient to apply the schema migration. continue exp_ids_in_modified_story = ( story.story_contents.get_all_linked_exp_ids()) exp_ids_removed_from_story = list( set(exp_ids_in_old_story).difference(exp_ids_in_modified_story)) exp_ids_added_to_story = list( set(exp_ids_in_modified_story).difference(exp_ids_in_old_story)) return story, exp_ids_removed_from_story, exp_ids_added_to_story except Exception as e: logging.error( '%s %s %s %s' % ( e.__class__.__name__, e, story_id, change_list) ) raise def does_story_exist_with_url_fragment(url_fragment): """Checks if the url fragment for the story exists. Args: url_fragment: str. The url_fragment of the story. Returns: bool. Whether the the url fragment for the story exists or not. """ story = story_fetchers.get_story_by_url_fragment(url_fragment) return story is not None def validate_explorations_for_story(exp_ids, raise_error): """Validates the explorations in the given story and checks whether they are compatible with the mobile app and ready for publishing. Args: exp_ids: list(str). The exp IDs to validate. raise_error: bool. Whether to raise an Exception when a validation error is encountered. If not, a list of the error messages are returned. raise_error should be True when this is called before saving the story and False when this function is called from the frontend. Returns: list(str). The various validation error messages (if raise_error is False). Raises: ValidationError. Expected story to only reference valid explorations. ValidationError. Exploration with ID is not public. Please publish explorations before adding them to a story. ValidationError. All explorations in a story should be of the same category. ValidationError. Invalid language found for exploration. ValidationError. Expected no exploration to have parameter values in it. ValidationError. Invalid interaction in exploration. ValidationError. RTE content in state of exploration with ID is not supported on mobile. """ validation_error_messages = [] # Strict = False, since the existence of explorations is checked below. exps_dict = ( exp_fetchers.get_multiple_explorations_by_id(exp_ids, strict=False)) exp_rights = ( rights_manager.get_multiple_exploration_rights_by_ids(exp_ids)) exp_rights_dict = {} for rights in exp_rights: if rights is not None: exp_rights_dict[rights.id] = rights.status for exp_id in exp_ids: if exp_id not in exps_dict: error_string = ( 'Expected story to only reference valid explorations, but found' ' a reference to an invalid exploration with ID: %s' % exp_id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) else: if exp_rights_dict[exp_id] != constants.ACTIVITY_STATUS_PUBLIC: error_string = ( 'Exploration with ID %s is not public. Please publish ' 'explorations before adding them to a story.' % exp_id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if exps_dict: for exp_id in exp_ids: if exp_id in exps_dict: sample_exp_id = exp_id break common_exp_category = exps_dict[sample_exp_id].category for exp_id in exps_dict: exp = exps_dict[exp_id] if exp.category != common_exp_category: error_string = ( 'All explorations in a story should be of the ' 'same category. The explorations with ID %s and %s have' ' different categories.' % (sample_exp_id, exp_id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if ( exp.language_code not in android_validation_constants.SUPPORTED_LANGUAGES): error_string = ( 'Invalid language %s found for exploration ' 'with ID %s.' % (exp.language_code, exp_id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if exp.param_specs or exp.param_changes: error_string = ( 'Expected no exploration to have parameter ' 'values in it. Invalid exploration: %s' % exp.id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if not exp.correctness_feedback_enabled: error_string = ( 'Expected all explorations to have correctness feedback ' 'enabled. Invalid exploration: %s' % exp.id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) for state_name in exp.states: state = exp.states[state_name] if not state.interaction.is_supported_on_android_app(): error_string = ( 'Invalid interaction %s in exploration ' 'with ID: %s.' % (state.interaction.id, exp.id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if not state.is_rte_content_supported_on_android(): error_string = ( 'RTE content in state %s of exploration ' 'with ID %s is not supported on mobile.' % (state_name, exp.id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if state.interaction.id == 'EndExploration': recommended_exploration_ids = ( state.interaction.customization_args[ 'recommendedExplorationIds'].value) if len(recommended_exploration_ids) != 0: error_string = ( 'Exploration with ID: %s contains exploration ' 'recommendations in its EndExploration interaction.' % (exp.id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) return validation_error_messages def _save_story( committer_id, story, commit_message, change_list, story_is_published): """Validates a story and commits it to persistent storage. If successful, increments the version number of the incoming story domain object by 1. Args: committer_id: str. ID of the given committer. story: Story. The story domain object to be saved. commit_message: str. The commit message. change_list: list(StoryChange). List of changes applied to a story. story_is_published: bool. Whether the supplied story is published. Raises: ValidationError. An invalid exploration was referenced in the story. Exception. The story model and the incoming story domain object have different version numbers. """ if not change_list: raise Exception( 'Unexpected error: received an invalid change list when trying to ' 'save story %s: %s' % (story.id, change_list)) story.validate() if story_is_published: exp_ids = [] for node in story.story_contents.nodes: if not node.exploration_id: raise Exception( 'Story node with id %s does not contain an ' 'exploration id.' % node.id) exp_ids.append(node.exploration_id) validate_explorations_for_story(exp_ids, True) # Story model cannot be None as story is passed as parameter here and that # is only possible if a story model with that story id exists. Also this is # a private function and so it cannot be called independently with any # story object. story_model = story_models.StoryModel.get(story.id) if story.version > story_model.version: raise Exception( 'Unexpected error: trying to update version %s of story ' 'from version %s. Please reload the page and try again.' % (story_model.version, story.version)) elif story.version < story_model.version: raise Exception( 'Trying to update version %s of story from version %s, ' 'which is too old. Please reload the page and try again.' % (story_model.version, story.version)) story_model.description = story.description story_model.title = story.title story_model.thumbnail_bg_color = story.thumbnail_bg_color story_model.thumbnail_filename = story.thumbnail_filename story_model.notes = story.notes story_model.language_code = story.language_code story_model.story_contents_schema_version = ( story.story_contents_schema_version) story_model.story_contents = story.story_contents.to_dict() story_model.corresponding_topic_id = story.corresponding_topic_id story_model.version = story.version story_model.url_fragment = story.url_fragment story_model.meta_tag_content = story.meta_tag_content change_dicts = [change.to_dict() for change in change_list] story_model.commit(committer_id, commit_message, change_dicts) caching_services.delete_multi( caching_services.CACHE_NAMESPACE_STORY, None, [story.id]) story.version += 1 def _is_story_published_and_present_in_topic(story): """Returns whether a story is published. Raises an exception if the story is not present in the corresponding topic's story references. Args: story: Story. The story domain object. Returns: bool. Whether the supplied story is published. """ topic = topic_fetchers.get_topic_by_id( story.corresponding_topic_id, strict=False) if topic is None: raise utils.ValidationError( 'Expected story to only belong to a valid topic, but found no ' 'topic with ID: %s' % story.corresponding_topic_id) story_is_published = False story_is_present_in_topic = False for story_reference in topic.get_all_story_references(): if story_reference.story_id == story.id: story_is_present_in_topic = True story_is_published = story_reference.story_is_published if not story_is_present_in_topic: raise Exception( 'Expected story to belong to the topic %s, but it is ' 'neither a part of the canonical stories or the additional ' 'stories of the topic.' % story.corresponding_topic_id) return story_is_published def update_story( committer_id, story_id, change_list, commit_message): """Updates a story. Commits changes. Args: committer_id: str. The id of the user who is performing the update action. story_id: str. The story id. change_list: list(StoryChange). These changes are applied in sequence to produce the resulting story. commit_message: str or None. A description of changes made to the story. Raises: ValidationError. Exploration is already linked to a different story. """ if not commit_message: raise ValueError('Expected a commit message but received none.') old_story = story_fetchers.get_story_by_id(story_id) new_story, exp_ids_removed_from_story, exp_ids_added_to_story = ( apply_change_list(story_id, change_list)) story_is_published = _is_story_published_and_present_in_topic(new_story) if ( old_story.url_fragment != new_story.url_fragment and does_story_exist_with_url_fragment(new_story.url_fragment)): raise utils.ValidationError( 'Story Url Fragment is not unique across the site.') _save_story( committer_id, new_story, commit_message, change_list, story_is_published) create_story_summary(new_story.id) if story_is_published and _is_topic_published(new_story): opportunity_services.update_exploration_opportunities( old_story, new_story) suggestion_services.auto_reject_translation_suggestions_for_exp_ids( exp_ids_removed_from_story) exploration_context_models_to_be_deleted = ( exp_models.ExplorationContextModel.get_multi( exp_ids_removed_from_story)) exploration_context_models_to_be_deleted = [ model for model in exploration_context_models_to_be_deleted if model is not None] exp_models.ExplorationContextModel.delete_multi( exploration_context_models_to_be_deleted) exploration_context_models_collisions_list = ( exp_models.ExplorationContextModel.get_multi( exp_ids_added_to_story)) for context_model in exploration_context_models_collisions_list: if context_model is not None and context_model.story_id != story_id: raise utils.ValidationError( 'The exploration with ID %s is already linked to story ' 'with ID %s' % (context_model.id, context_model.story_id)) new_exploration_context_models = [exp_models.ExplorationContextModel( id=exp_id, story_id=story_id ) for exp_id in exp_ids_added_to_story] exp_models.ExplorationContextModel.update_timestamps_multi( new_exploration_context_models) exp_models.ExplorationContextModel.put_multi(new_exploration_context_models) def _is_topic_published(story): """Returns whether the story's corresponding topic is published. Args: story: Story. The story domain object. Returns: bool. Whether the the story's corresponding topic is published. """ topic_rights = topic_fetchers.get_topic_rights(story.corresponding_topic_id) return topic_rights.topic_is_published def delete_story(committer_id, story_id, force_deletion=False): """Deletes the story with the given story_id. Args: committer_id: str. ID of the committer. story_id: str. ID of the story to be deleted. force_deletion: bool. If true, the story and its history are fully deleted and are unrecoverable. Otherwise, the story and all its history are marked as deleted, but the corresponding models are still retained in the datastore. This last option is the preferred one. """ story_model = story_models.StoryModel.get(story_id) story = story_fetchers.get_story_from_model(story_model) exp_ids = story.story_contents.get_all_linked_exp_ids() story_model.delete( committer_id, feconf.COMMIT_MESSAGE_STORY_DELETED, force_deletion=force_deletion) exp_ids_to_be_removed = [] for node in story.story_contents.nodes: exp_ids_to_be_removed.append(node.exploration_id) exploration_context_models_to_be_deleted = ( exp_models.ExplorationContextModel.get_multi( exp_ids_to_be_removed)) exploration_context_models_to_be_deleted = [ model for model in exploration_context_models_to_be_deleted if model is not None] exp_models.ExplorationContextModel.delete_multi( exploration_context_models_to_be_deleted) # This must come after the story is retrieved. Otherwise the memcache # key will be reinstated. caching_services.delete_multi( caching_services.CACHE_NAMESPACE_STORY, None, [story_id]) # Delete the summary of the story (regardless of whether # force_deletion is True or not). delete_story_summary(story_id) # Delete the opportunities available and reject the suggestions related to # the exploration used in the story. opportunity_services.delete_exploration_opportunities(exp_ids) suggestion_services.auto_reject_translation_suggestions_for_exp_ids( exp_ids) def delete_story_summary(story_id): """Delete a story summary model. Args: story_id: str. ID of the story whose story summary is to be deleted. """ story_models.StorySummaryModel.get(story_id).delete() def compute_summary_of_story(story): """Create a StorySummary domain object for a given Story domain object and return it. Args: story: Story. The story object, for which the summary is to be computed. Returns: StorySummary. The computed summary for the given story. """ story_model_node_titles = [ node.title for node in story.story_contents.nodes] story_summary = story_domain.StorySummary( story.id, story.title, story.description, story.language_code, story.version, story_model_node_titles, story.thumbnail_bg_color, story.thumbnail_filename, story.url_fragment, story.created_on, story.last_updated ) return story_summary def create_story_summary(story_id): """Creates and stores a summary of the given story. Args: story_id: str. ID of the story. """ story = story_fetchers.get_story_by_id(story_id) story_summary = compute_summary_of_story(story) save_story_summary(story_summary) def save_story_summary(story_summary): """Save a story summary domain object as a StorySummaryModel entity in the datastore. Args: story_summary: StorySummary. The story summary object to be saved in the datastore. """ story_summary_dict = { 'title': story_summary.title, 'description': story_summary.description, 'language_code': story_summary.language_code, 'version': story_summary.version, 'node_titles': story_summary.node_titles, 'thumbnail_bg_color': story_summary.thumbnail_bg_color, 'thumbnail_filename': story_summary.thumbnail_filename, 'url_fragment': story_summary.url_fragment, 'story_model_last_updated': ( story_summary.story_model_last_updated), 'story_model_created_on': ( story_summary.story_model_created_on) } story_summary_model = ( story_models.StorySummaryModel.get_by_id(story_summary.id)) if story_summary_model is not None: story_summary_model.populate(**story_summary_dict) story_summary_model.update_timestamps() story_summary_model.put() else: story_summary_dict['id'] = story_summary.id model = story_models.StorySummaryModel(**story_summary_dict) model.update_timestamps() model.put() def record_completed_node_in_story_context(user_id, story_id, node_id): """Records a node by a given user in a given story context as having been played. Args: user_id: str. ID of the given user. story_id: str. ID of the given story. node_id: str. ID of the given node. """ progress_model = user_models.StoryProgressModel.get_or_create( user_id, story_id) if node_id not in progress_model.completed_node_ids: progress_model.completed_node_ids.append(node_id) progress_model.update_timestamps() progress_model.put()
41.759943
80
0.661689
from __future__ import absolute_import from __future__ import unicode_literals import logging from constants import constants from core.domain import android_validation_constants from core.domain import caching_services from core.domain import exp_fetchers from core.domain import opportunity_services from core.domain import rights_manager from core.domain import story_domain from core.domain import story_fetchers from core.domain import suggestion_services from core.domain import topic_fetchers from core.platform import models import feconf import utils (exp_models, story_models, user_models,) = models.Registry.import_models( [models.NAMES.exploration, models.NAMES.story, models.NAMES.user]) def get_new_story_id(): return story_models.StoryModel.get_new_id('') def _create_story(committer_id, story, commit_message, commit_cmds): story.validate() model = story_models.StoryModel( id=story.id, description=story.description, title=story.title, thumbnail_bg_color=story.thumbnail_bg_color, thumbnail_filename=story.thumbnail_filename, language_code=story.language_code, story_contents_schema_version=story.story_contents_schema_version, notes=story.notes, story_contents=story.story_contents.to_dict(), corresponding_topic_id=story.corresponding_topic_id, url_fragment=story.url_fragment, meta_tag_content=story.meta_tag_content ) commit_cmd_dicts = [commit_cmd.to_dict() for commit_cmd in commit_cmds] model.commit(committer_id, commit_message, commit_cmd_dicts) story.version += 1 create_story_summary(story.id) def save_new_story(committer_id, story): commit_message = ( 'New story created with title \'%s\'.' % story.title) _create_story( committer_id, story, commit_message, [story_domain.StoryChange({ 'cmd': story_domain.CMD_CREATE_NEW, 'title': story.title })]) def apply_change_list(story_id, change_list): story = story_fetchers.get_story_by_id(story_id) exp_ids_in_old_story = story.story_contents.get_all_linked_exp_ids() try: for change in change_list: if not isinstance(change, story_domain.StoryChange): raise Exception('Expected change to be of type StoryChange') if change.cmd == story_domain.CMD_ADD_STORY_NODE: story.add_node(change.node_id, change.title) elif change.cmd == story_domain.CMD_DELETE_STORY_NODE: story.delete_node(change.node_id) elif (change.cmd == story_domain.CMD_UPDATE_STORY_NODE_OUTLINE_STATUS): if change.new_value: story.mark_node_outline_as_finalized(change.node_id) else: story.mark_node_outline_as_unfinalized(change.node_id) elif change.cmd == story_domain.CMD_UPDATE_STORY_NODE_PROPERTY: if (change.property_name == story_domain.STORY_NODE_PROPERTY_OUTLINE): story.update_node_outline(change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_TITLE): story.update_node_title(change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_DESCRIPTION): story.update_node_description( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_THUMBNAIL_FILENAME): story.update_node_thumbnail_filename( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_THUMBNAIL_BG_COLOR): story.update_node_thumbnail_bg_color( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_ACQUIRED_SKILL_IDS): story.update_node_acquired_skill_ids( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_PREREQUISITE_SKILL_IDS): story.update_node_prerequisite_skill_ids( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_DESTINATION_NODE_IDS): story.update_node_destination_node_ids( change.node_id, change.new_value) elif (change.property_name == story_domain.STORY_NODE_PROPERTY_EXPLORATION_ID): story.update_node_exploration_id( change.node_id, change.new_value) elif change.cmd == story_domain.CMD_UPDATE_STORY_PROPERTY: if (change.property_name == story_domain.STORY_PROPERTY_TITLE): story.update_title(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_THUMBNAIL_FILENAME): story.update_thumbnail_filename(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_THUMBNAIL_BG_COLOR): story.update_thumbnail_bg_color(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_DESCRIPTION): story.update_description(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_NOTES): story.update_notes(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_LANGUAGE_CODE): story.update_language_code(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_URL_FRAGMENT): story.update_url_fragment(change.new_value) elif (change.property_name == story_domain.STORY_PROPERTY_META_TAG_CONTENT): story.update_meta_tag_content(change.new_value) elif change.cmd == story_domain.CMD_UPDATE_STORY_CONTENTS_PROPERTY: if (change.property_name == story_domain.INITIAL_NODE_ID): story.update_initial_node(change.new_value) if change.property_name == story_domain.NODE: story.rearrange_node_in_story( change.old_value, change.new_value) elif ( change.cmd == story_domain.CMD_MIGRATE_SCHEMA_TO_LATEST_VERSION): continue exp_ids_in_modified_story = ( story.story_contents.get_all_linked_exp_ids()) exp_ids_removed_from_story = list( set(exp_ids_in_old_story).difference(exp_ids_in_modified_story)) exp_ids_added_to_story = list( set(exp_ids_in_modified_story).difference(exp_ids_in_old_story)) return story, exp_ids_removed_from_story, exp_ids_added_to_story except Exception as e: logging.error( '%s %s %s %s' % ( e.__class__.__name__, e, story_id, change_list) ) raise def does_story_exist_with_url_fragment(url_fragment): story = story_fetchers.get_story_by_url_fragment(url_fragment) return story is not None def validate_explorations_for_story(exp_ids, raise_error): validation_error_messages = [] exps_dict = ( exp_fetchers.get_multiple_explorations_by_id(exp_ids, strict=False)) exp_rights = ( rights_manager.get_multiple_exploration_rights_by_ids(exp_ids)) exp_rights_dict = {} for rights in exp_rights: if rights is not None: exp_rights_dict[rights.id] = rights.status for exp_id in exp_ids: if exp_id not in exps_dict: error_string = ( 'Expected story to only reference valid explorations, but found' ' a reference to an invalid exploration with ID: %s' % exp_id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) else: if exp_rights_dict[exp_id] != constants.ACTIVITY_STATUS_PUBLIC: error_string = ( 'Exploration with ID %s is not public. Please publish ' 'explorations before adding them to a story.' % exp_id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if exps_dict: for exp_id in exp_ids: if exp_id in exps_dict: sample_exp_id = exp_id break common_exp_category = exps_dict[sample_exp_id].category for exp_id in exps_dict: exp = exps_dict[exp_id] if exp.category != common_exp_category: error_string = ( 'All explorations in a story should be of the ' 'same category. The explorations with ID %s and %s have' ' different categories.' % (sample_exp_id, exp_id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if ( exp.language_code not in android_validation_constants.SUPPORTED_LANGUAGES): error_string = ( 'Invalid language %s found for exploration ' 'with ID %s.' % (exp.language_code, exp_id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if exp.param_specs or exp.param_changes: error_string = ( 'Expected no exploration to have parameter ' 'values in it. Invalid exploration: %s' % exp.id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if not exp.correctness_feedback_enabled: error_string = ( 'Expected all explorations to have correctness feedback ' 'enabled. Invalid exploration: %s' % exp.id) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) for state_name in exp.states: state = exp.states[state_name] if not state.interaction.is_supported_on_android_app(): error_string = ( 'Invalid interaction %s in exploration ' 'with ID: %s.' % (state.interaction.id, exp.id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if not state.is_rte_content_supported_on_android(): error_string = ( 'RTE content in state %s of exploration ' 'with ID %s is not supported on mobile.' % (state_name, exp.id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) if state.interaction.id == 'EndExploration': recommended_exploration_ids = ( state.interaction.customization_args[ 'recommendedExplorationIds'].value) if len(recommended_exploration_ids) != 0: error_string = ( 'Exploration with ID: %s contains exploration ' 'recommendations in its EndExploration interaction.' % (exp.id)) if raise_error: raise utils.ValidationError(error_string) validation_error_messages.append(error_string) return validation_error_messages def _save_story( committer_id, story, commit_message, change_list, story_is_published): if not change_list: raise Exception( 'Unexpected error: received an invalid change list when trying to ' 'save story %s: %s' % (story.id, change_list)) story.validate() if story_is_published: exp_ids = [] for node in story.story_contents.nodes: if not node.exploration_id: raise Exception( 'Story node with id %s does not contain an ' 'exploration id.' % node.id) exp_ids.append(node.exploration_id) validate_explorations_for_story(exp_ids, True) story_model = story_models.StoryModel.get(story.id) if story.version > story_model.version: raise Exception( 'Unexpected error: trying to update version %s of story ' 'from version %s. Please reload the page and try again.' % (story_model.version, story.version)) elif story.version < story_model.version: raise Exception( 'Trying to update version %s of story from version %s, ' 'which is too old. Please reload the page and try again.' % (story_model.version, story.version)) story_model.description = story.description story_model.title = story.title story_model.thumbnail_bg_color = story.thumbnail_bg_color story_model.thumbnail_filename = story.thumbnail_filename story_model.notes = story.notes story_model.language_code = story.language_code story_model.story_contents_schema_version = ( story.story_contents_schema_version) story_model.story_contents = story.story_contents.to_dict() story_model.corresponding_topic_id = story.corresponding_topic_id story_model.version = story.version story_model.url_fragment = story.url_fragment story_model.meta_tag_content = story.meta_tag_content change_dicts = [change.to_dict() for change in change_list] story_model.commit(committer_id, commit_message, change_dicts) caching_services.delete_multi( caching_services.CACHE_NAMESPACE_STORY, None, [story.id]) story.version += 1 def _is_story_published_and_present_in_topic(story): topic = topic_fetchers.get_topic_by_id( story.corresponding_topic_id, strict=False) if topic is None: raise utils.ValidationError( 'Expected story to only belong to a valid topic, but found no ' 'topic with ID: %s' % story.corresponding_topic_id) story_is_published = False story_is_present_in_topic = False for story_reference in topic.get_all_story_references(): if story_reference.story_id == story.id: story_is_present_in_topic = True story_is_published = story_reference.story_is_published if not story_is_present_in_topic: raise Exception( 'Expected story to belong to the topic %s, but it is ' 'neither a part of the canonical stories or the additional ' 'stories of the topic.' % story.corresponding_topic_id) return story_is_published def update_story( committer_id, story_id, change_list, commit_message): if not commit_message: raise ValueError('Expected a commit message but received none.') old_story = story_fetchers.get_story_by_id(story_id) new_story, exp_ids_removed_from_story, exp_ids_added_to_story = ( apply_change_list(story_id, change_list)) story_is_published = _is_story_published_and_present_in_topic(new_story) if ( old_story.url_fragment != new_story.url_fragment and does_story_exist_with_url_fragment(new_story.url_fragment)): raise utils.ValidationError( 'Story Url Fragment is not unique across the site.') _save_story( committer_id, new_story, commit_message, change_list, story_is_published) create_story_summary(new_story.id) if story_is_published and _is_topic_published(new_story): opportunity_services.update_exploration_opportunities( old_story, new_story) suggestion_services.auto_reject_translation_suggestions_for_exp_ids( exp_ids_removed_from_story) exploration_context_models_to_be_deleted = ( exp_models.ExplorationContextModel.get_multi( exp_ids_removed_from_story)) exploration_context_models_to_be_deleted = [ model for model in exploration_context_models_to_be_deleted if model is not None] exp_models.ExplorationContextModel.delete_multi( exploration_context_models_to_be_deleted) exploration_context_models_collisions_list = ( exp_models.ExplorationContextModel.get_multi( exp_ids_added_to_story)) for context_model in exploration_context_models_collisions_list: if context_model is not None and context_model.story_id != story_id: raise utils.ValidationError( 'The exploration with ID %s is already linked to story ' 'with ID %s' % (context_model.id, context_model.story_id)) new_exploration_context_models = [exp_models.ExplorationContextModel( id=exp_id, story_id=story_id ) for exp_id in exp_ids_added_to_story] exp_models.ExplorationContextModel.update_timestamps_multi( new_exploration_context_models) exp_models.ExplorationContextModel.put_multi(new_exploration_context_models) def _is_topic_published(story): topic_rights = topic_fetchers.get_topic_rights(story.corresponding_topic_id) return topic_rights.topic_is_published def delete_story(committer_id, story_id, force_deletion=False): story_model = story_models.StoryModel.get(story_id) story = story_fetchers.get_story_from_model(story_model) exp_ids = story.story_contents.get_all_linked_exp_ids() story_model.delete( committer_id, feconf.COMMIT_MESSAGE_STORY_DELETED, force_deletion=force_deletion) exp_ids_to_be_removed = [] for node in story.story_contents.nodes: exp_ids_to_be_removed.append(node.exploration_id) exploration_context_models_to_be_deleted = ( exp_models.ExplorationContextModel.get_multi( exp_ids_to_be_removed)) exploration_context_models_to_be_deleted = [ model for model in exploration_context_models_to_be_deleted if model is not None] exp_models.ExplorationContextModel.delete_multi( exploration_context_models_to_be_deleted) caching_services.delete_multi( caching_services.CACHE_NAMESPACE_STORY, None, [story_id]) delete_story_summary(story_id) opportunity_services.delete_exploration_opportunities(exp_ids) suggestion_services.auto_reject_translation_suggestions_for_exp_ids( exp_ids) def delete_story_summary(story_id): story_models.StorySummaryModel.get(story_id).delete() def compute_summary_of_story(story): story_model_node_titles = [ node.title for node in story.story_contents.nodes] story_summary = story_domain.StorySummary( story.id, story.title, story.description, story.language_code, story.version, story_model_node_titles, story.thumbnail_bg_color, story.thumbnail_filename, story.url_fragment, story.created_on, story.last_updated ) return story_summary def create_story_summary(story_id): story = story_fetchers.get_story_by_id(story_id) story_summary = compute_summary_of_story(story) save_story_summary(story_summary) def save_story_summary(story_summary): story_summary_dict = { 'title': story_summary.title, 'description': story_summary.description, 'language_code': story_summary.language_code, 'version': story_summary.version, 'node_titles': story_summary.node_titles, 'thumbnail_bg_color': story_summary.thumbnail_bg_color, 'thumbnail_filename': story_summary.thumbnail_filename, 'url_fragment': story_summary.url_fragment, 'story_model_last_updated': ( story_summary.story_model_last_updated), 'story_model_created_on': ( story_summary.story_model_created_on) } story_summary_model = ( story_models.StorySummaryModel.get_by_id(story_summary.id)) if story_summary_model is not None: story_summary_model.populate(**story_summary_dict) story_summary_model.update_timestamps() story_summary_model.put() else: story_summary_dict['id'] = story_summary.id model = story_models.StorySummaryModel(**story_summary_dict) model.update_timestamps() model.put() def record_completed_node_in_story_context(user_id, story_id, node_id): progress_model = user_models.StoryProgressModel.get_or_create( user_id, story_id) if node_id not in progress_model.completed_node_ids: progress_model.completed_node_ids.append(node_id) progress_model.update_timestamps() progress_model.put()
true
true
1c486d9dbfb62cdd738c2ff2e418532e0d2734b8
6,437
py
Python
data/p3BR/R2/benchmark/startQiskit_noisy239.py
UCLA-SEAL/QDiff
d968cbc47fe926b7f88b4adf10490f1edd6f8819
[ "BSD-3-Clause" ]
null
null
null
data/p3BR/R2/benchmark/startQiskit_noisy239.py
UCLA-SEAL/QDiff
d968cbc47fe926b7f88b4adf10490f1edd6f8819
[ "BSD-3-Clause" ]
null
null
null
data/p3BR/R2/benchmark/startQiskit_noisy239.py
UCLA-SEAL/QDiff
d968cbc47fe926b7f88b4adf10490f1edd6f8819
[ "BSD-3-Clause" ]
null
null
null
# qubit number=3 # total number=45 import numpy as np from qiskit import QuantumCircuit, execute, Aer, QuantumRegister, ClassicalRegister, transpile, BasicAer, IBMQ from qiskit.visualization import plot_histogram from typing import * from pprint import pprint from math import log2 from collections import Counter from qiskit.test.mock import FakeVigo, FakeYorktown kernel = 'circuit/bernstein' def bitwise_xor(s: str, t: str) -> str: length = len(s) res = [] for i in range(length): res.append(str(int(s[i]) ^ int(t[i]))) return ''.join(res[::-1]) def bitwise_dot(s: str, t: str) -> str: length = len(s) res = 0 for i in range(length): res += int(s[i]) * int(t[i]) return str(res % 2) def build_oracle(n: int, f: Callable[[str], str]) -> QuantumCircuit: # implement the oracle O_f # NOTE: use multi_control_toffoli_gate ('noancilla' mode) # https://qiskit.org/documentation/_modules/qiskit/aqua/circuits/gates/multi_control_toffoli_gate.html # https://quantumcomputing.stackexchange.com/questions/3943/how-do-you-implement-the-toffoli-gate-using-only-single-qubit-and-cnot-gates # https://quantumcomputing.stackexchange.com/questions/2177/how-can-i-implement-an-n-bit-toffoli-gate controls = QuantumRegister(n, "ofc") target = QuantumRegister(1, "oft") oracle = QuantumCircuit(controls, target, name="Of") for i in range(2 ** n): rep = np.binary_repr(i, n) if f(rep) == "1": for j in range(n): if rep[j] == "0": oracle.x(controls[j]) oracle.mct(controls, target[0], None, mode='noancilla') for j in range(n): if rep[j] == "0": oracle.x(controls[j]) # oracle.barrier() # oracle.draw('mpl', filename=(kernel + '-oracle.png')) return oracle def build_circuit(n: int, f: Callable[[str], str]) -> QuantumCircuit: # implement the Bernstein-Vazirani circuit zero = np.binary_repr(0, n) b = f(zero) # initial n + 1 bits input_qubit = QuantumRegister(n+1, "qc") classicals = ClassicalRegister(n, "qm") prog = QuantumCircuit(input_qubit, classicals) # inverse last one (can be omitted if using O_f^\pm) prog.x(input_qubit[n]) # circuit begin prog.h(input_qubit[1]) # number=1 prog.h(input_qubit[2]) # number=38 prog.cz(input_qubit[0],input_qubit[2]) # number=39 prog.h(input_qubit[2]) # number=40 prog.h(input_qubit[2]) # number=42 prog.cz(input_qubit[0],input_qubit[2]) # number=43 prog.h(input_qubit[2]) # number=44 prog.cx(input_qubit[0],input_qubit[2]) # number=35 prog.x(input_qubit[2]) # number=36 prog.cx(input_qubit[0],input_qubit[2]) # number=37 prog.cx(input_qubit[0],input_qubit[2]) # number=33 prog.h(input_qubit[2]) # number=25 prog.cz(input_qubit[0],input_qubit[2]) # number=26 prog.h(input_qubit[2]) # number=27 prog.h(input_qubit[1]) # number=7 prog.cz(input_qubit[2],input_qubit[1]) # number=8 prog.rx(0.17592918860102857,input_qubit[2]) # number=34 prog.rx(-0.3989822670059037,input_qubit[1]) # number=30 prog.h(input_qubit[1]) # number=9 prog.h(input_qubit[1]) # number=18 prog.cz(input_qubit[2],input_qubit[1]) # number=19 prog.h(input_qubit[1]) # number=20 prog.y(input_qubit[1]) # number=14 prog.h(input_qubit[1]) # number=22 prog.cz(input_qubit[2],input_qubit[1]) # number=23 prog.h(input_qubit[1]) # number=24 prog.z(input_qubit[2]) # number=3 prog.z(input_qubit[1]) # number=41 prog.x(input_qubit[1]) # number=17 prog.y(input_qubit[2]) # number=5 prog.x(input_qubit[2]) # number=21 # apply H to get superposition for i in range(n): prog.h(input_qubit[i]) prog.h(input_qubit[n]) prog.barrier() # apply oracle O_f oracle = build_oracle(n, f) prog.append( oracle.to_gate(), [input_qubit[i] for i in range(n)] + [input_qubit[n]]) # apply H back (QFT on Z_2^n) for i in range(n): prog.h(input_qubit[i]) prog.barrier() # measure return prog def get_statevector(prog: QuantumCircuit) -> Any: state_backend = Aer.get_backend('statevector_simulator') statevec = execute(prog, state_backend).result() quantum_state = statevec.get_statevector() qubits = round(log2(len(quantum_state))) quantum_state = { "|" + np.binary_repr(i, qubits) + ">": quantum_state[i] for i in range(2 ** qubits) } return quantum_state def evaluate(backend_str: str, prog: QuantumCircuit, shots: int, b: str) -> Any: # Q: which backend should we use? # get state vector quantum_state = get_statevector(prog) # get simulate results # provider = IBMQ.load_account() # backend = provider.get_backend(backend_str) # qobj = compile(prog, backend, shots) # job = backend.run(qobj) # job.result() backend = Aer.get_backend(backend_str) # transpile/schedule -> assemble -> backend.run results = execute(prog, backend, shots=shots).result() counts = results.get_counts() a = Counter(counts).most_common(1)[0][0][::-1] return { "measurements": counts, # "state": statevec, "quantum_state": quantum_state, "a": a, "b": b } def bernstein_test_1(rep: str): """011 . x + 1""" a = "011" b = "1" return bitwise_xor(bitwise_dot(a, rep), b) def bernstein_test_2(rep: str): """000 . x + 0""" a = "000" b = "0" return bitwise_xor(bitwise_dot(a, rep), b) def bernstein_test_3(rep: str): """111 . x + 1""" a = "111" b = "1" return bitwise_xor(bitwise_dot(a, rep), b) if __name__ == "__main__": n = 2 a = "11" b = "1" f = lambda rep: \ bitwise_xor(bitwise_dot(a, rep), b) prog = build_circuit(n, f) sample_shot =4000 writefile = open("../data/startQiskit_noisy239.csv", "w") # prog.draw('mpl', filename=(kernel + '.png')) backend = FakeYorktown() circuit1 = transpile(prog, FakeYorktown()) circuit1.h(qubit=2) circuit1.x(qubit=3) circuit1.measure_all() info = execute(circuit1,backend=backend, shots=sample_shot).result().get_counts() print(info, file=writefile) print("results end", file=writefile) print(circuit1.depth(), file=writefile) print(circuit1, file=writefile) writefile.close()
30.799043
140
0.634457
import numpy as np from qiskit import QuantumCircuit, execute, Aer, QuantumRegister, ClassicalRegister, transpile, BasicAer, IBMQ from qiskit.visualization import plot_histogram from typing import * from pprint import pprint from math import log2 from collections import Counter from qiskit.test.mock import FakeVigo, FakeYorktown kernel = 'circuit/bernstein' def bitwise_xor(s: str, t: str) -> str: length = len(s) res = [] for i in range(length): res.append(str(int(s[i]) ^ int(t[i]))) return ''.join(res[::-1]) def bitwise_dot(s: str, t: str) -> str: length = len(s) res = 0 for i in range(length): res += int(s[i]) * int(t[i]) return str(res % 2) def build_oracle(n: int, f: Callable[[str], str]) -> QuantumCircuit: controls = QuantumRegister(n, "ofc") target = QuantumRegister(1, "oft") oracle = QuantumCircuit(controls, target, name="Of") for i in range(2 ** n): rep = np.binary_repr(i, n) if f(rep) == "1": for j in range(n): if rep[j] == "0": oracle.x(controls[j]) oracle.mct(controls, target[0], None, mode='noancilla') for j in range(n): if rep[j] == "0": oracle.x(controls[j]) return oracle def build_circuit(n: int, f: Callable[[str], str]) -> QuantumCircuit: zero = np.binary_repr(0, n) b = f(zero) input_qubit = QuantumRegister(n+1, "qc") classicals = ClassicalRegister(n, "qm") prog = QuantumCircuit(input_qubit, classicals) prog.x(input_qubit[n]) prog.h(input_qubit[1]) prog.h(input_qubit[2]) prog.cz(input_qubit[0],input_qubit[2]) prog.h(input_qubit[2]) prog.h(input_qubit[2]) prog.cz(input_qubit[0],input_qubit[2]) prog.h(input_qubit[2]) prog.cx(input_qubit[0],input_qubit[2]) prog.x(input_qubit[2]) prog.cx(input_qubit[0],input_qubit[2]) prog.cx(input_qubit[0],input_qubit[2]) prog.h(input_qubit[2]) prog.cz(input_qubit[0],input_qubit[2]) prog.h(input_qubit[2]) prog.h(input_qubit[1]) prog.cz(input_qubit[2],input_qubit[1]) prog.rx(0.17592918860102857,input_qubit[2]) prog.rx(-0.3989822670059037,input_qubit[1]) prog.h(input_qubit[1]) prog.h(input_qubit[1]) prog.cz(input_qubit[2],input_qubit[1]) prog.h(input_qubit[1]) prog.y(input_qubit[1]) prog.h(input_qubit[1]) prog.cz(input_qubit[2],input_qubit[1]) prog.h(input_qubit[1]) prog.z(input_qubit[2]) prog.z(input_qubit[1]) prog.x(input_qubit[1]) prog.y(input_qubit[2]) prog.x(input_qubit[2]) for i in range(n): prog.h(input_qubit[i]) prog.h(input_qubit[n]) prog.barrier() oracle = build_oracle(n, f) prog.append( oracle.to_gate(), [input_qubit[i] for i in range(n)] + [input_qubit[n]]) for i in range(n): prog.h(input_qubit[i]) prog.barrier() return prog def get_statevector(prog: QuantumCircuit) -> Any: state_backend = Aer.get_backend('statevector_simulator') statevec = execute(prog, state_backend).result() quantum_state = statevec.get_statevector() qubits = round(log2(len(quantum_state))) quantum_state = { "|" + np.binary_repr(i, qubits) + ">": quantum_state[i] for i in range(2 ** qubits) } return quantum_state def evaluate(backend_str: str, prog: QuantumCircuit, shots: int, b: str) -> Any: quantum_state = get_statevector(prog) backend = Aer.get_backend(backend_str) results = execute(prog, backend, shots=shots).result() counts = results.get_counts() a = Counter(counts).most_common(1)[0][0][::-1] return { "measurements": counts, "quantum_state": quantum_state, "a": a, "b": b } def bernstein_test_1(rep: str): a = "011" b = "1" return bitwise_xor(bitwise_dot(a, rep), b) def bernstein_test_2(rep: str): a = "000" b = "0" return bitwise_xor(bitwise_dot(a, rep), b) def bernstein_test_3(rep: str): a = "111" b = "1" return bitwise_xor(bitwise_dot(a, rep), b) if __name__ == "__main__": n = 2 a = "11" b = "1" f = lambda rep: \ bitwise_xor(bitwise_dot(a, rep), b) prog = build_circuit(n, f) sample_shot =4000 writefile = open("../data/startQiskit_noisy239.csv", "w") backend = FakeYorktown() circuit1 = transpile(prog, FakeYorktown()) circuit1.h(qubit=2) circuit1.x(qubit=3) circuit1.measure_all() info = execute(circuit1,backend=backend, shots=sample_shot).result().get_counts() print(info, file=writefile) print("results end", file=writefile) print(circuit1.depth(), file=writefile) print(circuit1, file=writefile) writefile.close()
true
true
1c486da49bd95ed0382213bd70102161815cc3de
3,912
py
Python
google/cloud/pubsublite/v1/pubsublite-v1-py/google/cloud/pubsublite_v1/types/publisher.py
googleapis/googleapis-gen
d84824c78563d59b0e58d5664bfaa430e9ad7e7a
[ "Apache-2.0" ]
7
2021-02-21T10:39:41.000Z
2021-12-07T07:31:28.000Z
google/cloud/pubsublite/v1/pubsublite-v1-py/google/cloud/pubsublite_v1/types/publisher.py
googleapis/googleapis-gen
d84824c78563d59b0e58d5664bfaa430e9ad7e7a
[ "Apache-2.0" ]
6
2021-02-02T23:46:11.000Z
2021-11-15T01:46:02.000Z
google/cloud/pubsublite/v1/pubsublite-v1-py/google/cloud/pubsublite_v1/types/publisher.py
googleapis/googleapis-gen
d84824c78563d59b0e58d5664bfaa430e9ad7e7a
[ "Apache-2.0" ]
4
2021-01-28T23:25:45.000Z
2021-08-30T01:55:16.000Z
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import proto # type: ignore from google.cloud.pubsublite_v1.types import common __protobuf__ = proto.module( package='google.cloud.pubsublite.v1', manifest={ 'InitialPublishRequest', 'InitialPublishResponse', 'MessagePublishRequest', 'MessagePublishResponse', 'PublishRequest', 'PublishResponse', }, ) class InitialPublishRequest(proto.Message): r"""The first request that must be sent on a newly-opened stream. Attributes: topic (str): The topic to which messages will be written. partition (int): The partition within the topic to which messages will be written. Partitions are zero indexed, so ``partition`` must be in the range [0, topic.num_partitions). """ topic = proto.Field( proto.STRING, number=1, ) partition = proto.Field( proto.INT64, number=2, ) class InitialPublishResponse(proto.Message): r"""Response to an InitialPublishRequest. """ class MessagePublishRequest(proto.Message): r"""Request to publish messages to the topic. Attributes: messages (Sequence[google.cloud.pubsublite_v1.types.PubSubMessage]): The messages to publish. """ messages = proto.RepeatedField( proto.MESSAGE, number=1, message=common.PubSubMessage, ) class MessagePublishResponse(proto.Message): r"""Response to a MessagePublishRequest. Attributes: start_cursor (google.cloud.pubsublite_v1.types.Cursor): The cursor of the first published message in the batch. The cursors for any remaining messages in the batch are guaranteed to be sequential. """ start_cursor = proto.Field( proto.MESSAGE, number=1, message=common.Cursor, ) class PublishRequest(proto.Message): r"""Request sent from the client to the server on a stream. Attributes: initial_request (google.cloud.pubsublite_v1.types.InitialPublishRequest): Initial request on the stream. message_publish_request (google.cloud.pubsublite_v1.types.MessagePublishRequest): Request to publish messages. """ initial_request = proto.Field( proto.MESSAGE, number=1, oneof='request_type', message='InitialPublishRequest', ) message_publish_request = proto.Field( proto.MESSAGE, number=2, oneof='request_type', message='MessagePublishRequest', ) class PublishResponse(proto.Message): r"""Response to a PublishRequest. Attributes: initial_response (google.cloud.pubsublite_v1.types.InitialPublishResponse): Initial response on the stream. message_response (google.cloud.pubsublite_v1.types.MessagePublishResponse): Response to publishing messages. """ initial_response = proto.Field( proto.MESSAGE, number=1, oneof='response_type', message='InitialPublishResponse', ) message_response = proto.Field( proto.MESSAGE, number=2, oneof='response_type', message='MessagePublishResponse', ) __all__ = tuple(sorted(__protobuf__.manifest))
27.356643
89
0.660532
import proto from google.cloud.pubsublite_v1.types import common __protobuf__ = proto.module( package='google.cloud.pubsublite.v1', manifest={ 'InitialPublishRequest', 'InitialPublishResponse', 'MessagePublishRequest', 'MessagePublishResponse', 'PublishRequest', 'PublishResponse', }, ) class InitialPublishRequest(proto.Message): topic = proto.Field( proto.STRING, number=1, ) partition = proto.Field( proto.INT64, number=2, ) class InitialPublishResponse(proto.Message): class MessagePublishRequest(proto.Message): messages = proto.RepeatedField( proto.MESSAGE, number=1, message=common.PubSubMessage, ) class MessagePublishResponse(proto.Message): start_cursor = proto.Field( proto.MESSAGE, number=1, message=common.Cursor, ) class PublishRequest(proto.Message): initial_request = proto.Field( proto.MESSAGE, number=1, oneof='request_type', message='InitialPublishRequest', ) message_publish_request = proto.Field( proto.MESSAGE, number=2, oneof='request_type', message='MessagePublishRequest', ) class PublishResponse(proto.Message): initial_response = proto.Field( proto.MESSAGE, number=1, oneof='response_type', message='InitialPublishResponse', ) message_response = proto.Field( proto.MESSAGE, number=2, oneof='response_type', message='MessagePublishResponse', ) __all__ = tuple(sorted(__protobuf__.manifest))
true
true
1c486f1cf2689297fa04af34f08c7bca55b19f3d
15,971
py
Python
src/c4/cmany/project.py
biojppm/cmany
b20c24169d60077122ae29a0c09526913340fd5c
[ "MIT" ]
20
2017-05-17T18:43:08.000Z
2021-02-13T16:20:53.000Z
src/c4/cmany/project.py
biojppm/cmany
b20c24169d60077122ae29a0c09526913340fd5c
[ "MIT" ]
8
2017-06-04T17:01:06.000Z
2022-03-17T12:43:32.000Z
src/c4/cmany/project.py
biojppm/cmany
b20c24169d60077122ae29a0c09526913340fd5c
[ "MIT" ]
1
2017-06-04T13:09:19.000Z
2017-06-04T13:09:19.000Z
#!/usr/bin/env python3 import os import glob import json import copy import timeit from collections import OrderedDict as odict from ruamel import yaml as yaml from ruamel.yaml.comments import CommentedMap as CommentedMap from . import util from . import conf from .build_flags import BuildFlags from .build_item import BuildItem from .build_type import BuildType from .system import System from .architecture import Architecture from .compiler import Compiler from .variant import Variant from .build import Build from .combination_rules import CombinationRules from .cmake import getcachevars from . import cmake from . import err from .util import path_exists as _pexists from .util import logdbg as dbg # ----------------------------------------------------------------------------- def _getdir(attr_name, default, kwargs, cwd): d = kwargs.get(attr_name) if d is None: d = os.path.join(cwd, default) else: if not os.path.isabs(d): d = os.path.join(cwd, d) d = util.abspath(d) return d # ----------------------------------------------------------------------------- class Project: def __init__(self, **kwargs): # self.kwargs = kwargs self.num_jobs = kwargs.get('jobs') self.targets = kwargs.get('target') self.continue_on_fail = kwargs.get('continue') # cwd = util.abspath(os.getcwd()) pdir = kwargs.get('proj_dir') dbg("cwd:", cwd) dbg("proj_dir:", pdir) if pdir is None: raise err.ProjDirNotFound(None) if pdir == ".": pdir = cwd pdir = util.abspath(pdir) dbg("proj_dir, abs:", pdir) # if not _pexists(pdir): raise err.ProjDirNotFound(pdir) # self.cmakelists = os.path.join(pdir, "CMakeLists.txt") cmakecache = None if _pexists(self.cmakelists): dbg("found CMakeLists.txt:", self.cmakelists) self.build_dir = _getdir('build_dir', 'build', kwargs, cwd) self.install_dir = _getdir('install_dir', 'install', kwargs, cwd) self.root_dir = pdir elif _pexists(pdir, "CMakeCache.txt"): cmakecache = os.path.join(pdir, "CMakeCache.txt") dbg("found CMakeCache.txt:", cmakecache) ch = cmake.CMakeCache(pdir) self.build_dir = os.path.dirname(pdir) self.install_dir = os.path.dirname(ch['CMAKE_INSTALL_PREFIX'].val) self.root_dir = ch['CMAKE_HOME_DIRECTORY'].val self.cmakelists = os.path.join(self.root_dir, "CMakeLists.txt") # self.root_dir = os.path.realpath(self.root_dir) self.build_dir = os.path.realpath(self.build_dir) self.install_dir = os.path.realpath(self.install_dir) self.cmakelists = os.path.realpath(self.cmakelists) # dbg("root_dir:", self.root_dir) dbg("build_dir:", self.build_dir) dbg("install_dir:", self.install_dir) dbg("CMakeLists.txt:", self.cmakelists) # if not _pexists(self.cmakelists): raise err.CMakeListsNotFound(pdir) # if cmakecache is not None: self._init_with_build_dir(os.path.dirname(cmakecache), **kwargs) elif cmake.hascache(self.root_dir): self._init_with_build_dir(self.root_dir, **kwargs) elif kwargs.get('glob'): self._init_with_glob(**kwargs) else: self.load_configs() self._init_with_build_items(**kwargs) def _init_with_build_dir(self, pdir, **kwargs): build = Build.deserialize(pdir) self.builds = [build] def _init_with_glob(self, **kwargs): g = kwargs.get('glob') self.builds = [] for pattern in g: bp = os.path.join(self.build_dir, pattern) li = glob.glob(bp) for b in li: build = Build.deserialize(b) self.builds.append(build) def _init_with_build_items(self, **kwargs): s, a, c, t, v = __class__.get_build_items(**kwargs) # cr = CombinationRules(kwargs.get('combination_rules', [])) combs = cr.valid_combinations(s, a, c, t, v) dbg("combinations:", combs) self.combination_rules = cr # self.builds = [] for comb in combs: dbg("adding build from combination:", comb) self.add_build(*comb) #s_, a_, c_, t_, v_) # self.systems = s self.architectures = a self.compilers = c self.build_types = t self.variants = v # # add new build params as needed to deal with adjusted builds def _addnew(b, name): a = getattr(b, name) ali = getattr(self, name + 's') if not [elm for elm in ali if str(elm) == str(a)]: ali.append(a) for b in self.builds: if not b.adjusted: continue _addnew(b, 'system') _addnew(b, 'architecture') _addnew(b, 'build_type') _addnew(b, 'compiler') _addnew(b, 'variant') @staticmethod def get_build_items(**kwargs): d = odict() for c, cls in ( ('systems', System), ('architectures', Architecture), ('compilers', Compiler), ('build_types', BuildType), ('variants', Variant)): d[c] = (cls, kwargs.get(c)) coll = BuildItem.create(d) s = coll['systems'] a = coll['architectures'] c = coll['compilers'] t = coll['build_types'] v = coll['variants'] return s, a, c, t, v def load_configs(self): seq = [os.path.join(d, "cmany.yml") for d in ( conf.CONF_DIR, conf.USER_DIR, self.root_dir)] if self.kwargs.get('no_default_config'): seq = [] for f in self.kwargs.get('config_file', []): ff = f if not os.path.isabs(ff): ff = os.path.join(self.root_dir, ff) if not os.path.exists(ff): raise err.ConfigFileNotFound(ff) seq.append(f) self.configs = conf.Configs.load_seq(seq) def save_configs(self): # c = Configs() pass def create_proj(self): yml = CommentedMap() yml['project'] = CommentedMap() # def _add(name): items = getattr(self, name) #if BuildItem.trivial_item(items): # yml['project'][name] = "_default_" #elif BuildItem.no_flags_in_collection(items): if BuildItem.no_flags_in_collection(items): out = [] for s in items: out.append(s.name) yml['project'][name] = out else: out = [] for s in items: cm = CommentedMap() cm[s.name] = CommentedMap() s.save_config(cm[s.name]) out.append(cm) yml['project'][name] = out # _add('systems') _add('architectures') _add('compilers') _add('build_types') _add('variants') txt = yaml.round_trip_dump(yml) fn = self.kwargs['output_file'] if not os.path.isabs(fn): fn = os.path.join(self.root_dir, fn) with open(fn, "w") as f: f.write(txt) def add_build(self, system, arch, compiler, build_type, variant): # duplicate the build items, as they may be mutated due # to translation of their flags for the compiler def _dup_item(item): i = copy.deepcopy(item) i.flags.resolve_flag_aliases(compiler, aliases=self.configs.flag_aliases) return i s = _dup_item(system) a = _dup_item(arch) t = _dup_item(build_type) c = _dup_item(compiler) v = _dup_item(variant) # f = BuildFlags('all_builds', **self.kwargs) f.resolve_flag_aliases(compiler, aliases=self.configs.flag_aliases) # # create the build dbg("adding build:", s, a, t, c, v, f) b = Build(self.root_dir, self.build_dir, self.install_dir, s, a, t, c, v, f, self.num_jobs, dict(self.kwargs)) # # When a build is created, its parameters may have been adjusted # because of an incompatible generator specification. # So drop this build if an equal one already exists if b.adjusted and self.exists(b): return False # a similar build already exists # # finally, this. self.builds.append(b) return True # build successfully added def exists(self, build): for b in self.builds: if str(b.tag) == str(build.tag): return True return False def select(self, **kwargs): out = [b for b in self.builds] def _h(kw, attr): global out g = kwargs.get(kw) if g is not None: lo = [] for b in out: if str(getattr(b, attr)) == str(g): lo.append(b) out = lo _h("sys", "system") _h("arch", "architecture") _h("compiler", "compiler") _h("build_type", "build_type") _h("variant", "variant") return out def create_tree(self, **restrict_to): builds = self.select(**restrict_to) for b in builds: b.create_dir() b.create_preload_file() # print(b, ":", d) def configure(self, **restrict_to): if not os.path.exists(self.build_dir): os.makedirs(self.build_dir) self._execute(Build.configure, "Configure", silent=False, **restrict_to) def reconfigure(self, **restrict_to): if not os.path.exists(self.build_dir): os.makedirs(self.build_dir) self._execute(Build.reconfigure, "Reconfigure", silent=False, **restrict_to) def export_compile_commands(self, **restrict_to): if not os.path.exists(self.build_dir): os.makedirs(self.build_dir) self._execute(Build.export_compile_commands, "Export compile commands", silent=False, **restrict_to) def build(self, **restrict_to): def do_build(build): build.build(self.targets) self._execute(do_build, "Build", silent=False, **restrict_to) def rebuild(self, **restrict_to): def do_rebuild(build): build.rebuild(self.targets) self._execute(do_rebuild, "Rebuild", silent=False, **restrict_to) def clean(self, **restrict_to): self._execute(Build.clean, "Clean", silent=False, **restrict_to) def install(self, **restrict_to): self._execute(Build.install, "Install", silent=False, **restrict_to) def reinstall(self, **restrict_to): self._execute(Build.reinstall, "Reinstall", silent=False, **restrict_to) def run_cmd(self, cmd, **subprocess_args): def run_it(build): build.run_custom_cmd(cmd, **subprocess_args) self._execute(run_it, "Run cmd", silent=False) def export_vs(self): confs = [] for b in self.builds: confs.append(b.json_data()) jd = odict([('configurations', confs)]) with open(self.configfile, 'w') as f: json.dump(jd, f, indent=2) def show_vars(self, varlist): varv = odict() pat = os.path.join(self.build_dir, '*', 'CMakeCache.txt') g = glob.glob(pat) md = 0 mv = 0 for p in g: d = os.path.dirname(p) b = os.path.basename(d) md = max(md, len(b)) vars = getcachevars(d, varlist) for k, v in vars.items(): sk = str(k) if not varv.get(sk): varv[sk] = odict() varv[sk][b] = v mv = max(mv, len(sk)) # fmt = "{:" + str(mv) + "}[{:" + str(md) + "}]={}" for var, sysvalues in varv.items(): for s, v in sysvalues.items(): print(fmt.format(var, s, v)) def show_build_names(self): for b in self.builds: print(b) def show_build_dirs(self): for b in self.builds: print(b.builddir) def show_builds(self): for b in self.builds: b.show_properties() def show_targets(self): for t in self.builds[0].get_targets(): print(t) def _execute(self, fn, msg, silent, **restrict_to): builds = self.select(**restrict_to) failed = odict() durations = odict() num = len(builds) if not silent: if num == 0: print("no builds selected") if num == 0: return def nt(*args, **kwargs): # notice if silent: return util.lognotice(*args, **kwargs) def dn(*args, **kwargs): # done if silent: return util.logdone(*args, **kwargs) def er(*args, **kwargs): # error if silent: return util.logerr(*args, **kwargs) # if num > 1: nt("") nt("===============================================") nt(msg + ": start", num, "builds:") for b in builds: nt(b) nt("===============================================") # for i, b in enumerate(builds): if i > 0: nt("\n") nt("-----------------------------------------------") if num > 1: nt(msg + ": build #{} of {}:".format(i + 1, num), b) else: nt(msg, b) nt("-----------------------------------------------") # t = timeit.default_timer() try: # this is where it happens fn(b) # <-- here word, logger = "finished", dn # exceptions thrown from builds inherit this type except err.BuildError as e: word, logger = "failed", er util.logerr(f"{b} failed! {e}") failed[b] = e if not self.continue_on_fail: raise t = timeit.default_timer() - t hrt = util.human_readable_time(t) durations[b] = (t, hrt) if num > 1: ip1 = i + 1 info = f"{word} build #{ip1} of {num} ({hrt})" else: info = f"{word} building ({hrt})" logger(msg + ": " + info + ":", b) # nt("-----------------------------------------------") if num > 1: if failed: dn(msg + ": processed", num, "builds: (with failures)") else: dn(msg + ": finished", num, "builds:") tot = 0. for _, (d, _) in durations.items(): tot += d for b in builds: dur, hrt = durations[b] times = "({}, {:.3f}%, {:.3f}x avg)".format( hrt, dur / tot * 100., dur / (tot / float(num)) ) fail = failed.get(b) if fail: er(b, times, "[FAIL]!!!", fail) else: dn(b, times) if failed: msg = "{}/{} builds failed ({:.1f}%)!" er(msg.format(len(failed), num, float(len(failed)) / num * 100.0)) else: dn(f"all {num} builds succeeded!") dn("total time:", util.human_readable_time(tot)) nt("===============================================") if failed: raise Exception(failed)
34.420259
108
0.508547
import os import glob import json import copy import timeit from collections import OrderedDict as odict from ruamel import yaml as yaml from ruamel.yaml.comments import CommentedMap as CommentedMap from . import util from . import conf from .build_flags import BuildFlags from .build_item import BuildItem from .build_type import BuildType from .system import System from .architecture import Architecture from .compiler import Compiler from .variant import Variant from .build import Build from .combination_rules import CombinationRules from .cmake import getcachevars from . import cmake from . import err from .util import path_exists as _pexists from .util import logdbg as dbg def _getdir(attr_name, default, kwargs, cwd): d = kwargs.get(attr_name) if d is None: d = os.path.join(cwd, default) else: if not os.path.isabs(d): d = os.path.join(cwd, d) d = util.abspath(d) return d class Project: def __init__(self, **kwargs): self.kwargs = kwargs self.num_jobs = kwargs.get('jobs') self.targets = kwargs.get('target') self.continue_on_fail = kwargs.get('continue') cwd = util.abspath(os.getcwd()) pdir = kwargs.get('proj_dir') dbg("cwd:", cwd) dbg("proj_dir:", pdir) if pdir is None: raise err.ProjDirNotFound(None) if pdir == ".": pdir = cwd pdir = util.abspath(pdir) dbg("proj_dir, abs:", pdir) if not _pexists(pdir): raise err.ProjDirNotFound(pdir) self.cmakelists = os.path.join(pdir, "CMakeLists.txt") cmakecache = None if _pexists(self.cmakelists): dbg("found CMakeLists.txt:", self.cmakelists) self.build_dir = _getdir('build_dir', 'build', kwargs, cwd) self.install_dir = _getdir('install_dir', 'install', kwargs, cwd) self.root_dir = pdir elif _pexists(pdir, "CMakeCache.txt"): cmakecache = os.path.join(pdir, "CMakeCache.txt") dbg("found CMakeCache.txt:", cmakecache) ch = cmake.CMakeCache(pdir) self.build_dir = os.path.dirname(pdir) self.install_dir = os.path.dirname(ch['CMAKE_INSTALL_PREFIX'].val) self.root_dir = ch['CMAKE_HOME_DIRECTORY'].val self.cmakelists = os.path.join(self.root_dir, "CMakeLists.txt") self.root_dir = os.path.realpath(self.root_dir) self.build_dir = os.path.realpath(self.build_dir) self.install_dir = os.path.realpath(self.install_dir) self.cmakelists = os.path.realpath(self.cmakelists) dbg("root_dir:", self.root_dir) dbg("build_dir:", self.build_dir) dbg("install_dir:", self.install_dir) dbg("CMakeLists.txt:", self.cmakelists) if not _pexists(self.cmakelists): raise err.CMakeListsNotFound(pdir) if cmakecache is not None: self._init_with_build_dir(os.path.dirname(cmakecache), **kwargs) elif cmake.hascache(self.root_dir): self._init_with_build_dir(self.root_dir, **kwargs) elif kwargs.get('glob'): self._init_with_glob(**kwargs) else: self.load_configs() self._init_with_build_items(**kwargs) def _init_with_build_dir(self, pdir, **kwargs): build = Build.deserialize(pdir) self.builds = [build] def _init_with_glob(self, **kwargs): g = kwargs.get('glob') self.builds = [] for pattern in g: bp = os.path.join(self.build_dir, pattern) li = glob.glob(bp) for b in li: build = Build.deserialize(b) self.builds.append(build) def _init_with_build_items(self, **kwargs): s, a, c, t, v = __class__.get_build_items(**kwargs) cr = CombinationRules(kwargs.get('combination_rules', [])) combs = cr.valid_combinations(s, a, c, t, v) dbg("combinations:", combs) self.combination_rules = cr self.builds = [] for comb in combs: dbg("adding build from combination:", comb) self.add_build(*comb) self.systems = s self.architectures = a self.compilers = c self.build_types = t self.variants = v def _addnew(b, name): a = getattr(b, name) ali = getattr(self, name + 's') if not [elm for elm in ali if str(elm) == str(a)]: ali.append(a) for b in self.builds: if not b.adjusted: continue _addnew(b, 'system') _addnew(b, 'architecture') _addnew(b, 'build_type') _addnew(b, 'compiler') _addnew(b, 'variant') @staticmethod def get_build_items(**kwargs): d = odict() for c, cls in ( ('systems', System), ('architectures', Architecture), ('compilers', Compiler), ('build_types', BuildType), ('variants', Variant)): d[c] = (cls, kwargs.get(c)) coll = BuildItem.create(d) s = coll['systems'] a = coll['architectures'] c = coll['compilers'] t = coll['build_types'] v = coll['variants'] return s, a, c, t, v def load_configs(self): seq = [os.path.join(d, "cmany.yml") for d in ( conf.CONF_DIR, conf.USER_DIR, self.root_dir)] if self.kwargs.get('no_default_config'): seq = [] for f in self.kwargs.get('config_file', []): ff = f if not os.path.isabs(ff): ff = os.path.join(self.root_dir, ff) if not os.path.exists(ff): raise err.ConfigFileNotFound(ff) seq.append(f) self.configs = conf.Configs.load_seq(seq) def save_configs(self): pass def create_proj(self): yml = CommentedMap() yml['project'] = CommentedMap() def _add(name): items = getattr(self, name) if BuildItem.no_flags_in_collection(items): out = [] for s in items: out.append(s.name) yml['project'][name] = out else: out = [] for s in items: cm = CommentedMap() cm[s.name] = CommentedMap() s.save_config(cm[s.name]) out.append(cm) yml['project'][name] = out _add('systems') _add('architectures') _add('compilers') _add('build_types') _add('variants') txt = yaml.round_trip_dump(yml) fn = self.kwargs['output_file'] if not os.path.isabs(fn): fn = os.path.join(self.root_dir, fn) with open(fn, "w") as f: f.write(txt) def add_build(self, system, arch, compiler, build_type, variant): def _dup_item(item): i = copy.deepcopy(item) i.flags.resolve_flag_aliases(compiler, aliases=self.configs.flag_aliases) return i s = _dup_item(system) a = _dup_item(arch) t = _dup_item(build_type) c = _dup_item(compiler) v = _dup_item(variant) f = BuildFlags('all_builds', **self.kwargs) f.resolve_flag_aliases(compiler, aliases=self.configs.flag_aliases) dbg("adding build:", s, a, t, c, v, f) b = Build(self.root_dir, self.build_dir, self.install_dir, s, a, t, c, v, f, self.num_jobs, dict(self.kwargs)) if b.adjusted and self.exists(b): return False self.builds.append(b) return True def exists(self, build): for b in self.builds: if str(b.tag) == str(build.tag): return True return False def select(self, **kwargs): out = [b for b in self.builds] def _h(kw, attr): global out g = kwargs.get(kw) if g is not None: lo = [] for b in out: if str(getattr(b, attr)) == str(g): lo.append(b) out = lo _h("sys", "system") _h("arch", "architecture") _h("compiler", "compiler") _h("build_type", "build_type") _h("variant", "variant") return out def create_tree(self, **restrict_to): builds = self.select(**restrict_to) for b in builds: b.create_dir() b.create_preload_file() def configure(self, **restrict_to): if not os.path.exists(self.build_dir): os.makedirs(self.build_dir) self._execute(Build.configure, "Configure", silent=False, **restrict_to) def reconfigure(self, **restrict_to): if not os.path.exists(self.build_dir): os.makedirs(self.build_dir) self._execute(Build.reconfigure, "Reconfigure", silent=False, **restrict_to) def export_compile_commands(self, **restrict_to): if not os.path.exists(self.build_dir): os.makedirs(self.build_dir) self._execute(Build.export_compile_commands, "Export compile commands", silent=False, **restrict_to) def build(self, **restrict_to): def do_build(build): build.build(self.targets) self._execute(do_build, "Build", silent=False, **restrict_to) def rebuild(self, **restrict_to): def do_rebuild(build): build.rebuild(self.targets) self._execute(do_rebuild, "Rebuild", silent=False, **restrict_to) def clean(self, **restrict_to): self._execute(Build.clean, "Clean", silent=False, **restrict_to) def install(self, **restrict_to): self._execute(Build.install, "Install", silent=False, **restrict_to) def reinstall(self, **restrict_to): self._execute(Build.reinstall, "Reinstall", silent=False, **restrict_to) def run_cmd(self, cmd, **subprocess_args): def run_it(build): build.run_custom_cmd(cmd, **subprocess_args) self._execute(run_it, "Run cmd", silent=False) def export_vs(self): confs = [] for b in self.builds: confs.append(b.json_data()) jd = odict([('configurations', confs)]) with open(self.configfile, 'w') as f: json.dump(jd, f, indent=2) def show_vars(self, varlist): varv = odict() pat = os.path.join(self.build_dir, '*', 'CMakeCache.txt') g = glob.glob(pat) md = 0 mv = 0 for p in g: d = os.path.dirname(p) b = os.path.basename(d) md = max(md, len(b)) vars = getcachevars(d, varlist) for k, v in vars.items(): sk = str(k) if not varv.get(sk): varv[sk] = odict() varv[sk][b] = v mv = max(mv, len(sk)) fmt = "{:" + str(mv) + "}[{:" + str(md) + "}]={}" for var, sysvalues in varv.items(): for s, v in sysvalues.items(): print(fmt.format(var, s, v)) def show_build_names(self): for b in self.builds: print(b) def show_build_dirs(self): for b in self.builds: print(b.builddir) def show_builds(self): for b in self.builds: b.show_properties() def show_targets(self): for t in self.builds[0].get_targets(): print(t) def _execute(self, fn, msg, silent, **restrict_to): builds = self.select(**restrict_to) failed = odict() durations = odict() num = len(builds) if not silent: if num == 0: print("no builds selected") if num == 0: return def nt(*args, **kwargs): if silent: return util.lognotice(*args, **kwargs) def dn(*args, **kwargs): if silent: return util.logdone(*args, **kwargs) def er(*args, **kwargs): if silent: return util.logerr(*args, **kwargs) if num > 1: nt("") nt("===============================================") nt(msg + ": start", num, "builds:") for b in builds: nt(b) nt("===============================================") for i, b in enumerate(builds): if i > 0: nt("\n") nt("-----------------------------------------------") if num > 1: nt(msg + ": build #{} of {}:".format(i + 1, num), b) else: nt(msg, b) nt("-----------------------------------------------") t = timeit.default_timer() try: fn(b) word, logger = "finished", dn except err.BuildError as e: word, logger = "failed", er util.logerr(f"{b} failed! {e}") failed[b] = e if not self.continue_on_fail: raise t = timeit.default_timer() - t hrt = util.human_readable_time(t) durations[b] = (t, hrt) if num > 1: ip1 = i + 1 info = f"{word} build #{ip1} of {num} ({hrt})" else: info = f"{word} building ({hrt})" logger(msg + ": " + info + ":", b) nt("-----------------------------------------------") if num > 1: if failed: dn(msg + ": processed", num, "builds: (with failures)") else: dn(msg + ": finished", num, "builds:") tot = 0. for _, (d, _) in durations.items(): tot += d for b in builds: dur, hrt = durations[b] times = "({}, {:.3f}%, {:.3f}x avg)".format( hrt, dur / tot * 100., dur / (tot / float(num)) ) fail = failed.get(b) if fail: er(b, times, "[FAIL]!!!", fail) else: dn(b, times) if failed: msg = "{}/{} builds failed ({:.1f}%)!" er(msg.format(len(failed), num, float(len(failed)) / num * 100.0)) else: dn(f"all {num} builds succeeded!") dn("total time:", util.human_readable_time(tot)) nt("===============================================") if failed: raise Exception(failed)
true
true
1c486f22ea70cd1af40bf4a804d81f728c367b5b
10,630
py
Python
salt/states/tomcat.py
yuriks/salt
d2a5bd8adddb98ec1718d79384aa13b4f37e8028
[ "Apache-2.0", "MIT" ]
1
2020-03-31T22:51:16.000Z
2020-03-31T22:51:16.000Z
salt/states/tomcat.py
yuriks/salt
d2a5bd8adddb98ec1718d79384aa13b4f37e8028
[ "Apache-2.0", "MIT" ]
null
null
null
salt/states/tomcat.py
yuriks/salt
d2a5bd8adddb98ec1718d79384aa13b4f37e8028
[ "Apache-2.0", "MIT" ]
1
2021-09-30T07:00:01.000Z
2021-09-30T07:00:01.000Z
# -*- coding: utf-8 -*- ''' Manage Apache Tomcat web applications ===================================== .. note:: This state requires the Tomcat Manager webapp to be installed and running. The following grains/pillars must be set for communication with Tomcat Manager to work: .. code-block:: yaml tomcat-manager: user: 'tomcat-manager' passwd: 'Passw0rd' Configuring Tomcat Manager -------------------------- To manage webapps via the Tomcat Manager, you'll need to configure a valid user in the file ``conf/tomcat-users.xml``. .. code-block:: xml :caption: conf/tomcat-users.xml <?xml version='1.0' encoding='utf-8'?> <tomcat-users> <role rolename="manager-script"/> <user username="tomcat-manager" password="Passw0rd" roles="manager-script"/> </tomcat-users> Notes ----- - Using multiple versions (aka. parallel deployments) on the same context path is not supported. - More information about the Tomcat Manager: http://tomcat.apache.org/tomcat-7.0-doc/manager-howto.html - If you use only this module for deployments you might want to restrict access to the manager so it's only accessible via localhost. For more info: http://tomcat.apache.org/tomcat-7.0-doc/manager-howto.html#Configuring_Manager_Application_Access - Last tested on: Tomcat Version: Apache Tomcat/7.0.54 JVM Vendor: Oracle Corporation JVM Version: 1.8.0_101-b13 OS Architecture: amd64 OS Name: Linux OS Version: 3.10.0-327.22.2.el7.x86_64 ''' from __future__ import absolute_import, unicode_literals, print_function from salt.ext import six # Private def __virtual__(): ''' Load if the module tomcat exists ''' return 'tomcat' if 'tomcat.status' in __salt__ else False # Functions def war_deployed(name, war, force=False, url='http://localhost:8080/manager', timeout=180, temp_war_location=None, version=True): ''' Enforce that the WAR will be deployed and started in the context path, while making use of WAR versions in the filename. .. note:: For more info about Tomcats file paths and context naming, please see http://tomcat.apache.org/tomcat-7.0-doc/config/context.html#Naming name The context path to deploy (incl. forward slash) the WAR to. war Absolute path to WAR file (should be accessible by the user running Tomcat) or a path supported by the ``salt.modules.cp.get_url`` function. force : False Force deployment even if the version strings are the same. Disabled by default. url : http://localhost:8080/manager The URL of the Tomcat Web Application Manager. timeout : 180 Timeout for HTTP requests to the Tomcat Manager. temp_war_location : None Use another location to temporarily copy the WAR file to. By default the system's temp directory is used. version : '' Specify the WAR version. If this argument is provided, it overrides the version encoded in the WAR file name, if one is present. .. versionadded:: 2015.8.6 Use ``False`` or blank value to prevent guessing the version and keeping it blank. .. versionadded:: 2016.11.0 Example: .. code-block:: yaml jenkins: tomcat.war_deployed: - name: /salt-powered-jenkins - war: salt://jenkins-1.2.4.war - require: - service: application-service .. note:: Be aware that in the above example the WAR ``jenkins-1.2.4.war`` will be deployed to the context path ``salt-powered-jenkins##1.2.4``. To avoid this either specify a version yourself, or set version to ``False``. ''' # Prepare ret = {'name': name, 'result': True, 'changes': {}, 'comment': ''} # if version is defined or False, we don't want to overwrite if version is True: version = __salt__['tomcat.extract_war_version'](war) or '' elif not version: version = '' webapps = __salt__['tomcat.ls'](url, timeout) deploy = False undeploy = False status = True # Gathered/specified new WAR version string specified_ver = 'version {0}'.format(version) if version else 'no version' # Determine what to do try: # Printed version strings, here to throw exception if no webapps[name] current_ver = 'version ' + webapps[name]['version'] \ if webapps[name]['version'] else 'no version' # `endswith` on the supposed string will cause Exception if empty if (not webapps[name]['version'].endswith(version) or (version == '' and webapps[name]['version'] != version) or force): deploy = True undeploy = True ret['changes']['undeploy'] = ('undeployed {0} with {1}'. format(name, current_ver)) ret['changes']['deploy'] = ('will deploy {0} with {1}'. format(name, specified_ver)) else: deploy = False ret['comment'] = ('{0} with {1} is already deployed'. format(name, specified_ver)) if webapps[name]['mode'] != 'running': ret['changes']['start'] = 'starting {0}'.format(name) status = False else: return ret except Exception: # pylint: disable=broad-except deploy = True ret['changes']['deploy'] = ('deployed {0} with {1}'. format(name, specified_ver)) # Test if __opts__['test']: ret['result'] = None return ret # make sure the webapp is up if deployed if deploy is False: if status is False: ret['comment'] = __salt__['tomcat.start'](name, url, timeout=timeout) ret['result'] = ret['comment'].startswith('OK') return ret # Undeploy if undeploy: un = __salt__['tomcat.undeploy'](name, url, timeout=timeout) if un.startswith('FAIL'): ret['result'] = False ret['comment'] = un return ret # Deploy deploy_res = __salt__['tomcat.deploy_war'](war, name, 'yes', url, __env__, timeout, temp_war_location=temp_war_location, version=version) # Return if deploy_res.startswith('OK'): ret['result'] = True ret['comment'] = six.text_type(__salt__['tomcat.ls'](url, timeout)[name]) ret['changes']['deploy'] = ('deployed {0} with {1}'. format(name, specified_ver)) else: ret['result'] = False ret['comment'] = deploy_res ret['changes'].pop('deploy') return ret def wait(name, url='http://localhost:8080/manager', timeout=180): ''' Wait for the Tomcat Manager to load. Notice that if tomcat is not running we won't wait for it start and the state will fail. This state can be required in the tomcat.war_deployed state to make sure tomcat is running and that the manager is running as well and ready for deployment. url : http://localhost:8080/manager The URL of the server with the Tomcat Manager webapp. timeout : 180 Timeout for HTTP request to the Tomcat Manager. Example: .. code-block:: yaml tomcat-service: service.running: - name: tomcat - enable: True wait-for-tomcatmanager: tomcat.wait: - timeout: 300 - require: - service: tomcat-service jenkins: tomcat.war_deployed: - name: /ran - war: salt://jenkins-1.2.4.war - require: - tomcat: wait-for-tomcatmanager ''' result = __salt__['tomcat.status'](url, timeout) ret = {'name': name, 'result': result, 'changes': {}, 'comment': ('tomcat manager is ready' if result else 'tomcat manager is not ready') } return ret def mod_watch(name, url='http://localhost:8080/manager', timeout=180): ''' The tomcat watcher, called to invoke the watch command. When called, it will reload the webapp in question .. note:: This state exists to support special handling of the ``watch`` :ref:`requisite <requisites>`. It should not be called directly. Parameters for this function should be set by the state being triggered. ''' msg = __salt__['tomcat.reload'](name, url, timeout) result = msg.startswith('OK') ret = {'name': name, 'result': result, 'changes': {name: result}, 'comment': msg } return ret def undeployed(name, url='http://localhost:8080/manager', timeout=180): ''' Enforce that the WAR will be undeployed from the server name The context path to undeploy. url : http://localhost:8080/manager The URL of the server with the Tomcat Manager webapp. timeout : 180 Timeout for HTTP request to the Tomcat Manager. Example: .. code-block:: yaml jenkins: tomcat.undeployed: - name: /ran - require: - service: application-service ''' # Prepare ret = {'name': name, 'result': True, 'changes': {}, 'comment': ''} if not __salt__['tomcat.status'](url, timeout): ret['comment'] = 'Tomcat Manager does not respond' ret['result'] = False return ret try: version = __salt__['tomcat.ls'](url, timeout)[name]['version'] ret['changes'] = {'undeploy': version} except KeyError: return ret # Test if __opts__['test']: ret['result'] = None return ret undeploy = __salt__['tomcat.undeploy'](name, url, timeout=timeout) if undeploy.startswith('FAIL'): ret['result'] = False ret['comment'] = undeploy return ret return ret
30.371429
114
0.564722
from __future__ import absolute_import, unicode_literals, print_function from salt.ext import six def __virtual__(): return 'tomcat' if 'tomcat.status' in __salt__ else False def war_deployed(name, war, force=False, url='http://localhost:8080/manager', timeout=180, temp_war_location=None, version=True): ret = {'name': name, 'result': True, 'changes': {}, 'comment': ''} if version is True: version = __salt__['tomcat.extract_war_version'](war) or '' elif not version: version = '' webapps = __salt__['tomcat.ls'](url, timeout) deploy = False undeploy = False status = True # Gathered/specified new WAR version string specified_ver = 'version {0}'.format(version) if version else 'no version' # Determine what to do try: # Printed version strings, here to throw exception if no webapps[name] current_ver = 'version ' + webapps[name]['version'] \ if webapps[name]['version'] else 'no version' # `endswith` on the supposed string will cause Exception if empty if (not webapps[name]['version'].endswith(version) or (version == '' and webapps[name]['version'] != version) or force): deploy = True undeploy = True ret['changes']['undeploy'] = ('undeployed {0} with {1}'. format(name, current_ver)) ret['changes']['deploy'] = ('will deploy {0} with {1}'. format(name, specified_ver)) else: deploy = False ret['comment'] = ('{0} with {1} is already deployed'. format(name, specified_ver)) if webapps[name]['mode'] != 'running': ret['changes']['start'] = 'starting {0}'.format(name) status = False else: return ret except Exception: # pylint: disable=broad-except deploy = True ret['changes']['deploy'] = ('deployed {0} with {1}'. format(name, specified_ver)) # Test if __opts__['test']: ret['result'] = None return ret # make sure the webapp is up if deployed if deploy is False: if status is False: ret['comment'] = __salt__['tomcat.start'](name, url, timeout=timeout) ret['result'] = ret['comment'].startswith('OK') return ret # Undeploy if undeploy: un = __salt__['tomcat.undeploy'](name, url, timeout=timeout) if un.startswith('FAIL'): ret['result'] = False ret['comment'] = un return ret # Deploy deploy_res = __salt__['tomcat.deploy_war'](war, name, 'yes', url, __env__, timeout, temp_war_location=temp_war_location, version=version) # Return if deploy_res.startswith('OK'): ret['result'] = True ret['comment'] = six.text_type(__salt__['tomcat.ls'](url, timeout)[name]) ret['changes']['deploy'] = ('deployed {0} with {1}'. format(name, specified_ver)) else: ret['result'] = False ret['comment'] = deploy_res ret['changes'].pop('deploy') return ret def wait(name, url='http://localhost:8080/manager', timeout=180): result = __salt__['tomcat.status'](url, timeout) ret = {'name': name, 'result': result, 'changes': {}, 'comment': ('tomcat manager is ready' if result else 'tomcat manager is not ready') } return ret def mod_watch(name, url='http://localhost:8080/manager', timeout=180): msg = __salt__['tomcat.reload'](name, url, timeout) result = msg.startswith('OK') ret = {'name': name, 'result': result, 'changes': {name: result}, 'comment': msg } return ret def undeployed(name, url='http://localhost:8080/manager', timeout=180): # Prepare ret = {'name': name, 'result': True, 'changes': {}, 'comment': ''} if not __salt__['tomcat.status'](url, timeout): ret['comment'] = 'Tomcat Manager does not respond' ret['result'] = False return ret try: version = __salt__['tomcat.ls'](url, timeout)[name]['version'] ret['changes'] = {'undeploy': version} except KeyError: return ret # Test if __opts__['test']: ret['result'] = None return ret undeploy = __salt__['tomcat.undeploy'](name, url, timeout=timeout) if undeploy.startswith('FAIL'): ret['result'] = False ret['comment'] = undeploy return ret return ret
true
true
1c486f58fa570c63d63f4cb3f130a23e65f259d7
7,188
py
Python
bc/navigation/migrations/0001_initial.py
Buckinghamshire-Digital-Service/buckinghamshire-council
bbbdb52b515bcdfc79a2bd9198dfa4828405370e
[ "BSD-3-Clause" ]
1
2021-02-27T07:27:17.000Z
2021-02-27T07:27:17.000Z
bc/navigation/migrations/0001_initial.py
Buckinghamshire-Digital-Service/buckinghamshire-council
bbbdb52b515bcdfc79a2bd9198dfa4828405370e
[ "BSD-3-Clause" ]
null
null
null
bc/navigation/migrations/0001_initial.py
Buckinghamshire-Digital-Service/buckinghamshire-council
bbbdb52b515bcdfc79a2bd9198dfa4828405370e
[ "BSD-3-Clause" ]
1
2021-06-09T15:56:54.000Z
2021-06-09T15:56:54.000Z
# -*- coding: utf-8 -*- # Generated by Django 1.11.8 on 2017-12-20 09:28 from __future__ import unicode_literals import django.db.models.deletion from django.db import migrations, models import wagtail.core.blocks import wagtail.core.fields class Migration(migrations.Migration): initial = True dependencies = [("wagtailcore", "0040_page_draft_title")] operations = [ migrations.CreateModel( name="NavigationSettings", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ( "primary_navigation", wagtail.core.fields.StreamField( ( ( "link", wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ), ), ), blank=True, help_text="Main site navigation", ), ), ( "secondary_navigation", wagtail.core.fields.StreamField( ( ( "link", wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ), ), ), blank=True, help_text="Alternative navigation", ), ), ( "footer_navigation", wagtail.core.fields.StreamField( ( ( "column", wagtail.core.blocks.StructBlock( ( ( "heading", wagtail.core.blocks.CharBlock( required=False, help_text="Leave blank if no header required.", ), ), ( "links", wagtail.core.blocks.ListBlock( wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ) ), ), ) ), ), ), blank=True, help_text="Multiple columns of footer links with optional header.", ), ), ( "footer_links", wagtail.core.fields.StreamField( ( ( "link", wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ), ), ), blank=True, help_text="Single list of elements at the base of the page.", ), ), ( "site", models.OneToOneField( editable=False, on_delete=django.db.models.deletion.CASCADE, to="wagtailcore.Site", ), ), ], options={"abstract": False}, ) ]
43.041916
116
0.229967
from __future__ import unicode_literals import django.db.models.deletion from django.db import migrations, models import wagtail.core.blocks import wagtail.core.fields class Migration(migrations.Migration): initial = True dependencies = [("wagtailcore", "0040_page_draft_title")] operations = [ migrations.CreateModel( name="NavigationSettings", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ( "primary_navigation", wagtail.core.fields.StreamField( ( ( "link", wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ), ), ), blank=True, help_text="Main site navigation", ), ), ( "secondary_navigation", wagtail.core.fields.StreamField( ( ( "link", wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ), ), ), blank=True, help_text="Alternative navigation", ), ), ( "footer_navigation", wagtail.core.fields.StreamField( ( ( "column", wagtail.core.blocks.StructBlock( ( ( "heading", wagtail.core.blocks.CharBlock( required=False, help_text="Leave blank if no header required.", ), ), ( "links", wagtail.core.blocks.ListBlock( wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ) ), ), ) ), ), ), blank=True, help_text="Multiple columns of footer links with optional header.", ), ), ( "footer_links", wagtail.core.fields.StreamField( ( ( "link", wagtail.core.blocks.StructBlock( ( ( "page", wagtail.core.blocks.PageChooserBlock(), ), ( "title", wagtail.core.blocks.CharBlock( help_text="Leave blank to use the page's own title", required=False, ), ), ) ), ), ), blank=True, help_text="Single list of elements at the base of the page.", ), ), ( "site", models.OneToOneField( editable=False, on_delete=django.db.models.deletion.CASCADE, to="wagtailcore.Site", ), ), ], options={"abstract": False}, ) ]
true
true
1c48703c6ae531496c99cd091196e12d31f42352
787
py
Python
src/aoc_day03.py
samuelrothen/advent_of_code_2020
22cfc62314d0d19fe91cd1e606064a8828ef0bc7
[ "MIT" ]
1
2021-01-24T21:17:22.000Z
2021-01-24T21:17:22.000Z
src/aoc_day03.py
samuelrothen/advent_of_code_2020
22cfc62314d0d19fe91cd1e606064a8828ef0bc7
[ "MIT" ]
null
null
null
src/aoc_day03.py
samuelrothen/advent_of_code_2020
22cfc62314d0d19fe91cd1e606064a8828ef0bc7
[ "MIT" ]
null
null
null
# Advent of Code Day 3 def check_trees(tree_map, n_steps_r, n_steps_d): x = 0 n_trees = 0 map_border = len(tree_map[0]) for y in range(0, len(tree_map)-1, n_steps_d): x = (x + n_steps_r) % map_border field = tree_map[n_steps_d + y][x] if field == '#': n_trees += 1 return n_trees with open ('../input/day03.txt', 'r') as f: tree_map = f.read().splitlines() # Right 1, down 1 r1_d1 = check_trees(tree_map, 1, 1) # Right 3, down 1 (Part 1) r3_d1 = check_trees(tree_map, 3, 1) # Right 5, down 1 r5_d1 = check_trees(tree_map, 5, 1) # Right 7, down 1 r7_d1 = check_trees(tree_map, 7, 1) # Right 1, down 2 r1_d2 = check_trees(tree_map, 1, 2) print(f'Part 1: {r3_d1}') print(f'Part 2: {r1_d1 * r3_d1 * r5_d1 * r7_d1 * r1_d2}')
23.848485
57
0.614994
def check_trees(tree_map, n_steps_r, n_steps_d): x = 0 n_trees = 0 map_border = len(tree_map[0]) for y in range(0, len(tree_map)-1, n_steps_d): x = (x + n_steps_r) % map_border field = tree_map[n_steps_d + y][x] if field == '#': n_trees += 1 return n_trees with open ('../input/day03.txt', 'r') as f: tree_map = f.read().splitlines() r1_d1 = check_trees(tree_map, 1, 1) r3_d1 = check_trees(tree_map, 3, 1) r5_d1 = check_trees(tree_map, 5, 1) r7_d1 = check_trees(tree_map, 7, 1) r1_d2 = check_trees(tree_map, 1, 2) print(f'Part 1: {r3_d1}') print(f'Part 2: {r1_d1 * r3_d1 * r5_d1 * r7_d1 * r1_d2}')
true
true
1c4870e4e14175611c397b227767f34f313a3296
29,015
py
Python
vnpy/trader/engine.py
crystalphi/vnpy
14eb6ff626b07230830248ee6aea8923a5a887f4
[ "MIT" ]
1
2021-11-24T09:19:57.000Z
2021-11-24T09:19:57.000Z
vnpy/trader/engine.py
crystalphi/vnpy
14eb6ff626b07230830248ee6aea8923a5a887f4
[ "MIT" ]
null
null
null
vnpy/trader/engine.py
crystalphi/vnpy
14eb6ff626b07230830248ee6aea8923a5a887f4
[ "MIT" ]
null
null
null
""" """ import logging from logging import Logger import smtplib import os from abc import ABC from datetime import datetime from email.message import EmailMessage from queue import Empty, Queue from threading import Thread from typing import Any, Sequence, Type, Dict, List, Optional from vnpy.event import Event, EventEngine from .app import BaseApp from .event import ( EVENT_TICK, EVENT_ORDER, EVENT_TRADE, EVENT_POSITION, EVENT_ACCOUNT, EVENT_CONTRACT, EVENT_LOG ) from .gateway import BaseGateway from .object import ( Direction, Exchange, CancelRequest, LogData, OrderRequest, SubscribeRequest, HistoryRequest, OrderData, BarData, TickData, TradeData, PositionData, AccountData, ContractData ) from .setting import SETTINGS from .utility import get_folder_path, TRADER_DIR # 专有的logger文件 from .util_logger import setup_logger class MainEngine: """ Acts as the core of VN Trader. """ def __init__(self, event_engine: EventEngine = None): """""" if event_engine: self.event_engine: EventEngine = event_engine else: self.event_engine = EventEngine() self.event_engine.start() self.gateways: Dict[str, BaseGateway] = {} self.engines: Dict[str, BaseEngine] = {} self.apps: Dict[str, BaseApp] = {} self.exchanges: List[Exchange] = [] self.rm_engine = None self.algo_engine = None self.rpc_service = None os.chdir(TRADER_DIR) # Change working directory self.init_engines() # Initialize function engines def add_engine(self, engine_class: Any) -> "BaseEngine": """ Add function engine. """ engine = engine_class(self, self.event_engine) self.engines[engine.engine_name] = engine return engine def add_gateway(self, gateway_class: Type[BaseGateway], gateway_name: str = None) -> BaseGateway: """ Add gateway. """ if gateway_name: # 使用指定的gateway_name, 可以区分相同接口不同账号的gateway同时接入 gateway = gateway_class(self.event_engine, gateway_name=gateway_name) else: # 缺省使用了接口自己定义的gateway_name gateway = gateway_class(self.event_engine) gateway_name = gateway.gateway_name self.gateways[gateway_name] = gateway # Add gateway supported exchanges into engine for exchange in gateway.exchanges: if exchange not in self.exchanges: self.exchanges.append(exchange) return gateway def add_app(self, app_class: Type[BaseApp]) -> "BaseEngine": """ Add app. """ app = app_class() self.apps[app.app_name] = app engine = self.add_engine(app.engine_class) if app.app_name == "RiskManager": self.rm_engine = engine elif app.app_name == "AlgoTrading": self.algo_engine = engine elif app.app_name == 'RpcService': self.rpc_service = engine return engine def init_engines(self) -> None: """ Init all engines. """ self.add_engine(LogEngine) self.add_engine(OmsEngine) self.add_engine(EmailEngine) def write_log(self, msg: str, source: str = "") -> None: """ Put log event with specific message. """ log = LogData(msg=msg, gateway_name=source) event = Event(EVENT_LOG, log) self.event_engine.put(event) def get_gateway(self, gateway_name: str) -> BaseGateway: """ Return gateway object by name. """ gateway = self.gateways.get(gateway_name, None) if not gateway: self.write_log(f"找不到底层接口:{gateway_name}") return gateway def get_engine(self, engine_name: str) -> "BaseEngine": """ Return engine object by name. """ engine = self.engines.get(engine_name, None) if not engine: self.write_log(f"找不到引擎:{engine_name}") return engine def get_default_setting(self, gateway_name: str) -> Optional[Dict[str, Any]]: """ Get default setting dict of a specific gateway. """ gateway = self.get_gateway(gateway_name) if gateway: return gateway.get_default_setting() return None def get_all_gateway_names(self) -> List[str]: """ Get all names of gatewasy added in main engine. """ return list(self.gateways.keys()) def get_all_gateway_status(self) -> List[dict]: """ Get all gateway status :return: """ return list([{k: v.get_status()} for k, v in self.gateways.items()]) def get_all_apps(self) -> List[BaseApp]: """ Get all app objects. """ return list(self.apps.values()) def get_all_exchanges(self) -> List[Exchange]: """ Get all exchanges. """ return self.exchanges def connect(self, setting: dict, gateway_name: str) -> None: """ Start connection of a specific gateway. """ gateway = self.get_gateway(gateway_name) if gateway: gateway.connect(setting) def subscribe(self, req: SubscribeRequest, gateway_name: str) -> None: """ Subscribe tick data update of a specific gateway. 如果没有指定gateway,那么所有的gateway都会接收改订阅请求 """ if gateway_name: gateway = self.get_gateway(gateway_name) if gateway: gateway.subscribe(req) else: for gateway in self.gateways.values(): if gateway: gateway.subscribe(req) def send_order(self, req: OrderRequest, gateway_name: str) -> str: """ Send new order request to a specific gateway. 扩展支持自定义套利合约。 由cta_strategy_pro发出算法单委托,由算法引擎进行处理 """ # 自定义套利合约,交给算法引擎处理 if self.algo_engine and req.exchange == Exchange.SPD: return self.algo_engine.send_spd_order( req=req, gateway_name=gateway_name) gateway = self.get_gateway(gateway_name) if gateway: return gateway.send_order(req) else: return "" def cancel_order(self, req: CancelRequest, gateway_name: str) -> bool: """ Send cancel order request to a specific gateway. """ # 自定义套利合约,交给算法引擎处理 if self.algo_engine and req.exchange == Exchange.SPD: return self.algo_engine.cancel_spd_order( req=req) gateway = self.get_gateway(gateway_name) if gateway: return gateway.cancel_order(req) return False def send_orders(self, reqs: Sequence[OrderRequest], gateway_name: str) -> List[str]: """ 批量发单 """ gateway = self.get_gateway(gateway_name) if gateway: return gateway.send_orders(reqs) else: return ["" for req in reqs] def cancel_orders(self, reqs: Sequence[CancelRequest], gateway_name: str) -> None: """ """ gateway = self.get_gateway(gateway_name) if gateway: gateway.cancel_orders(reqs) def query_history(self, req: HistoryRequest, gateway_name: str) -> Optional[List[BarData]]: """ Send cancel order request to a specific gateway. """ gateway = self.get_gateway(gateway_name) if gateway: return gateway.query_history(req) else: self.write_log(f'网关为空,请检查合约得网关是否与连接得网关一致') return None def close(self) -> None: """ Make sure every gateway and app is closed properly before programme exit. """ if hasattr(self, 'save_contracts'): self.save_contracts() # Stop event engine first to prevent new timer event. self.event_engine.stop() for engine in self.engines.values(): engine.close() for gateway in self.gateways.values(): gateway.close() class BaseEngine(ABC): """ Abstract class for implementing an function engine. """ def __init__( self, main_engine: MainEngine, event_engine: EventEngine, engine_name: str, ): """""" self.main_engine = main_engine self.event_engine = event_engine self.engine_name = engine_name self.logger = None self.create_logger(engine_name) def create_logger(self, logger_name: str = 'base_engine'): """ 创建engine独有的日志 :param logger_name: 日志名,缺省为engine的名称 :return: """ log_path = get_folder_path("log") log_filename = str(log_path.joinpath(logger_name)) print(u'create logger:{}'.format(log_filename)) self.logger = setup_logger(file_name=log_filename, name=logger_name, log_level=SETTINGS.get('log.level', logging.DEBUG)) def write_log(self, msg: str, source: str = "", level: int = logging.DEBUG): """ 写入日志 :param msg: 日志内容 :param source: 来源 :param level: 日志级别 :return: """ if self.logger: if len(source) > 0: msg = f'[{source}]{msg}' self.logger.log(level, msg) else: log = LogData(msg=msg, level=level, gateway_name='') event = Event(EVENT_LOG, log) self.event_engine.put(event) def close(self): """""" pass class LogEngine(BaseEngine): """ Processes log event and output with logging module. """ def __init__(self, main_engine: MainEngine, event_engine: EventEngine): """""" super(LogEngine, self).__init__(main_engine, event_engine, "log") if not SETTINGS["log.active"]: return self.level: int = SETTINGS["log.level"] self.logger: Logger = logging.getLogger("VN Trader") self.logger.setLevel(self.level) self.formatter = logging.Formatter( "%(asctime)s %(levelname)s: %(message)s" ) self.add_null_handler() if SETTINGS["log.console"]: self.add_console_handler() if SETTINGS["log.file"]: self.add_file_handler() self.register_event() def add_null_handler(self) -> None: """ Add null handler for logger. """ null_handler = logging.NullHandler() self.logger.addHandler(null_handler) def add_console_handler(self) -> None: """ Add console output of log. """ console_handler = logging.StreamHandler() console_handler.setLevel(self.level) console_handler.setFormatter(self.formatter) self.logger.addHandler(console_handler) def add_file_handler(self) -> None: """ Add file output of log. """ today_date = datetime.now().strftime("%Y%m%d") filename = f"vt_{today_date}.log" log_path = get_folder_path("log") file_path = log_path.joinpath(filename) file_handler = logging.FileHandler( file_path, mode="a", encoding="utf8" ) file_handler.setLevel(self.level) file_handler.setFormatter(self.formatter) self.logger.addHandler(file_handler) def register_event(self) -> None: """""" self.event_engine.register(EVENT_LOG, self.process_log_event) def process_log_event(self, event: Event) -> None: """ Process log event. """ log = event.data self.logger.log(log.level, log.msg) class OmsEngine(BaseEngine): """ Provides order management system function for VN Trader. """ def __init__(self, main_engine: MainEngine, event_engine: EventEngine): """""" super(OmsEngine, self).__init__(main_engine, event_engine, "oms") self.ticks: Dict[str, TickData] = {} self.orders: Dict[str, OrderData] = {} self.trades: Dict[str, TradeData] = {} self.positions: Dict[str, PositionData] = {} self.accounts: Dict[str, AccountData] = {} self.contracts: Dict[str, ContractData] = {} self.today_contracts: Dict[str, ContractData] = {} # 自定义合约 self.custom_contracts = {} # vt_symbol: ContractData self.custom_settings = {} # symbol: dict self.symbol_spd_maping = {} # symbol: [spd_symbol] self.prices = {} self.active_orders: Dict[str, OrderData] = {} self.add_function() self.register_event() self.load_contracts() def __del__(self): """保存缓存""" self.save_contracts() def load_contracts(self) -> None: """从本地缓存加载合约字典""" import bz2 import pickle contract_file_name = 'vn_contract.pkb2' if os.path.exists(contract_file_name): try: with bz2.BZ2File(contract_file_name, 'rb') as f: self.contracts = pickle.load(f) self.write_log(f'加载缓存合约字典:{contract_file_name}') except Exception as ex: self.write_log(f'加载缓存合约异常:{str(ex)}') # 更新自定义合约 custom_contracts = self.get_all_custom_contracts() self.get_all_custom_contracts(rtn_setting=True) for contract in custom_contracts.values(): # 更新合约缓存 self.contracts.update({contract.symbol: contract}) self.contracts.update({contract.vt_symbol: contract}) self.today_contracts[contract.vt_symbol] = contract self.today_contracts[contract.symbol] = contract # 获取自定义合约的主动腿/被动腿 setting = self.custom_settings.get(contract.symbol, {}) leg1_symbol = setting.get('leg1_symbol') leg2_symbol = setting.get('leg2_symbol') # 构建映射关系 for symbol in [leg1_symbol, leg2_symbol]: spd_mapping_list = self.symbol_spd_maping.get(symbol, []) # 更新映射 symbol => spd_symbol if (not contract.symbol.endswith('.SPD')) and contract.symbol not in spd_mapping_list: spd_mapping_list.append(contract.symbol) self.symbol_spd_maping.update({symbol: spd_mapping_list}) def save_contracts(self) -> None: """持久化合约对象到缓存文件""" import bz2 import pickle contract_file_name = 'vn_contract.pkb2' with bz2.BZ2File(contract_file_name, 'wb') as f: if len(self.today_contracts) > 0: self.write_log(f'保存今日合约对象到缓存文件') pickle.dump(self.today_contracts, f) else: pickle.dump(self.contracts, f) def add_function(self) -> None: """Add query function to main engine.""" self.main_engine.get_tick = self.get_tick self.main_engine.get_order = self.get_order self.main_engine.get_price = self.get_price self.main_engine.get_trade = self.get_trade self.main_engine.get_position = self.get_position self.main_engine.get_account = self.get_account self.main_engine.get_contract = self.get_contract self.main_engine.get_exchange = self.get_exchange self.main_engine.get_custom_contract = self.get_custom_contract self.main_engine.get_all_ticks = self.get_all_ticks self.main_engine.get_all_orders = self.get_all_orders self.main_engine.get_all_trades = self.get_all_trades self.main_engine.get_all_positions = self.get_all_positions self.main_engine.get_all_accounts = self.get_all_accounts self.main_engine.get_all_contracts = self.get_all_contracts self.main_engine.get_all_active_orders = self.get_all_active_orders self.main_engine.get_all_custom_contracts = self.get_all_custom_contracts self.main_engine.get_mapping_spd = self.get_mapping_spd self.main_engine.save_contracts = self.save_contracts def register_event(self) -> None: """""" self.event_engine.register(EVENT_TICK, self.process_tick_event) self.event_engine.register(EVENT_ORDER, self.process_order_event) self.event_engine.register(EVENT_TRADE, self.process_trade_event) self.event_engine.register(EVENT_POSITION, self.process_position_event) self.event_engine.register(EVENT_ACCOUNT, self.process_account_event) self.event_engine.register(EVENT_CONTRACT, self.process_contract_event) def process_tick_event(self, event: Event) -> None: """""" tick = event.data self.ticks[tick.vt_symbol] = tick if '&' not in tick.symbol and tick.last_price: self.prices[tick.vt_symbol] = tick.last_price def process_order_event(self, event: Event) -> None: """""" order = event.data self.orders[order.vt_orderid] = order # If order is active, then update data in dict. if order.is_active(): self.active_orders[order.vt_orderid] = order # Otherwise, pop inactive order from in dict elif order.vt_orderid in self.active_orders: self.active_orders.pop(order.vt_orderid) def process_trade_event(self, event: Event) -> None: """""" trade = event.data self.trades[trade.vt_tradeid] = trade def process_position_event(self, event: Event) -> None: """""" position = event.data self.positions[position.vt_positionid] = position if position.exchange != Exchange.SPD: self.create_spd_position_event(position.symbol, position.direction) def reverse_direction(self, direction): """返回反向持仓""" if direction == Direction.LONG: return Direction.SHORT elif direction == Direction.SHORT: return Direction.LONG return direction def create_spd_position_event(self, symbol, direction): """创建自定义品种对持仓信息""" spd_symbols = self.symbol_spd_maping.get(symbol, []) if not spd_symbols: return for spd_symbol in spd_symbols: spd_setting = self.custom_settings.get(spd_symbol, None) if not spd_setting: continue leg1_symbol = spd_setting.get('leg1_symbol') leg2_symbol = spd_setting.get('leg2_symbol') leg1_contract = self.contracts.get(leg1_symbol) leg2_contract = self.contracts.get(leg2_symbol) spd_contract = self.contracts.get(spd_symbol) if leg1_contract is None or leg2_contract is None: continue leg1_ratio = spd_setting.get('leg1_ratio', 1) leg2_ratio = spd_setting.get('leg2_ratio', 1) # 找出leg1,leg2的持仓,并判断出spd的方向 spd_pos = None if leg1_symbol == symbol: k1 = f"{leg1_contract.gateway_name}.{leg1_contract.vt_symbol}.{direction.value}" leg1_pos = self.positions.get(k1) k2 = f"{leg2_contract.gateway_name}.{leg2_contract.vt_symbol}.{self.reverse_direction(direction).value}" leg2_pos = self.positions.get(k2) spd_direction = direction k3 = f"{spd_contract.gateway_name}.{spd_symbol}.{Exchange.SPD.value}.{spd_direction.value}" spd_pos = self.positions.get(k3) elif leg2_symbol == symbol: k1 = f"{leg1_contract.gateway_name}.{leg1_contract.vt_symbol}.{self.reverse_direction(direction).value}" leg1_pos = self.positions.get(k1) k2 = f"{leg2_contract.gateway_name}.{leg2_contract.vt_symbol}.{direction.value}" leg2_pos = self.positions.get(k2) spd_direction = self.reverse_direction(direction) k3 = f"{spd_contract.gateway_name}.{spd_symbol}.{Exchange.SPD.value}.{spd_direction.value}" spd_pos = self.positions.get(k3) else: continue if leg1_pos is None or leg2_pos is None: # or leg1_pos.volume ==0 or leg2_pos.volume == 0: continue # 根据leg1/leg2的volume ratio,计算出最小spd_volume spd_volume = min(int(leg1_pos.volume / leg1_ratio), int(leg2_pos.volume / leg2_ratio)) if spd_volume <= 0 and spd_pos is None: continue if spd_setting.get('is_ratio', False) and leg2_pos.price > 0: spd_price = 100 * (leg2_pos.price * leg1_ratio) / (leg2_pos.price * leg2_ratio) elif spd_setting.get('is_spread', False): spd_price = leg1_pos.price * leg1_ratio - leg2_pos.price * leg2_ratio else: spd_price = 0 spd_pos = PositionData( gateway_name=spd_contract.gateway_name, accountid=leg1_pos.accountid, symbol=spd_symbol, exchange=Exchange.SPD, direction=spd_direction, volume=spd_volume, price=spd_price ) event = Event(EVENT_POSITION, data=spd_pos) self.event_engine.put(event) def process_account_event(self, event: Event) -> None: """""" account = event.data self.accounts[account.vt_accountid] = account def process_contract_event(self, event: Event) -> None: """""" contract = event.data self.contracts[contract.vt_symbol] = contract self.contracts[contract.symbol] = contract self.today_contracts[contract.vt_symbol] = contract self.today_contracts[contract.symbol] = contract def get_exchange(self, symbol: str) -> Exchange: """获取合约对应的交易所""" contract = self.contracts.get(symbol, None) if contract is None: return Exchange.LOCAL return contract.exchange def get_tick(self, vt_symbol: str) -> Optional[TickData]: """ Get latest market tick data by vt_symbol. """ return self.ticks.get(vt_symbol, None) def get_price(self, vt_symbol): """ get the lastest price by vt_symbol :param vt_symbol: :return: """ return self.prices.get(vt_symbol, None) def get_order(self, vt_orderid) -> Optional[OrderData]: """ Get latest order data by vt_orderid. """ return self.orders.get(vt_orderid, None) def get_trade(self, vt_tradeid: str) -> Optional[TradeData]: """ Get trade data by vt_tradeid. """ return self.trades.get(vt_tradeid, None) def get_position(self, vt_positionid: str) -> Optional[PositionData]: """ Get latest position data by vt_positionid. """ return self.positions.get(vt_positionid, None) def get_account(self, vt_accountid: str) -> Optional[AccountData]: """ Get latest account data by vt_accountid. """ return self.accounts.get(vt_accountid, None) def get_contract(self, vt_symbol: str) -> Optional[ContractData]: """ Get contract data by vt_symbol. """ return self.contracts.get(vt_symbol, None) def get_all_ticks(self) -> List[TickData]: """ Get all tick data. """ return list(self.ticks.values()) def get_all_orders(self) -> List[OrderData]: """ Get all order data. """ return list(self.orders.values()) def get_all_trades(self) -> List[TradeData]: """ Get all trade data. """ return list(self.trades.values()) def get_all_positions(self) -> List[PositionData]: """ Get all position data. """ return list(self.positions.values()) def get_all_accounts(self) -> List[AccountData]: """ Get all account data. """ return list(self.accounts.values()) def get_all_contracts(self) -> List[ContractData]: """ Get all contract data. """ return list(self.contracts.values()) def get_all_active_orders(self, vt_symbol: str = "") -> List[OrderData]: """ Get all active orders by vt_symbol. If vt_symbol is empty, return all active orders. """ if not vt_symbol: return list(self.active_orders.values()) else: active_orders = [ order for order in self.active_orders.values() if order.vt_symbol == vt_symbol ] return active_orders def get_custom_contract(self, symbol): """ 获取自定义合约的设置 :param symbol: "pb2012-1-pb2101-1-CJ" :return: { "name": "pb跨期价差", "exchange": "SPD", "leg1_symbol": "pb2012", "leg1_exchange": "SHFE", "leg1_ratio": 1, "leg2_symbol": "pb2101", "leg2_exchange": "SHFE", "leg2_ratio": 1, "is_spread": true, "size": 1, "margin_rate": 0.1, "price_tick": 5 } """ return self.custom_settings.get(symbol, None) def get_all_custom_contracts(self, rtn_setting=False): """ 获取所有自定义合约 :return: """ if rtn_setting: if len(self.custom_settings) == 0: c = CustomContract() self.custom_settings = c.get_config() return self.custom_settings if len(self.custom_contracts) == 0: c = CustomContract() self.custom_settings = c.get_config() self.custom_contracts = c.get_contracts() return self.custom_contracts def get_mapping_spd(self, symbol): """根据主动腿/被动腿symbol,获取自定义套利对的symbol list""" return self.symbol_spd_maping.get(symbol, []) class CustomContract(object): """ 定制合约 # 适用于初始化系统时,补充到本地合约信息文件中 contracts.vt # 适用于CTP网关,加载自定义的套利合约,做内部行情撮合 """ # 运行本地目录下,定制合约的配置文件(dict) file_name = 'custom_contracts.json' def __init__(self): """构造函数""" from vnpy.trader.utility import load_json self.setting = load_json(self.file_name) # 所有设置 def get_config(self): """获取配置""" return self.setting def get_contracts(self): """获取所有合约信息""" d = {} from vnpy.trader.object import ContractData, Exchange for symbol, setting in self.setting.items(): gateway_name = setting.get('gateway_name', None) if gateway_name is None: gateway_name = SETTINGS.get('gateway_name', '') vn_exchange = Exchange(setting.get('exchange', 'SPD')) contract = ContractData( gateway_name=gateway_name, symbol=symbol, exchange=vn_exchange, name=setting.get('name', symbol), size=setting.get('size', 100), product=None, pricetick=setting.get('price_tick', 0.01), margin_rate=setting.get('margin_rate', 0.1) ) d[contract.vt_symbol] = contract return d class EmailEngine(BaseEngine): """ Provides email sending function for VN Trader. """ def __init__(self, main_engine: MainEngine, event_engine: EventEngine): """""" super(EmailEngine, self).__init__(main_engine, event_engine, "email") self.thread: Thread = Thread(target=self.run) self.queue: Queue = Queue() self.active: bool = False self.main_engine.send_email = self.send_email def send_email(self, subject: str, content: str, receiver: str = "") -> None: """""" # Start email engine when sending first email. if not self.active: self.start() # Use default receiver if not specified. if not receiver: receiver = SETTINGS["email.receiver"] msg = EmailMessage() msg["From"] = SETTINGS["email.sender"] msg["To"] = receiver msg["Subject"] = subject msg.set_content(content) self.queue.put(msg) def run(self) -> None: """""" while self.active: try: msg = self.queue.get(block=True, timeout=1) with smtplib.SMTP_SSL( SETTINGS["email.server"], SETTINGS["email.port"] ) as smtp: smtp.login( SETTINGS["email.username"], SETTINGS["email.password"] ) smtp.send_message(msg) except Empty: pass def start(self) -> None: """""" self.active = True self.thread.start() def close(self) -> None: """""" if not self.active: return self.active = False self.thread.join()
32.131783
120
0.592418
import logging from logging import Logger import smtplib import os from abc import ABC from datetime import datetime from email.message import EmailMessage from queue import Empty, Queue from threading import Thread from typing import Any, Sequence, Type, Dict, List, Optional from vnpy.event import Event, EventEngine from .app import BaseApp from .event import ( EVENT_TICK, EVENT_ORDER, EVENT_TRADE, EVENT_POSITION, EVENT_ACCOUNT, EVENT_CONTRACT, EVENT_LOG ) from .gateway import BaseGateway from .object import ( Direction, Exchange, CancelRequest, LogData, OrderRequest, SubscribeRequest, HistoryRequest, OrderData, BarData, TickData, TradeData, PositionData, AccountData, ContractData ) from .setting import SETTINGS from .utility import get_folder_path, TRADER_DIR from .util_logger import setup_logger class MainEngine: def __init__(self, event_engine: EventEngine = None): if event_engine: self.event_engine: EventEngine = event_engine else: self.event_engine = EventEngine() self.event_engine.start() self.gateways: Dict[str, BaseGateway] = {} self.engines: Dict[str, BaseEngine] = {} self.apps: Dict[str, BaseApp] = {} self.exchanges: List[Exchange] = [] self.rm_engine = None self.algo_engine = None self.rpc_service = None os.chdir(TRADER_DIR) self.init_engines() def add_engine(self, engine_class: Any) -> "BaseEngine": engine = engine_class(self, self.event_engine) self.engines[engine.engine_name] = engine return engine def add_gateway(self, gateway_class: Type[BaseGateway], gateway_name: str = None) -> BaseGateway: if gateway_name: gateway = gateway_class(self.event_engine, gateway_name=gateway_name) else: gateway = gateway_class(self.event_engine) gateway_name = gateway.gateway_name self.gateways[gateway_name] = gateway for exchange in gateway.exchanges: if exchange not in self.exchanges: self.exchanges.append(exchange) return gateway def add_app(self, app_class: Type[BaseApp]) -> "BaseEngine": app = app_class() self.apps[app.app_name] = app engine = self.add_engine(app.engine_class) if app.app_name == "RiskManager": self.rm_engine = engine elif app.app_name == "AlgoTrading": self.algo_engine = engine elif app.app_name == 'RpcService': self.rpc_service = engine return engine def init_engines(self) -> None: self.add_engine(LogEngine) self.add_engine(OmsEngine) self.add_engine(EmailEngine) def write_log(self, msg: str, source: str = "") -> None: log = LogData(msg=msg, gateway_name=source) event = Event(EVENT_LOG, log) self.event_engine.put(event) def get_gateway(self, gateway_name: str) -> BaseGateway: gateway = self.gateways.get(gateway_name, None) if not gateway: self.write_log(f"找不到底层接口:{gateway_name}") return gateway def get_engine(self, engine_name: str) -> "BaseEngine": engine = self.engines.get(engine_name, None) if not engine: self.write_log(f"找不到引擎:{engine_name}") return engine def get_default_setting(self, gateway_name: str) -> Optional[Dict[str, Any]]: gateway = self.get_gateway(gateway_name) if gateway: return gateway.get_default_setting() return None def get_all_gateway_names(self) -> List[str]: return list(self.gateways.keys()) def get_all_gateway_status(self) -> List[dict]: return list([{k: v.get_status()} for k, v in self.gateways.items()]) def get_all_apps(self) -> List[BaseApp]: return list(self.apps.values()) def get_all_exchanges(self) -> List[Exchange]: return self.exchanges def connect(self, setting: dict, gateway_name: str) -> None: gateway = self.get_gateway(gateway_name) if gateway: gateway.connect(setting) def subscribe(self, req: SubscribeRequest, gateway_name: str) -> None: if gateway_name: gateway = self.get_gateway(gateway_name) if gateway: gateway.subscribe(req) else: for gateway in self.gateways.values(): if gateway: gateway.subscribe(req) def send_order(self, req: OrderRequest, gateway_name: str) -> str: if self.algo_engine and req.exchange == Exchange.SPD: return self.algo_engine.send_spd_order( req=req, gateway_name=gateway_name) gateway = self.get_gateway(gateway_name) if gateway: return gateway.send_order(req) else: return "" def cancel_order(self, req: CancelRequest, gateway_name: str) -> bool: if self.algo_engine and req.exchange == Exchange.SPD: return self.algo_engine.cancel_spd_order( req=req) gateway = self.get_gateway(gateway_name) if gateway: return gateway.cancel_order(req) return False def send_orders(self, reqs: Sequence[OrderRequest], gateway_name: str) -> List[str]: gateway = self.get_gateway(gateway_name) if gateway: return gateway.send_orders(reqs) else: return ["" for req in reqs] def cancel_orders(self, reqs: Sequence[CancelRequest], gateway_name: str) -> None: gateway = self.get_gateway(gateway_name) if gateway: gateway.cancel_orders(reqs) def query_history(self, req: HistoryRequest, gateway_name: str) -> Optional[List[BarData]]: gateway = self.get_gateway(gateway_name) if gateway: return gateway.query_history(req) else: self.write_log(f'网关为空,请检查合约得网关是否与连接得网关一致') return None def close(self) -> None: if hasattr(self, 'save_contracts'): self.save_contracts() self.event_engine.stop() for engine in self.engines.values(): engine.close() for gateway in self.gateways.values(): gateway.close() class BaseEngine(ABC): def __init__( self, main_engine: MainEngine, event_engine: EventEngine, engine_name: str, ): self.main_engine = main_engine self.event_engine = event_engine self.engine_name = engine_name self.logger = None self.create_logger(engine_name) def create_logger(self, logger_name: str = 'base_engine'): log_path = get_folder_path("log") log_filename = str(log_path.joinpath(logger_name)) print(u'create logger:{}'.format(log_filename)) self.logger = setup_logger(file_name=log_filename, name=logger_name, log_level=SETTINGS.get('log.level', logging.DEBUG)) def write_log(self, msg: str, source: str = "", level: int = logging.DEBUG): if self.logger: if len(source) > 0: msg = f'[{source}]{msg}' self.logger.log(level, msg) else: log = LogData(msg=msg, level=level, gateway_name='') event = Event(EVENT_LOG, log) self.event_engine.put(event) def close(self): pass class LogEngine(BaseEngine): def __init__(self, main_engine: MainEngine, event_engine: EventEngine): super(LogEngine, self).__init__(main_engine, event_engine, "log") if not SETTINGS["log.active"]: return self.level: int = SETTINGS["log.level"] self.logger: Logger = logging.getLogger("VN Trader") self.logger.setLevel(self.level) self.formatter = logging.Formatter( "%(asctime)s %(levelname)s: %(message)s" ) self.add_null_handler() if SETTINGS["log.console"]: self.add_console_handler() if SETTINGS["log.file"]: self.add_file_handler() self.register_event() def add_null_handler(self) -> None: null_handler = logging.NullHandler() self.logger.addHandler(null_handler) def add_console_handler(self) -> None: console_handler = logging.StreamHandler() console_handler.setLevel(self.level) console_handler.setFormatter(self.formatter) self.logger.addHandler(console_handler) def add_file_handler(self) -> None: today_date = datetime.now().strftime("%Y%m%d") filename = f"vt_{today_date}.log" log_path = get_folder_path("log") file_path = log_path.joinpath(filename) file_handler = logging.FileHandler( file_path, mode="a", encoding="utf8" ) file_handler.setLevel(self.level) file_handler.setFormatter(self.formatter) self.logger.addHandler(file_handler) def register_event(self) -> None: self.event_engine.register(EVENT_LOG, self.process_log_event) def process_log_event(self, event: Event) -> None: log = event.data self.logger.log(log.level, log.msg) class OmsEngine(BaseEngine): def __init__(self, main_engine: MainEngine, event_engine: EventEngine): super(OmsEngine, self).__init__(main_engine, event_engine, "oms") self.ticks: Dict[str, TickData] = {} self.orders: Dict[str, OrderData] = {} self.trades: Dict[str, TradeData] = {} self.positions: Dict[str, PositionData] = {} self.accounts: Dict[str, AccountData] = {} self.contracts: Dict[str, ContractData] = {} self.today_contracts: Dict[str, ContractData] = {} self.custom_contracts = {} self.custom_settings = {} self.symbol_spd_maping = {} self.prices = {} self.active_orders: Dict[str, OrderData] = {} self.add_function() self.register_event() self.load_contracts() def __del__(self): self.save_contracts() def load_contracts(self) -> None: import bz2 import pickle contract_file_name = 'vn_contract.pkb2' if os.path.exists(contract_file_name): try: with bz2.BZ2File(contract_file_name, 'rb') as f: self.contracts = pickle.load(f) self.write_log(f'加载缓存合约字典:{contract_file_name}') except Exception as ex: self.write_log(f'加载缓存合约异常:{str(ex)}') custom_contracts = self.get_all_custom_contracts() self.get_all_custom_contracts(rtn_setting=True) for contract in custom_contracts.values(): self.contracts.update({contract.symbol: contract}) self.contracts.update({contract.vt_symbol: contract}) self.today_contracts[contract.vt_symbol] = contract self.today_contracts[contract.symbol] = contract setting = self.custom_settings.get(contract.symbol, {}) leg1_symbol = setting.get('leg1_symbol') leg2_symbol = setting.get('leg2_symbol') for symbol in [leg1_symbol, leg2_symbol]: spd_mapping_list = self.symbol_spd_maping.get(symbol, []) if (not contract.symbol.endswith('.SPD')) and contract.symbol not in spd_mapping_list: spd_mapping_list.append(contract.symbol) self.symbol_spd_maping.update({symbol: spd_mapping_list}) def save_contracts(self) -> None: import bz2 import pickle contract_file_name = 'vn_contract.pkb2' with bz2.BZ2File(contract_file_name, 'wb') as f: if len(self.today_contracts) > 0: self.write_log(f'保存今日合约对象到缓存文件') pickle.dump(self.today_contracts, f) else: pickle.dump(self.contracts, f) def add_function(self) -> None: self.main_engine.get_tick = self.get_tick self.main_engine.get_order = self.get_order self.main_engine.get_price = self.get_price self.main_engine.get_trade = self.get_trade self.main_engine.get_position = self.get_position self.main_engine.get_account = self.get_account self.main_engine.get_contract = self.get_contract self.main_engine.get_exchange = self.get_exchange self.main_engine.get_custom_contract = self.get_custom_contract self.main_engine.get_all_ticks = self.get_all_ticks self.main_engine.get_all_orders = self.get_all_orders self.main_engine.get_all_trades = self.get_all_trades self.main_engine.get_all_positions = self.get_all_positions self.main_engine.get_all_accounts = self.get_all_accounts self.main_engine.get_all_contracts = self.get_all_contracts self.main_engine.get_all_active_orders = self.get_all_active_orders self.main_engine.get_all_custom_contracts = self.get_all_custom_contracts self.main_engine.get_mapping_spd = self.get_mapping_spd self.main_engine.save_contracts = self.save_contracts def register_event(self) -> None: self.event_engine.register(EVENT_TICK, self.process_tick_event) self.event_engine.register(EVENT_ORDER, self.process_order_event) self.event_engine.register(EVENT_TRADE, self.process_trade_event) self.event_engine.register(EVENT_POSITION, self.process_position_event) self.event_engine.register(EVENT_ACCOUNT, self.process_account_event) self.event_engine.register(EVENT_CONTRACT, self.process_contract_event) def process_tick_event(self, event: Event) -> None: tick = event.data self.ticks[tick.vt_symbol] = tick if '&' not in tick.symbol and tick.last_price: self.prices[tick.vt_symbol] = tick.last_price def process_order_event(self, event: Event) -> None: order = event.data self.orders[order.vt_orderid] = order if order.is_active(): self.active_orders[order.vt_orderid] = order elif order.vt_orderid in self.active_orders: self.active_orders.pop(order.vt_orderid) def process_trade_event(self, event: Event) -> None: trade = event.data self.trades[trade.vt_tradeid] = trade def process_position_event(self, event: Event) -> None: position = event.data self.positions[position.vt_positionid] = position if position.exchange != Exchange.SPD: self.create_spd_position_event(position.symbol, position.direction) def reverse_direction(self, direction): if direction == Direction.LONG: return Direction.SHORT elif direction == Direction.SHORT: return Direction.LONG return direction def create_spd_position_event(self, symbol, direction): spd_symbols = self.symbol_spd_maping.get(symbol, []) if not spd_symbols: return for spd_symbol in spd_symbols: spd_setting = self.custom_settings.get(spd_symbol, None) if not spd_setting: continue leg1_symbol = spd_setting.get('leg1_symbol') leg2_symbol = spd_setting.get('leg2_symbol') leg1_contract = self.contracts.get(leg1_symbol) leg2_contract = self.contracts.get(leg2_symbol) spd_contract = self.contracts.get(spd_symbol) if leg1_contract is None or leg2_contract is None: continue leg1_ratio = spd_setting.get('leg1_ratio', 1) leg2_ratio = spd_setting.get('leg2_ratio', 1) spd_pos = None if leg1_symbol == symbol: k1 = f"{leg1_contract.gateway_name}.{leg1_contract.vt_symbol}.{direction.value}" leg1_pos = self.positions.get(k1) k2 = f"{leg2_contract.gateway_name}.{leg2_contract.vt_symbol}.{self.reverse_direction(direction).value}" leg2_pos = self.positions.get(k2) spd_direction = direction k3 = f"{spd_contract.gateway_name}.{spd_symbol}.{Exchange.SPD.value}.{spd_direction.value}" spd_pos = self.positions.get(k3) elif leg2_symbol == symbol: k1 = f"{leg1_contract.gateway_name}.{leg1_contract.vt_symbol}.{self.reverse_direction(direction).value}" leg1_pos = self.positions.get(k1) k2 = f"{leg2_contract.gateway_name}.{leg2_contract.vt_symbol}.{direction.value}" leg2_pos = self.positions.get(k2) spd_direction = self.reverse_direction(direction) k3 = f"{spd_contract.gateway_name}.{spd_symbol}.{Exchange.SPD.value}.{spd_direction.value}" spd_pos = self.positions.get(k3) else: continue if leg1_pos is None or leg2_pos is None: continue spd_volume = min(int(leg1_pos.volume / leg1_ratio), int(leg2_pos.volume / leg2_ratio)) if spd_volume <= 0 and spd_pos is None: continue if spd_setting.get('is_ratio', False) and leg2_pos.price > 0: spd_price = 100 * (leg2_pos.price * leg1_ratio) / (leg2_pos.price * leg2_ratio) elif spd_setting.get('is_spread', False): spd_price = leg1_pos.price * leg1_ratio - leg2_pos.price * leg2_ratio else: spd_price = 0 spd_pos = PositionData( gateway_name=spd_contract.gateway_name, accountid=leg1_pos.accountid, symbol=spd_symbol, exchange=Exchange.SPD, direction=spd_direction, volume=spd_volume, price=spd_price ) event = Event(EVENT_POSITION, data=spd_pos) self.event_engine.put(event) def process_account_event(self, event: Event) -> None: account = event.data self.accounts[account.vt_accountid] = account def process_contract_event(self, event: Event) -> None: contract = event.data self.contracts[contract.vt_symbol] = contract self.contracts[contract.symbol] = contract self.today_contracts[contract.vt_symbol] = contract self.today_contracts[contract.symbol] = contract def get_exchange(self, symbol: str) -> Exchange: contract = self.contracts.get(symbol, None) if contract is None: return Exchange.LOCAL return contract.exchange def get_tick(self, vt_symbol: str) -> Optional[TickData]: return self.ticks.get(vt_symbol, None) def get_price(self, vt_symbol): return self.prices.get(vt_symbol, None) def get_order(self, vt_orderid) -> Optional[OrderData]: return self.orders.get(vt_orderid, None) def get_trade(self, vt_tradeid: str) -> Optional[TradeData]: return self.trades.get(vt_tradeid, None) def get_position(self, vt_positionid: str) -> Optional[PositionData]: return self.positions.get(vt_positionid, None) def get_account(self, vt_accountid: str) -> Optional[AccountData]: return self.accounts.get(vt_accountid, None) def get_contract(self, vt_symbol: str) -> Optional[ContractData]: return self.contracts.get(vt_symbol, None) def get_all_ticks(self) -> List[TickData]: return list(self.ticks.values()) def get_all_orders(self) -> List[OrderData]: return list(self.orders.values()) def get_all_trades(self) -> List[TradeData]: return list(self.trades.values()) def get_all_positions(self) -> List[PositionData]: return list(self.positions.values()) def get_all_accounts(self) -> List[AccountData]: return list(self.accounts.values()) def get_all_contracts(self) -> List[ContractData]: return list(self.contracts.values()) def get_all_active_orders(self, vt_symbol: str = "") -> List[OrderData]: if not vt_symbol: return list(self.active_orders.values()) else: active_orders = [ order for order in self.active_orders.values() if order.vt_symbol == vt_symbol ] return active_orders def get_custom_contract(self, symbol): return self.custom_settings.get(symbol, None) def get_all_custom_contracts(self, rtn_setting=False): if rtn_setting: if len(self.custom_settings) == 0: c = CustomContract() self.custom_settings = c.get_config() return self.custom_settings if len(self.custom_contracts) == 0: c = CustomContract() self.custom_settings = c.get_config() self.custom_contracts = c.get_contracts() return self.custom_contracts def get_mapping_spd(self, symbol): return self.symbol_spd_maping.get(symbol, []) class CustomContract(object): file_name = 'custom_contracts.json' def __init__(self): from vnpy.trader.utility import load_json self.setting = load_json(self.file_name) def get_config(self): return self.setting def get_contracts(self): d = {} from vnpy.trader.object import ContractData, Exchange for symbol, setting in self.setting.items(): gateway_name = setting.get('gateway_name', None) if gateway_name is None: gateway_name = SETTINGS.get('gateway_name', '') vn_exchange = Exchange(setting.get('exchange', 'SPD')) contract = ContractData( gateway_name=gateway_name, symbol=symbol, exchange=vn_exchange, name=setting.get('name', symbol), size=setting.get('size', 100), product=None, pricetick=setting.get('price_tick', 0.01), margin_rate=setting.get('margin_rate', 0.1) ) d[contract.vt_symbol] = contract return d class EmailEngine(BaseEngine): def __init__(self, main_engine: MainEngine, event_engine: EventEngine): super(EmailEngine, self).__init__(main_engine, event_engine, "email") self.thread: Thread = Thread(target=self.run) self.queue: Queue = Queue() self.active: bool = False self.main_engine.send_email = self.send_email def send_email(self, subject: str, content: str, receiver: str = "") -> None: if not self.active: self.start() if not receiver: receiver = SETTINGS["email.receiver"] msg = EmailMessage() msg["From"] = SETTINGS["email.sender"] msg["To"] = receiver msg["Subject"] = subject msg.set_content(content) self.queue.put(msg) def run(self) -> None: while self.active: try: msg = self.queue.get(block=True, timeout=1) with smtplib.SMTP_SSL( SETTINGS["email.server"], SETTINGS["email.port"] ) as smtp: smtp.login( SETTINGS["email.username"], SETTINGS["email.password"] ) smtp.send_message(msg) except Empty: pass def start(self) -> None: self.active = True self.thread.start() def close(self) -> None: if not self.active: return self.active = False self.thread.join()
true
true
1c4871f0c878ab1c35feefbb79fe50299bde90c4
729
py
Python
apex/functions/notify/line_notify.py
Koichi-Kobayashi/aws_iot_line_notify
a745db492d01107b2b8675f6febe332bbb4307af
[ "Apache-2.0" ]
null
null
null
apex/functions/notify/line_notify.py
Koichi-Kobayashi/aws_iot_line_notify
a745db492d01107b2b8675f6febe332bbb4307af
[ "Apache-2.0" ]
null
null
null
apex/functions/notify/line_notify.py
Koichi-Kobayashi/aws_iot_line_notify
a745db492d01107b2b8675f6febe332bbb4307af
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- import requests import conf def create_message(button_event): if button_event['clickType'] == 'SINGLE': msg = conf.clickType_single elif button_event['clickType'] == 'DOUBLE': msg = conf.clickType_double else: msg = conf.clickType_long return msg def lambda_handler(event, context): url = "https://notify-api.line.me/api/notify" token = conf.token headers = {"Authorization": "Bearer " + token} button_event = event['deviceEvent']['buttonClicked'] payload = {"message": create_message(button_event)} r = requests.post(url, headers=headers, params=payload) if __name__ == '__main__': event = '' lambda_handler(event, None)
22.090909
59
0.657064
import requests import conf def create_message(button_event): if button_event['clickType'] == 'SINGLE': msg = conf.clickType_single elif button_event['clickType'] == 'DOUBLE': msg = conf.clickType_double else: msg = conf.clickType_long return msg def lambda_handler(event, context): url = "https://notify-api.line.me/api/notify" token = conf.token headers = {"Authorization": "Bearer " + token} button_event = event['deviceEvent']['buttonClicked'] payload = {"message": create_message(button_event)} r = requests.post(url, headers=headers, params=payload) if __name__ == '__main__': event = '' lambda_handler(event, None)
true
true
1c48739fad3abf032b3b1f31759a4c33c5bb2c20
398
py
Python
Chapter-2/Example_2-2.py
vishwesh5/Learning-OpenCV-Book
59ad4e190274f92735075013e9276b8159d6ca2e
[ "MIT" ]
2
2018-12-21T17:36:53.000Z
2018-12-22T17:13:00.000Z
Chapter-2/Example_2-2.py
vishwesh5/Learning-OpenCV-Book
59ad4e190274f92735075013e9276b8159d6ca2e
[ "MIT" ]
null
null
null
Chapter-2/Example_2-2.py
vishwesh5/Learning-OpenCV-Book
59ad4e190274f92735075013e9276b8159d6ca2e
[ "MIT" ]
null
null
null
import cv2 as cv import argparse parser = argparse.ArgumentParser() parser.add_argument("-i","--image",required=True,\ help="Path to image") args = vars(parser.parse_args()) img = cv.imread(args["image"]) if img is None: print("Image not found") else: cv.namedWindow("Example2",cv.WINDOW_AUTOSIZE) cv.imshow("Example2",img) cv.waitKey(0) cv.destroyWindow("Example2")
22.111111
50
0.68593
import cv2 as cv import argparse parser = argparse.ArgumentParser() parser.add_argument("-i","--image",required=True,\ help="Path to image") args = vars(parser.parse_args()) img = cv.imread(args["image"]) if img is None: print("Image not found") else: cv.namedWindow("Example2",cv.WINDOW_AUTOSIZE) cv.imshow("Example2",img) cv.waitKey(0) cv.destroyWindow("Example2")
true
true