text
string
size
int64
token_count
int64
import mysql.connector as mydb conn = mydb.connect( host='mysql_container', port='3306', user='docker', password='docker', database='my_db' ) conn.is_connected()
184
66
from django.db import models class Exam(models.Model): """Exam Model""" class Meta(object): verbose_name = 'Іспит' verbose_name_plural = 'Іспити' title = models.CharField( max_length=256, blank=False, verbose_name='Назва предмету') datetime = models.DateTimeField( blank=False, verbose_name='Дата і час проведення') teacher = models.CharField( max_length=256, blank=False, verbose_name='ПІБ викладача') exam_group = models.ForeignKey( 'Group', verbose_name='Група', blank=False, null=True, on_delete=models.PROTECT) def __str__(self): return '%s (приймає %s' % (self.title, self.teacher)
752
253
from __future__ import unicode_literals import frappe from frappe import _ from datetime import datetime, timedelta, date, time from frappe.utils import getdate, get_time, flt, now_datetime from frappe.utils import escape_html from frappe import throw, msgprint, _ @frappe.whitelist() def get_schedule(): time_schedule = frappe.get_doc('Club Settings') schedule = [] for time in time_schedule.club_tour_schedule: from_time_string = str(time.from_time) from_time_datetime = datetime.strptime(from_time_string, "%H:%M:%S") from_time = datetime.strftime(from_time_datetime, "%I:%M %p") to_time_string = str(time.to_time) to_time_datetime = datetime.strptime(to_time_string, "%H:%M:%S") to_time = datetime.strftime(to_time_datetime, "%I:%M %p") name = _('{0} - {1}').format(from_time, to_time) schedule.append({ "name" : name }) frappe.response["message"] = { "Preferred Time": schedule } @frappe.whitelist() def get_status(client_id): client = frappe.db.get("Client", {"email": frappe.session.user}) doc= frappe.get_all('Club Tour', filters={'client_id':client.name,'tour_status': "Pending"}, fields=["*"]) if doc: frappe.response["message"] = { "Status": 0, "Status Message": "Pending" } else: doc= frappe.get_all('Club Tour', filters={'client_id':client.name,'tour_status': "Scheduled"}, fields=["*"]) if doc: doc_1= doc[0] frappe.response["message"] = { "Status":1, "Status Message": "Scheduled", "From Time": doc_1.start_time, "To Time": doc_1.end_time } @frappe.whitelist() def create_clubtour(client_id,date,time): client = frappe.db.get("Client", {"email": frappe.session.user}) doc = frappe.get_doc({ 'doctype': 'Club Tour', 'client_id': client.name, 'preferred_date': date, 'preferred_time_between': time }) doc.save() frappe.response["message"] = { "Status":1, "Status Message": "Club Tour booking submitted" }
2,201
724
''' goalboost.model package The goalboost model package consists of MongoEngine models along with Marshmallow schemas. MongoEngine is our database ORM to MongoDB, and Marshmallow is a serialization library that helps us validate, consume, and expose these Orm objects for clients that need it at the API layer. For MongoEngine, see http://mongoengine.org/ For Marshmallow and the MongoEngine integration piece, see: https://marshmallow.readthedocs.org/en/latest/ https://github.com/touilleMan/marshmallow-mongoengine ''' from flask.ext.mongoengine import MongoEngine db = MongoEngine() def init_db(app): global db db.init_app(app)
647
191
# Copyright 2018 Huawei Technologies Co.,Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); you may not use # this file except in compliance with the License. You may obtain a copy of the # License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software distributed # under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR # CONDITIONS OF ANY KIND, either express or implied. See the License for the # specific language governing permissions and limitations under the License. import logging def snapshots_detail(conn): query = { 'limit': 10 } details = list(conn.block_store.snapshots(**query)) logging.info(details) def create_snapshot(conn): attr = { 'name': 'snap-001', 'description': 'Daily backup', 'volume_id': '5aa119a8-d25b-45a7-8d1b-88e127885635', 'force': False, 'metadata': {} } snapshot = conn.block_store.create_snapshot(**attr) logging.info(snapshot) def rollback_snapshot(conn): snapshot_id = 'snapshot-id' volume_id = 'volume-id' volume_name = 'volume-name' snapshot_rollback = conn.block_store.rollback_snapshot(volume_id, volume_name, snapshot_id) logging.info(snapshot_rollback) def update_snapshot(conn): snapshot_id = 'snapshot-id' attrs = { 'name': 'name_xx3', 'description': 'hello' } snapshot = conn.block_store.update_snapshot(snapshot_id, **attrs) logging.info(snapshot) def create_snapshot_metadata(conn): snapshot_id = 'snapshot-id' metadata = { 'metadata': { 'key1': 'value1', 'key2': 'value2' } } new_metadata = conn.block_store.create_snapshot_metadata(snapshot_id, **metadata) logging.info(new_metadata) def update_snapshot_metadata(conn): snapshot_id = 'snapshot-id' metadata = { 'metadata': { 'key1': 'value1', 'key2': 'value2' } } updated_metadata = conn.block_store.update_snapshot_metadata(snapshot_id, **metadata) logging.info(updated_metadata) def update_snapshot_metadata_with_key(conn): snapshot_id = 'snapshot-id' key = 'key1' metadata = { 'meta': { 'key1': 'value1', } } updated_metadata = conn.block_store.update_snapshot_metadata(snapshot_id, key=key, **metadata) logging.info(updated_metadata) def delete_snapshot_metadata(conn): snapshot_id = 'snapshot-id' key = 'key1' conn.block_store.delete_snapshot_metadata(snapshot_id, key) def get_snapshot_metadata(conn): snapshot_id = 'snapshot-id' metadata = conn.block_store.get_snapshot_metadata(snapshot_id) logging.info(metadata) def get_snapshot_metadata_with_key(conn): key = 'key1' snapshot_id = 'snapshot-id' metadata = conn.block_store.get_snapshot_metadata(snapshot_id, key) logging.info(metadata)
3,290
1,041
import boto3 from zipfile import ZipFile import argparse import json import os import shutil class LambdaMaker(object): def __init__(self, config_file, working_dir): # const vars self.creator='TomLambdaCreator_v1.0.0' os.chdir(working_dir) self.process_config_file(config_file) def process_config_file(self, fname): # read config file with open(fname, 'r') as f: self.contents = json.load(f) f.close() self.lambda_bucket = self.contents['S3Bucket'] self.key = self.contents['S3Key'] self.fname = self.contents['ZipLocalFname'] self.basename = self.contents['ZipBaseName'] self.buildDir = self.contents['BuildDir'] self.functionName=self.contents['FunctionName'] self.runTime=self.contents['Runtime'] self.iamRole=self.contents['Role'] self.handler=self.contents['Handler'] self.desc=self.contents['Description'] self.timeout=self.contents['Timeout'] self.memory=self.contents['MemorySize'] self.publish=self.contents['Publish'] self.vpnconfig = {} self.vpnconfig['SubnetIds'] = self.contents['SubnetIds'] self.vpnconfig['SecurityGroupIds'] = self.contents['SecurityGroupIds'] self.targetArn = self.contents['DeadLetterTargetArn'] self.env = self.contents['EnvironmentVariables'] self.tracingConfig = self.contents['TracingConfigMode'] self.keyarn = self.contents['KeyArn'] def install_python_dependancies(self): deps = self.contents['dependancies'] for dep in deps: cmd = (("pip install {0} -t .").format(dep)) os.system(cmd) def install_node_dependancies(self): deps = self.contents['dependancies'] deplen = len(deps) if deplen > 0: os.mkdir("node_modules") for dep in deps: cmd = (("npm install -s {0}").format(dep)) print(cmd) os.system(cmd) def make_zip_file(self): if (os.path.exists(self.buildDir)): # remove old build director shutil.rmtree(self.buildDir) # make the build dir os.mkdir(self.buildDir) #copy the source file source = self.contents['sourceFile'] shutil.copy(source, self.buildDir) source_files = [] source_files.append(source) os.chdir(self.buildDir) if 'node' in self.runTime: self.install_node_dependancies() else: self.install_python_dependancies() shutil.make_archive(self.basename, "zip") #with ZipFile(self.fname, 'w') as myzip: # for zipit in source_files: # print(("adding {0} to {1}").format(zipit, self.fname)) # myzip.write(zipit) def push_function_code_to_s3(self): self.make_zip_file() client = boto3.client('s3') response = client.put_object( Bucket=self.lambda_bucket, Body=open(self.fname, 'rb'), Key=self.key) metadata=response['ResponseMetadata'] print(("s3 code metadata = {0}").format(metadata)) self.s3version = response['VersionId'] print(('version = {0}').format(self.s3version)) # now that we pushed the code we can setup the S3 # info. self.setup_function_vars() print("pushed code to s3") def setup_function_vars(self): self.code = {} self.code['S3Bucket'] = self.lambda_bucket self.code['S3Key'] = self.key self.code['S3ObjectVersion'] = self.s3version self.desc="Get Location" self.deadcfg = {} self.deadcfg['TargetArn'] = self.targetArn self.variables = {} self.variables['Variables'] = self.env self.tracingMode = {} self.tracingMode['Mode'] = self.tracingConfig # Active needs special permissions #self.tracingConfig['Mode'] = 'Active' self.tags = {} self.tags['FunctionName'] = self.functionName self.tags['RunTime'] = self.runTime self.tags['Creator'] = self.creator def make_new_function(self): response = self.lambda_client.create_function( FunctionName=self.functionName, Runtime=self.runTime, Role=self.iamRole, Handler=self.handler, Code=self.code, Description=self.desc, Timeout=self.timeout, MemorySize=self.memory, Publish=self.publish, VpcConfig=self.vpnconfig, DeadLetterConfig=self.deadcfg, Environment=self.variables, #KMSKeyArn=self.keyarn, TracingConfig=self.tracingMode, Tags=self.tags ) print(("lambda create response = {0}").format(response)) def update_function_code(self): response = self.lambda_client.update_function_code( FunctionName=self.functionName, S3Bucket=self.lambda_bucket, S3Key=self.key, S3ObjectVersion=self.s3version, Publish=True, DryRun=False) print(("update_function_code response: {0}").format(response)) def push_code(self): self.lambda_client = boto3.client('lambda') newFunction = False try: response = self.lambda_client.get_function( FunctionName=self.functionName) #print(response['ResponseMetadata']) except Exception: newFunction = True # push the new code to S3 self.push_function_code_to_s3() if newFunction: # new function so make it self.make_new_function(); else: # function exists so just update code self.update_function_code() def main(): parser = argparse.ArgumentParser(description='aws lambda function creator') parser.add_argument('-f', required=True, help='json file') parser.add_argument('-w', required=True, help='working directory') args = parser.parse_args() config_file = args.f wdir = args.w LambdaMaker(config_file).push_code() if __name__ == "__main__": main()
6,270
1,879
# Lint as: python3 # Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Common modules for callbacks.""" from __future__ import absolute_import from __future__ import division # from __future__ import google_type_annotations from __future__ import print_function import functools import os from absl import logging import numpy as np import tensorflow as tf from typing import Any, List, Optional, MutableMapping from official.utils.misc import keras_utils from official.vision.image_classification.pruning.pruning_base_configs import ModelPruningConfig from tensorflow_model_optimization.python.core.keras import compat from tensorflow_model_optimization.python.core.sparsity.keras.cprune_registry import ConstraintRegistry def get_callbacks(model_checkpoint: bool = True, include_tensorboard: bool = True, time_history: bool = True, track_lr: bool = True, model_pruning_config: Optional[ModelPruningConfig] = None, write_model_weights: bool = True, batch_size: int = 0, log_steps: int = 0, model_dir: str = None) -> List[tf.keras.callbacks.Callback]: """Get all callbacks.""" model_dir = model_dir or '' callbacks = [] if model_checkpoint: ckpt_full_path = os.path.join(model_dir, 'model.ckpt-{epoch:04d}') callbacks.append( tf.keras.callbacks.ModelCheckpoint( ckpt_full_path, save_weights_only=True, verbose=1)) if include_tensorboard: callbacks.append( CustomTensorBoard( log_dir=model_dir, track_lr=track_lr, model_pruning_config=model_pruning_config, write_images=write_model_weights)) if time_history: callbacks.append( keras_utils.TimeHistory( batch_size, log_steps, logdir=model_dir if include_tensorboard else None)) return callbacks def get_scalar_from_tensor(t: tf.Tensor) -> int: """Utility function to convert a Tensor to a scalar.""" t = tf.keras.backend.get_value(t) if callable(t): return t() else: return t class CustomTensorBoard(tf.keras.callbacks.TensorBoard): """A customized TensorBoard callback that tracks additional datapoints. Metrics tracked: - Global learning rate Attributes: log_dir: the path of the directory where to save the log files to be parsed by TensorBoard. track_lr: `bool`, whether or not to track the global learning rate. **kwargs: Additional arguments for backwards compatibility. Possible key is `period`. """ # TODO(b/146499062): track params, flops, log lr, l2 loss, # classification loss def __init__(self, log_dir: str, track_lr: bool = False, model_pruning_config: Optional[ModelPruningConfig] = None, **kwargs): super(CustomTensorBoard, self).__init__(log_dir=log_dir, **kwargs) self._track_lr = track_lr self._model_pruning_config = model_pruning_config def _collect_learning_rate(self, logs): logs = logs or {} lr_schedule = getattr(self.model.optimizer, "lr", None) if isinstance(lr_schedule, tf.keras.optimizers.schedules.LearningRateSchedule): logs["learning_rate"] = tf.keras.backend.get_value( lr_schedule(self.model.optimizer.iterations) ) if isinstance(logs["learning_rate"], functools.partial): logs["learning_rate"] = logs["learning_rate"]() return logs def _log_metrics(self, logs, prefix, step): if self._track_lr: super()._log_metrics(self._collect_learning_rate(logs), prefix, step) def _log_pruning_metrics(self, logs, prefix, step): if compat.is_v1_apis(): # Safely depend on TF 1.X private API given # no more 1.X releases. self._write_custom_summaries(step, logs) else: # TF 2.X log_dir = self.log_dir + '/metrics' file_writer = tf.summary.create_file_writer(log_dir) file_writer.set_as_default() for name, value in logs.items(): tf.summary.scalar(name, value, step=step) file_writer.flush() def on_epoch_begin(self, epoch, logs=None): if logs is not None: super(CustomTensorBoard, self).on_epoch_begin(epoch, logs) if self._model_pruning_config: pruning_logs = {} params = [] postfixes = [] for layer_pruning_config in self._model_pruning_config.pruning: layer_name = layer_pruning_config.layer_name layer = self.model.get_layer(layer_name) for weight_pruning_config in layer_pruning_config.pruning: weight_name = weight_pruning_config.weight_name constraint_name = ConstraintRegistry.get_constraint_from_weight(weight_name) constraint = getattr(layer, constraint_name) params.append(constraint.mask) params.append(constraint.threshold) postfixes.append('/' + layer_name + '/' + weight_name) params.append(self.model.optimizer.iterations) values = tf.keras.backend.batch_get_value(params) iteration = values[-1] del values[-1] del params[-1] param_value_pairs = list(zip(params, values)) for (mask, mask_value), postfix in zip(param_value_pairs[::2], postfixes): pruning_logs.update({ 'mask_sparsity' + postfix: 1 - np.mean(mask_value) }) for (threshold, threshold_value), postfix in zip(param_value_pairs[1::2], postfixes): pruning_logs.update({'threshold' + postfix: threshold_value}) self._log_pruning_metrics(pruning_logs, '', iteration)
6,260
1,949
"""Initialize the index package""" # flake8: noqa from __future__ import absolute_import from .base import * from .typ import *
129
41
# -*- coding: utf-8 -*- class Saucer(object): """ Representa un plato de comida. """ def __init__(self, cadNombre, realPrecio, cadDescription=None, cadImagen=None, boolVegetariano=False, entCoccion=1): self.nombre = cadNombre self.precio = realPrecio self.descripcion = cadDescription self.imagen = cadImagen self.esVegetariano = boolVegetariano self.coccion = entCoccion def __str__(self): return "{nombre}{esVeg}: {precio:.2f}{desc}".format( nombre=self.nombre, desc=' (' + self.descripcion + ')' if self.descripcion else '', precio=self.precio, esVeg='*' if self.esVegetariano else '') burgerPython = Saucer("Hamburguesa de Python", 0.13, cadDescription="Barely an eigth of a byte") print(burgerPython)
885
296
import torch import torch.nn.functional as F from torch import nn from resnext.resnext101_regular import ResNeXt101 class _AttentionModule(nn.Module): def __init__(self): super(_AttentionModule, self).__init__() self.block1 = nn.Sequential( nn.Conv2d(64, 64, 1, bias=False), nn.BatchNorm2d(64), nn.ReLU(), nn.Conv2d(64, 64, 3, dilation=2, padding=2, groups=32, bias=False), nn.BatchNorm2d(64), nn.ReLU(), nn.Conv2d(64, 64, 1, bias=False), nn.BatchNorm2d(64) ) self.block2 = nn.Sequential( nn.Conv2d(64, 64, 1, bias=False), nn.BatchNorm2d(64), nn.ReLU(), nn.Conv2d(64, 64, 3, dilation=3, padding=3, groups=32, bias=False), nn.BatchNorm2d(64), nn.ReLU(), nn.Conv2d(64, 64, 1, bias=False), nn.BatchNorm2d(64) ) self.block3 = nn.Sequential( nn.Conv2d(64, 64, 1, bias=False), nn.BatchNorm2d(64), nn.ReLU(), nn.Conv2d(64, 64, 3, dilation=4, padding=4, groups=32, bias=False), nn.BatchNorm2d(64), nn.ReLU(), nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.down = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32) ) def forward(self, x): block1 = F.relu(self.block1(x) + x, True) block2 = F.relu(self.block2(block1) + block1, True) block3 = F.sigmoid(self.block3(block2) + self.down(block2)) return block3 class BDRAR(nn.Module): def __init__(self): super(BDRAR, self).__init__() resnext = ResNeXt101() self.layer0 = resnext.layer0 self.layer1 = resnext.layer1 self.layer2 = resnext.layer2 self.layer3 = resnext.layer3 self.layer4 = resnext.layer4 self.down4 = nn.Sequential( nn.Conv2d(2048, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU() ) self.down3 = nn.Sequential( nn.Conv2d(1024, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU() ) self.down2 = nn.Sequential( nn.Conv2d(512, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU() ) self.down1 = nn.Sequential( nn.Conv2d(256, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU() ) self.refine3_hl = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 3, padding=1, groups=32, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.refine2_hl = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 3, padding=1, groups=32, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.refine1_hl = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 3, padding=1, groups=32, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.attention3_hl = _AttentionModule() self.attention2_hl = _AttentionModule() self.attention1_hl = _AttentionModule() self.refine2_lh = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 3, padding=1, groups=32, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.refine4_lh = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 3, padding=1, groups=32, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.refine3_lh = nn.Sequential( nn.Conv2d(64, 32, 1, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 3, padding=1, groups=32, bias=False), nn.BatchNorm2d(32), nn.ReLU(), nn.Conv2d(32, 32, 1, bias=False), nn.BatchNorm2d(32) ) self.attention2_lh = _AttentionModule() self.attention3_lh = _AttentionModule() self.attention4_lh = _AttentionModule() self.fuse_attention = nn.Sequential( nn.Conv2d(64, 16, 3, padding=1, bias=False), nn.BatchNorm2d(16), nn.ReLU(), nn.Conv2d(16, 2, 1) ) self.predict = nn.Sequential( nn.Conv2d(32, 8, 3, padding=1, bias=False), nn.BatchNorm2d(8), nn.ReLU(), nn.Dropout(0.1), nn.Conv2d(8, 1, 1) ) # for m in self.modules(): # if isinstance(m, nn.ReLU) or isinstance(m, nn.Dropout): # m.inplace = True for m in self.modules(): if isinstance(m, nn.ReLU): m.inplace = True def forward(self, x): layer0 = self.layer0(x) layer1 = self.layer1(layer0) layer2 = self.layer2(layer1) layer3 = self.layer3(layer2) layer4 = self.layer4(layer3) down4 = self.down4(layer4) down3 = self.down3(layer3) down2 = self.down2(layer2) down1 = self.down1(layer1) down4 = F.upsample(down4, size=down3.size()[2:], mode='bilinear') refine3_hl_0 = F.relu(self.refine3_hl(torch.cat((down4, down3), 1)) + down4, True) refine3_hl_0 = (1 + self.attention3_hl(torch.cat((down4, down3), 1))) * refine3_hl_0 refine3_hl_1 = F.relu(self.refine3_hl(torch.cat((refine3_hl_0, down3), 1)) + refine3_hl_0, True) refine3_hl_1 = (1 + self.attention3_hl(torch.cat((refine3_hl_0, down3), 1))) * refine3_hl_1 refine3_hl_1 = F.upsample(refine3_hl_1, size=down2.size()[2:], mode='bilinear') refine2_hl_0 = F.relu(self.refine2_hl(torch.cat((refine3_hl_1, down2), 1)) + refine3_hl_1, True) refine2_hl_0 = (1 + self.attention2_hl(torch.cat((refine3_hl_1, down2), 1))) * refine2_hl_0 refine2_hl_1 = F.relu(self.refine2_hl(torch.cat((refine2_hl_0, down2), 1)) + refine2_hl_0, True) refine2_hl_1 = (1 + self.attention2_hl(torch.cat((refine2_hl_0, down2), 1))) * refine2_hl_1 refine2_hl_1 = F.upsample(refine2_hl_1, size=down1.size()[2:], mode='bilinear') refine1_hl_0 = F.relu(self.refine1_hl(torch.cat((refine2_hl_1, down1), 1)) + refine2_hl_1, True) refine1_hl_0 = (1 + self.attention1_hl(torch.cat((refine2_hl_1, down1), 1))) * refine1_hl_0 refine1_hl_1 = F.relu(self.refine1_hl(torch.cat((refine1_hl_0, down1), 1)) + refine1_hl_0, True) refine1_hl_1 = (1 + self.attention1_hl(torch.cat((refine1_hl_0, down1), 1))) * refine1_hl_1 down2 = F.upsample(down2, size=down1.size()[2:], mode='bilinear') refine2_lh_0 = F.relu(self.refine2_lh(torch.cat((down1, down2), 1)) + down1, True) refine2_lh_0 = (1 + self.attention2_lh(torch.cat((down1, down2), 1))) * refine2_lh_0 refine2_lh_1 = F.relu(self.refine2_lh(torch.cat((refine2_lh_0, down2), 1)) + refine2_lh_0, True) refine2_lh_1 = (1 + self.attention2_lh(torch.cat((refine2_lh_0, down2), 1))) * refine2_lh_1 down3 = F.upsample(down3, size=down1.size()[2:], mode='bilinear') refine3_lh_0 = F.relu(self.refine3_lh(torch.cat((refine2_lh_1, down3), 1)) + refine2_lh_1, True) refine3_lh_0 = (1 + self.attention3_lh(torch.cat((refine2_lh_1, down3), 1))) * refine3_lh_0 refine3_lh_1 = F.relu(self.refine3_lh(torch.cat((refine3_lh_0, down3), 1)) + refine3_lh_0, True) refine3_lh_1 = (1 + self.attention3_lh(torch.cat((refine3_lh_0, down3), 1))) * refine3_lh_1 down4 = F.upsample(down4, size=down1.size()[2:], mode='bilinear') refine4_lh_0 = F.relu(self.refine4_lh(torch.cat((refine3_lh_1, down4), 1)) + refine3_lh_1, True) refine4_lh_0 = (1 + self.attention4_lh(torch.cat((refine3_lh_1, down4), 1))) * refine4_lh_0 refine4_lh_1 = F.relu(self.refine4_lh(torch.cat((refine4_lh_0, down4), 1)) + refine4_lh_0, True) refine4_lh_1 = (1 + self.attention4_lh(torch.cat((refine4_lh_0, down4), 1))) * refine4_lh_1 refine3_hl_1 = F.upsample(refine3_hl_1, size=down1.size()[2:], mode='bilinear') predict4_hl = self.predict(down4) predict3_hl = self.predict(refine3_hl_1) predict2_hl = self.predict(refine2_hl_1) predict1_hl = self.predict(refine1_hl_1) predict1_lh = self.predict(down1) predict2_lh = self.predict(refine2_lh_1) predict3_lh = self.predict(refine3_lh_1) predict4_lh = self.predict(refine4_lh_1) fuse_attention = F.sigmoid(self.fuse_attention(torch.cat((refine1_hl_1, refine4_lh_1), 1))) fuse_predict = torch.sum(fuse_attention * torch.cat((predict1_hl, predict4_lh), 1), 1, True) predict4_hl = F.upsample(predict4_hl, size=x.size()[2:], mode='bilinear') predict3_hl = F.upsample(predict3_hl, size=x.size()[2:], mode='bilinear') predict2_hl = F.upsample(predict2_hl, size=x.size()[2:], mode='bilinear') predict1_hl = F.upsample(predict1_hl, size=x.size()[2:], mode='bilinear') predict1_lh = F.upsample(predict1_lh, size=x.size()[2:], mode='bilinear') predict2_lh = F.upsample(predict2_lh, size=x.size()[2:], mode='bilinear') predict3_lh = F.upsample(predict3_lh, size=x.size()[2:], mode='bilinear') predict4_lh = F.upsample(predict4_lh, size=x.size()[2:], mode='bilinear') fuse_predict = F.upsample(fuse_predict, size=x.size()[2:], mode='bilinear') if self.training: return fuse_predict, predict1_hl, predict2_hl, predict3_hl, predict4_hl, predict1_lh, predict2_lh, predict3_lh, predict4_lh return F.sigmoid(fuse_predict)
9,768
4,500
'Classes and functions for creating fonts and text buttons' #IMPORTS import pygame from pathlib import Path from .core import clip_surface, load_image from .colour import palette_swap #IMPORTS #VARIALBES __all__ = ['create_font', 'TextButton'] path = Path(__file__).resolve().parent #VARIABLES #CREATE_FONT def create_font(colour): '''A function to create small and large Font objects colour will be the colour of the text The first value in the returned tuple is the small font and the second value is the large font Returns: tuple''' if tuple(colour[:3]) == (0, 0, 0): small_font_image = palette_swap(load_image(path.joinpath('font/small.png')), {(255, 0, 0): colour[:3], tuple(colour[:3]): (255, 255, 255)}) large_font_image = palette_swap(load_image(path.joinpath('font/large.png')), {(255, 0, 0): colour[:3], tuple(colour[:3]): (255, 255, 255)}) return Font(small_font_image, background_colour=255), Font(large_font_image, background_colour=255) if tuple(colour[:3]) == (127, 127, 127): small_font_image = palette_swap(load_image(path.joinpath('font/small.png')), {(255, 0, 0): colour[:3], tuple(colour[:3]): (128, 128, 128)}) large_font_image = palette_swap(load_image(path.joinpath('font/large.png')), {(255, 0, 0): colour[:3], tuple(colour[:3]): (128, 128, 128)}) return Font(small_font_image, 128), Font(large_font_image, 128) small_font_image = palette_swap(load_image(path.joinpath('font/small.png')), {(255, 0, 0): colour[:3]}) large_font_image = palette_swap(load_image(path.joinpath('font/large.png')), {(255, 0, 0): colour[:3]}) return Font(small_font_image), Font(large_font_image) #CREATE_FONT #FONT class Font: '''A class to create a pixel art font It will get all the letters out of the image and render them The border between letters is usually (127, 127, 127) and the background is usually (0, 0, 0) change them if it is necessary The font is made by DaFluffyPotato Attributes: character_height characters font_image space_width''' #__INIT__ def __init__(self, font_image, border_colour=127, background_colour=0): 'Initialising a font object' self.font_image = font_image self.font_image.set_colorkey((0, 0, 0) if not background_colour else [background_colour for i in range(3)]) self.characters = {} current_width, character_count = 0, 0 character_order = ['A','B','C','D','E','F','G','H','I','J','K','L','M','N','O','P','Q','R','S','T','U','V','W','X','Y','Z','a','b','c','d','e','f','g','h','i','j','k','l','m','n','o','p','q','r','s','t','u','v','w','x','y','z','.','-',',',':','+','\'','!','?','0','1','2','3','4','5','6','7','8','9','(',')','/','_','=','\\','[',']','*','"','<','>',';'] for x in range(self.font_image.get_width()): colour = self.font_image.get_at((x, 0)) if colour[:3] == (border_colour, border_colour, border_colour): #IF THE TEXT COLOR IS (127, 127, 127), CHANGE BORDER_COLOR character_image = clip_surface(self.font_image, (x - current_width, 0), (current_width, self.font_image.get_height())) #CLIP EVERY CHARACTER OUT OF THE FONT IMAGE self.characters[character_order[character_count]] = character_image character_count += 1 current_width = 0 else: current_width += 1 self.space_width, self.character_height = self.characters['A'].get_size() #__INIT__ #__REPR__ def __repr__(self): '''Returns a string representation of the object Returns: str''' return 'pgeng.Font' #__REPR__ #GET_SIZE def get_size(self, text): '''Get the size that that a rendered string would use It will return the width and height Returns: tuple''' if type(text) is not str: raise TypeError('text has to be a string') width, height = 0, 0 for character in text: if character not in ('\n', ' ') and character in self.characters: width += self.characters[character].get_width() + 1 #+ 1 FOR SPACING elif character == ' ' or character not in ['\n']: width += self.space_width + 1 #+ 1 FOR SPACING else: width = 0 height += self.character_height + 1 #+ 1 FOR SPACING return width, height #GET_SIZE #RENDER def render(self, surface, text, location, scroll=pygame.Vector2()): 'Render a string on a surface at a location' if type(text) is not str: raise TypeError('text has to be a string') x_offset, y_offset = 0, 0 for character in text: if character not in ('\n', ' ') and character in self.characters: surface.blit(self.characters[character], (location[0] + x_offset - scroll[0], location[1] + y_offset - scroll[1])) x_offset += self.characters[character].get_width() + 1 #+ 1 FOR SPACING elif character == ' ' or character not in ['\n']: x_offset += self.space_width + 1 #+ 1 FOR SPACING else: x_offset = 0 y_offset += self.character_height + 1 #+ 1 FOR SPACING #RENDER #FONT #TEXTBUTTON class TextButton: '''A string of text that is also a button The collide function is to collide with the mouse and clicks It also needs a font size, it has to be either 'small' or 'large' Use the location variable instead of the rect values Attributes: location rect size test_font text''' #__INIT__ def __init__(self, text, location, font_size): 'Initialising a TextButton object' if font_size != 'small' and font_size != 'large': raise ValueError('font_size is not \'small\' or \'large\'') if type(text) is not str: raise TypeError('text is not a string') self.text = text self.location = pygame.Vector2(location) self.test_font = Font(load_image(path.joinpath(f'font/{font_size}.png'))) self.size = self.test_font.get_size(text) #__INIT__ #__REPR__ def __repr__(self): '''Returns a string representation of the object Returns: str''' return f'pgeng.TextButton({tuple(self.location)})' #__REPR__ #RECT @property def rect(self): '''Returns the pygame.Rect object of the TextButton Returns: pygame.Rect''' self.location = pygame.Vector2(self.location) return pygame.Rect(self.location, (self.size[0] - 1, self.size[1] + self.test_font.character_height)) #- 1 FOR THE EXTRA SPACING #RECT #SET_TEXT def set_text(self, text): '''Sets a new string as the text All the variables will be updated, so the functions can be used the same''' if type(text) is not str: raise TypeError('text is not a string') self.text = text self.size = self.test_font.get_size(text) #SET_TEXT #COLLIDE def collide(self, click, check_location=None): '''This will check collision with the mouse location and also if click is True with it A custom location can be set with location if pygame.mouse.get_pos() is not wished to be used The first value returns True if the mouse has collided with the button, the second one is if the mouse clicked on it Returns: tuple''' check_location = pygame.mouse.get_pos() if check_location is None else check_location if self.rect.collidepoint(check_location): if click: return True, True return True, False return False, False #COLLIDE #RENDER def render(self, surface, font, scroll=pygame.Vector2()): 'Renders the text from the button' if not isinstance(font, Font): raise TypeError('font is not a Font object') font.render(surface, self.text, self.location, scroll) #RENDER #TEXTBUTTON
7,413
2,958
""" This modules contain helper function to deal with PIL / Pillow Images. .. note:: Please note that the ``[PIL]`` (pillow) extra dependency must be installed to allow functions from this module to work. """ from . import guetzli def _to_pil_rgb_image(image): """Returns an PIL Image converted to the RGB color space. If the image has an alpha channel (transparency), it will be overlaid on a black background. :param image: the PIL image to convert :returns: The input image if it was already in RGB mode, or a new RGB image if converted. :raises ImportError: PIL / Pillow cannot be imported. """ if image.mode == "RGB": return image from PIL import Image image.load() rgb_image = Image.new("RGB", image.size, (0x00, 0x00, 0x00)) mask = None if image.mode == "RGBA": mask = image.split()[3] # bands: R=0, G=1, B=2, 1=3 rgb_image.paste(image, mask=mask) return rgb_image def process_pil_image(image, quality=guetzli.DEFAULT_JPEG_QUALITY): """Generates an optimized JPEG from a PIL image. If the image has an alpha channel (transparency), it will be overlaid on a black background. :param image: the PIL image :param quality: the output JPEG quality (default 95) :returns: Optimized JPEG bytes :rtype: bytes :raises ImportError: PIL / Pillow cannot be imported. .. code:: python import pyguetzli from PIL import Image image = Image.open("./test/image.jpg") optimized_jpeg = pyguetzli.process_pil_image(image) """ image_rgb = _to_pil_rgb_image(image) image_rgb_bytes = image_rgb.tobytes() return guetzli.process_rgb_bytes( image_rgb_bytes, *image.size, quality=quality )
1,814
598
# # @lc app=leetcode id=201 lang=python # # [201] Bitwise AND of Numbers Range # # https://leetcode.com/problems/bitwise-and-of-numbers-range/description/ # # algorithms # Medium (35.44%) # Total Accepted: 77.3K # Total Submissions: 217.7K # Testcase Example: '5\n7' # # Given a range [m, n] where 0 <= m <= n <= 2147483647, return the bitwise AND # of all numbers in this range, inclusive. # # Example 1: # # # Input: [5,7] # Output: 4 # # # Example 2: # # # Input: [0,1] # Output: 0 # class Solution(object): def rangeBitwiseAnd(self, m, n): """ :type m: int :type n: int :rtype: int """ # Extraordinary Solution max_num = 2 ** 31 - 1 while (max_num & m != max_num & n): max_num = max_num << 1 print max_num return max_num & m def _rangeBitwiseAnd(self, m, n): """ :type m: int :type n: int :rtype: int """ # memory error return reduce(lambda x,y: x & y, range(m, n+1))
1,034
409
''' dip @author Bardia Mojra - 1000766739 @brief ee-5323 - project - @date 10/31/21 code based on below YouTube tutorial and Pymotw.com documentation for socket mod. @link https://www.youtube.com/watch?v=3QiPPX-KeSc @link https://pymotw.com/2/socket/tcp.html python socket module documentation @link https://docs.python.org/3/library/socket.html @link https://docs.python.org/3/howto/sockets.html ''' import csv import math import numpy as np import os import pygame import pyglet from pyglet.window import key import pymunk import pymunk.constraints import pymunk.pygame_util import pandas as pd import pyglet.gl as gl ''' custom libs ''' import dvm import tcm ''' NBUG ''' from nbug import * ''' TEST CONFIG ''' TEST_ID = 'Test 903' SIM_DUR = 30.0 # in seconds OUT_DIR = '../out/' OUT_DATA = OUT_DIR+TEST_ID+'_data.csv' CONF_DIR = '../config/' # cart m_c = 0.5 all_friction = 0.2 ''' pendulum 1 ''' l_1 = 0.4 # 6, 5, 4, 7 -- 4 -> m_1 = 0.2 # 2, 3, 4 -- 1 -> stable m_1_moment = 0.01 m_1_radius = 0.05 ''' pendulum 2 ''' l_2 = 0.7 # 6, 5, 7 -- 3 -> unstable m_2 = 0.3 # 2, 3, 4 -- 2 -> unstable m_2_moment = 0.001 m_2_radius = 0.05 # other config output_labels=['t', 'x', 'dx', 'th_1', 'dth_1', 'th_2', 'dth_2'] # control config # K gain matrix and Nbar found from modelling via Jupyter # K = [16.91887353, 21.12423935, 137.96378003, -3.20040325, -259.72220049, -50.48383455] # Nbar = 17.0 K = [51.43763708, 54.12690472, 157.5467596, -21.67111679, -429.11603909, -88.73125241] Nbar = 51.5 tConfig = tcm.test_configuration(TEST_ID=TEST_ID, OUT_DIR=OUT_DIR, OUT_DATA=OUT_DATA, CONF_DIR=CONF_DIR, SIM_DUR=SIM_DUR, output_labels=output_labels, all_friction=all_friction, cart_mass=m_c, pend_1_length=l_1, pend_1_mass=m_1, pend_1_moment=m_1_moment, pend_2_length=l_2, pend_2_mass=m_2, pend_2_moment=m_2_moment, K=K, Nbar=Nbar) # log test config tcm.pkl(tConfig) ''' MOD CONFIG ''' SCREEN_WIDTH = 700 SCREEN_HEIGHT = 500 # sim config MAX_FORCE = 25 DT = 1 / 60.0 PPM = 200.0 # pxls per meter END_ = 1000 # samples used for plotting and analysis SHOW_ = True cart_size = 0.3, 0.2 white_color = (0,0,0,0) black_color = (255,255,255,255) green_color = (0,135,0,255) red_color = (135,0,0,255) blue_color = (0,0,135,255) ''' main ''' pygame.init() # screen = pygame.display.set_mode((SCREEN_WIDTH, SCREEN_HEIGHT)) surface = pygame.Surface((SCREEN_WIDTH, SCREEN_HEIGHT)) # clock = pygame.time.Clock() window = pyglet.window.Window(SCREEN_WIDTH, SCREEN_HEIGHT, vsync=False, caption='Double Inverted Pendulum Simulation') gl.glClearColor(255,255,255,255) # setup the space space = pymunk.Space() # options = pymunk.pygame_util.DrawOptions(surface) # space.debug_draw(options) space.gravity = 0, -9.81 # space.debug_draw(options) fil = pymunk.ShapeFilter(group=1) # screen.fill(pygame.Color("white")) # options = pymunk.pygame_util.DrawOptions(screen) # space.debug_draw(options) # ground ground = pymunk.Segment(space.static_body, (-4, -0.1), (4, -0.1), 0.1) # ground.color = pygame.Color("pink") ground.friction = all_friction ground.filter = fil space.add(ground) # space.debug_draw(options) # cart cart_moment = pymunk.moment_for_box(m_c, cart_size) cart_body = pymunk.Body(mass=m_c, moment=cart_moment) cart_body.position = 0.0, cart_size[1] / 2 cart_shape = pymunk.Poly.create_box(cart_body, cart_size) cart_shape.color = black_color # cart_shape.color = red_color # cart_shape.fill_color = red_color # cart_shape.color = black_color cart_shape.friction = ground.friction space.add(cart_body, cart_shape) # space.debug_draw(options) # pendulum 1 pend_1_body = pymunk.Body(mass=m_1, moment=m_1_moment) pend_1_body.position = cart_body.position[0], cart_body.position[1] + cart_size[1] / 2 + l_1 pend_shape = pymunk.Circle(pend_1_body, m_1_radius) pend_shape.filter = fil space.add(pend_1_body, pend_shape) # joint joint = pymunk.constraints.PivotJoint(cart_body, pend_1_body, cart_body.position + (0, cart_size[1] / 2)) joint.collide_bodies = False space.add(joint) # pendulum 2 pend_2_body = pymunk.Body(mass=m_2, moment=m_2_moment) pend_2_body.position = cart_body.position[0], cart_body.position[1] + cart_size[1] / 2 + (2 * l_2) pend_shape2 = pymunk.Circle(pend_2_body, m_2_radius) pend_shape2.filter = fil space.add(pend_2_body, pend_shape2) # joint 2 joint2 = pymunk.constraints.PivotJoint(pend_1_body, pend_2_body, cart_body.position + (0, cart_size[1] / 2 + l_2)) joint2.collide_bodies = False space.add(joint2) # space.debug_draw(options) print(f"cart mass = {cart_body.mass:0.1f} kg") print(f"pendulum 1 mass = {pend_1_body.mass:0.1f} kg, pendulum moment = {pend_1_body.moment:0.3f} kg*m^2") print(f"pendulum 2 mass = {pend_2_body.mass:0.1f} kg, pendulum moment = {pend_2_body.moment:0.3f} kg*m^2") force = 0.0 ref = 0.0 color = (200, 200, 200, 200) label_x = pyglet.text.Label(text='', font_size=12, color=color, x=10, y=SCREEN_HEIGHT - 28) label_th_1 = pyglet.text.Label(text='', font_size=12, color=color, x=10, y=SCREEN_HEIGHT - 58) label_th_2 = pyglet.text.Label(text='', font_size=12, color=color, x=10, y=SCREEN_HEIGHT - 88) label_force = pyglet.text.Label(text='', font_size=12, color=color, x=10, y=SCREEN_HEIGHT - 118) labels = [label_x, label_th_1, label_th_2, label_force] # data recorder so we can compare our results to our predictions if os.path.exists(OUT_DATA): os.remove(OUT_DATA) with open(OUT_DATA, 'w') as f: output_header = str() for i, s in enumerate(output_labels): if i == 0: output_header = s else: output_header += ', '+s output_header += '\n' f.write(output_header) f.close() currtime = 0.0 record_data = True def draw_body(offset, body): for shape in body.shapes: if isinstance(shape, pymunk.Circle): vertices = [] num_points = 10 for ii in range(num_points): angle = ii / num_points * 2 * math.pi vertices.append(body.position + (shape.radius * math.cos(angle), shape.radius * math.sin(angle))) points = [] for v in vertices: points.append(int(v[0] * PPM) + offset[0]) points.append(int(v[1] * PPM) + offset[1]) data = ('v2i', tuple(points)) gl.glColor3b(255,255,255) pyglet.graphics.draw(len(vertices), pyglet.gl.GL_LINE_LOOP, data) elif isinstance(shape, pymunk.Poly): # get vertices in world coordinates vertices = [v.rotated(body.angle) + body.position for v in shape.get_vertices()] # convert vertices to pixel coordinates points = [] for v in vertices: points.append(int(v[0] * PPM) + offset[0]) points.append(int(v[1] * PPM) + offset[1]) data = ('v2i', tuple(points)) gl.glColor3b(255,255,255) pyglet.graphics.draw(len(vertices), pyglet.gl.GL_LINE_LOOP, data) def draw_line_between(offset, pos1, pos2): vertices = [pos1, pos2] points = [] for v in vertices: points.append(int(v[0] * PPM) + offset[0]) points.append(int(v[1] * PPM) + offset[1]) data = ('v2i', tuple(points)) gl.glColor3b(255,255,255) pyglet.graphics.draw(len(vertices), pyglet.gl.GL_LINE_STRIP, data) def draw_ground(offset): vertices = [v + (0, ground.radius) for v in (ground.a, ground.b)] # convert vertices to pixel coordinates points = [] for v in vertices: points.append(int(v[0] * PPM) + offset[0]) points.append(int(v[1] * PPM) + offset[1]) data = ('v2i', tuple(points)) pyglet.graphics.draw(len(vertices), pyglet.gl.GL_LINES, data) @window.event def on_draw(): window.clear() # center view x around 0 offset = (250, 5) draw_body(offset, cart_body) draw_body(offset, pend_1_body) draw_line_between(offset, cart_body.position + (0, cart_size[1] / 2), pend_1_body.position) draw_body(offset, pend_2_body) draw_line_between(offset, pend_1_body.position, pend_2_body.position) draw_ground(offset) for label in labels: label.draw() @window.event def on_key_press(symbol, modifiers): # Symbolic names: if symbol == key.ESCAPE: window.close() def simulate(_): global currtime if currtime > SIM_DUR: window.close() # nprint('_',_) # ensure we get a consistent simulation step - ignore the input dt value dt = DT # simulate the world # NOTE: using substeps will mess up gains space.step(dt) # populate the current state posx = cart_body.position[0] velx = cart_body.velocity[0] th_1 = pend_1_body.angle th_1v = pend_1_body.angular_velocity th_2 = pend_2_body.angle th_2v = pend_2_body.angular_velocity # dump our data so we can plot if record_data: with open(OUT_DATA, 'a+') as f: f.write(f"{currtime:0.5f}, {posx:0.5f}, {velx:0.5f}, {th_1:0.5f}, {th_1v:0.5f}, {th_2:0.5f}, {th_2v:0.5f} \n") f.close() currtime += dt # calculate our gain based on the current state gain = K[0] * posx + K[1] * velx + K[2] * th_1 + K[3] * th_1v + K[4] * th_2 + K[5] * th_2v # calculate the force required global force force = ref * Nbar - gain # kill our motors if our angles get out of control if math.fabs(pend_1_body.angle) > 1.0 or math.fabs(pend_2_body.angle) > 1.0: force = 0.0 # cap our maximum force so it doesn't go crazy if math.fabs(force) > MAX_FORCE: force = math.copysign(MAX_FORCE, force) # apply force to cart center of mass cart_body.apply_force_at_local_point((force, 0.0), (0, 0)) def update_state_label(_): ''' function to store the current state to draw on screen ''' label_x.text = f'x: {cart_body.position[0]:0.3f} m' label_th_1.text = f'theta_1: {pend_1_body.angle:0.3f} rad' label_th_2.text = f'theta_2: {pend_2_body.angle:0.3f} rad' label_force.text = f'force: {force:0.1f} N' def update_reference(_, newref): global ref ref = newref # callback for simulation pyglet.clock.schedule_interval(simulate, DT) pyglet.clock.schedule_interval(update_state_label, 0.25) # schedule some small movements by updating our reference pyglet.clock.schedule_once(update_reference, 2, 0.2) pyglet.clock.schedule_once(update_reference, 7, 0.6) pyglet.clock.schedule_once(update_reference, 12, 0.2) pyglet.clock.schedule_once(update_reference, 17, 0.0) pyglet.app.run() f.close() # data recorder so we can compare our results to our predictions # f = open(OUT_DATA, 'r') # ['t', 'x', 'dx', 'th_1', 'dth_1', 'th_2', 'dth_2', 'L1', 'L2'] # for i in test_IDs: tConfig = tcm.unpkl(TEST_ID, CONF_DIR) df = pd.read_csv(tConfig.out_data) df = dvm.get_losses(df, dataPath=tConfig.data_path, lossPath=tConfig.loss_path) # plot pose # ['t', 'x', 'dx', 'th_1', 'dth_1', 'th_2', 'dth_2', 'L1', 'L2'] cols = [0, 1, 3, 5] xy_df = df.iloc[:,cols].copy() dvm.plot_df(xy_df, plot_title='State Position', labels=xy_df.columns, test_id=tConfig.id, out_dir=tConfig.out_dir, end=END_, show=SHOW_) # plot vel # ['t', 'x', 'dx', 'th_1', 'dth_1', 'th_2', 'dth_2', 'L1', 'L2'] cols = [0, 2, 4, 6] xy_df = df.iloc[:,cols].copy() dvm.plot_df(xy_df, plot_title='State Velocity', labels=xy_df.columns, test_id=tConfig.id, out_dir=tConfig.out_dir, end=END_, show=SHOW_) # plot losses # ['t', 'x', 'dx', 'th_1', 'dth_1', 'th_2', 'dth_2', 'L1', 'L2'] cols = [0, 7, 8] xy_df = df.iloc[:,cols].copy() dvm.plot_df(xy_df, plot_title='State Losses', labels=xy_df.columns, test_id=tConfig.id, out_dir=tConfig.out_dir, end=END_, show=SHOW_) # print losses dvm.print_losses(df)
12,053
5,044
""" A Command parser to parse over each jinja template for a given cmake command """ import os from apigen.Logger import Logger from jinja2 import Environment, PackageLoader, FileSystemLoader class CommandParser(object): def __init__(self, cmdfile: str, env: Environment, outdir: str): super().__init__() self.__log = Logger.getlogger() self.CommandFilePath = cmdfile self.__env = env self.OutputDirectory = outdir def ParseFile(self): cmd_basefilename = os.path.basename(self.CommandFilePath) self.__log.info("Parsing File: " + cmd_basefilename) cmd_name = os.path.splitext(cmd_basefilename)[0] cmd_outfile = os.path.join(self.OutputDirectory, cmd_basefilename) #if (cmd_basefilename != "add_executable.py"): # return # Render the command output from the template template = self.__env.get_template(cmd_basefilename) tmpl_output = template.render(CmdName=cmd_name) # Save the File with open(cmd_outfile, "w") as text_file: text_file.write(tmpl_output) return
1,128
334
#!/usr/bin/env python import sys print("Please tell me your favorite color: ") color = sys.stdin.readline() animal = raw_input("Please tell me your favorite animal: ") print(animal) sys.stdout.write("Your favorite color is: " + color + " favorite animal is: " + animal + "\n") print("*" * 10)
299
99
# Comet VOEvent Broker. # Event handlers. from comet.handler.relay import * from comet.handler.spawn import *
111
39
from rest_framework import status from rest_framework.reverse import reverse from rest_framework.test import APITestCase from measure_mate.models import Template from measure_mate.tests.factories import TemplateFactory class TemplateAPITestCases(APITestCase): def test_list_template(self): """ List all templates and check that all fields are returned """ template = TemplateFactory() template2 = TemplateFactory() url = reverse('template-list') response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) self.assertEqual(response.data[0]['id'], template.id) self.assertEqual(response.data[0]['name'], template.name) self.assertEqual(response.data[0]['taggable'], template.taggable) self.assertEqual(response.data[0]['enabled'], template.enabled) self.assertEqual(len(response.data), Template.objects.count())
954
255
# Generated by Django 4.0 on 2021-12-29 11:36 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('shop_management', '0002_alter_shop_type_alter_shoptype_name'), ] operations = [ migrations.AddField( model_name='shop', name='created_at', field=models.DateTimeField(auto_now_add=True, null=True, verbose_name='تاریخ ایجاد'), ), migrations.AddField( model_name='shop', name='last_updated', field=models.DateTimeField(auto_now=True, null=True, verbose_name='آخرین بروزرسانی'), ), ]
658
228
import torch import torch.nn as nn from lib.util import normal_log_density class ModelActor(nn.Module): def __init__(self, obs_size, act_size, active='tanh', hidden_size=128, lstd=-0.0): super(ModelActor, self).__init__() if active == 'tanh': self.active = torch.tanh else: pass self.linear1 = nn.Linear(obs_size, hidden_size) self.linear2 = nn.Linear(hidden_size, hidden_size) self.linear3 = nn.Linear(hidden_size, act_size) self.linear3.weight.data.mul_(0.1) self.linear3.bias.data.mul_(0.0) self.logstd = nn.Parameter(torch.ones(1, act_size) * lstd) def forward(self, x): x = self.linear1(x) x = self.active(x) x = self.linear2(x) x = self.active(x) mean = self.linear3(x) logstd = self.logstd.expand_as(mean) std = torch.exp(logstd) return mean, logstd, std def select_action(self, x): mean, _, std = self.forward(x) #action = mean + std * torch.normal(mean=torch.zeros_like(mean), std=torch.ones_like(std)) action = torch.normal(mean, std) #print(mean) #print(std) #print(action) #print(torch.normal(mean, std)) #torch.normal(mean, std) return action, mean, std def get_log_prob(self, x, action): mean, logstd, std = self.forward(x) return normal_log_density(action, mean, logstd, std) def get_kl(self, x, model): mean, logstd, std = self.forward(x) #mean_, logstd_, std_ = model(x) mean_, logstd_, std_ = mean.detach(), logstd.detach(), std.detach() kl = logstd - logstd_ + (std_ ** 2 + (mean_ - mean) ** 2) / (2.0 * std ** 2) - 0.5 return kl.sum(1, keepdim=True) def get_ent(self, x ): mean, logstd, std = self.forward(x) return (logstd + 0.5 * np.log(2.0*np.pi*np.e)).sum(-1) class ModelCritic(nn.Module): def __init__(self, obs_size, hidden_size = 128): super(ModelCritic, self).__init__() self.active = torch.tanh self.linear1 = nn.Linear(obs_size, hidden_size) self.linear2 = nn.Linear(hidden_size, hidden_size) self.linear3 = nn.Linear(hidden_size, 1) self.linear3.weight.data.mul_(0.1) self.linear3.bias.data.mul_(0.0) def forward(self, x): x = self.linear1(x) x = self.active(x) x - self.linear2(x) x - self.active(x) value = self.linear3(x) return value class ModelDCritic(nn.Module): def __init__(self, obs_size, act_size, hidden_size=128, ): super(ModelDCritic, self).__init__() self.active = torch.tanh self.linear1 = nn.Linear(obs_size+act_size, hidden_size) self.linear2 = nn.Linear(hidden_size, hidden_size) self.linear3 = nn.Linear(hidden_size, 1) self.linear3.weight.data.mul_(0.1) self.linear3.bias.data.mul_(0.0) def forward(self, x): x = self.linear1(x) x = self.active(x) x = self.linear2(x) x = self.active(x) value = torch.sigmoid(self.linear3(x)) return value class DDPGActor(nn.Module): def __init__(self, obs_size, act_size, active='tanh', hidden_size=128): super(ModelActor, self).__init__() self.active = torch.tanh self.linear1 = nn.Linear(obs_size, hidden_size) self.linear2 = nn.Linear(hidden_size, hidden_size) self.linear3 = nn.Linear(hidden_size, act_size) self.linear3.weight.data.mul_(0.1) self.linear3.bias.data.mul_(0.0) def forward(self, x): x = self.linear1(x) x = self.active(x) x = self.linear2(x) x = self.active(x) x = self.linear3(x) return self.active(x) def select_action(self, x): x = self.forward(x) def get_log_prob(self, ): pass def get_kl(self, ): pass def get_ent(self,): pass class DDPGCritic(nn.Module): def __init__(self, ): pass def forward(self, ): pass
4,085
1,506
import importlib import unittest solution = importlib.import_module('2020_04_2') class Test2020Day4Part1(unittest.TestCase): def test_example1(self): input = ( 'eyr:1972 cid:100\n' 'hcl:#18171d ecl:amb hgt:170 pid:186cm iyr:2018 byr:1926\n' '\n' 'iyr:2019\n' 'hcl:#602927 eyr:1967 hgt:170cm\n' 'ecl:grn pid:012533040 byr:1946\n' '\n' 'hcl:dab227 iyr:2012\n' 'ecl:brn hgt:182cm pid:021572410 eyr:2020 byr:1992 cid:277\n' '\n' 'hgt:59cm ecl:zzz\n' 'eyr:2038 hcl:74454a iyr:2023\n' 'pid:3556412378 byr:2007\n' '\n' 'pid:087499704 hgt:74in ecl:grn iyr:2012 eyr:2030 byr:1980\n' 'hcl:#623a2f\n' '\n' 'eyr:2029 ecl:blu cid:129 byr:1989\n' 'iyr:2014 pid:896056539 hcl:#a97842 hgt:165cm\n' '\n' 'hcl:#888785\n' 'hgt:164cm byr:2001 iyr:2015 cid:88\n' 'pid:545766238 ecl:hzl\n' 'eyr:2022\n' '\n' 'iyr:2010 hgt:158cm hcl:#b6652a ecl:blu byr:1944 eyr:2021 pid:093154719\n' ) self.assertEqual(solution.run(input), 4)
1,129
674
# Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Implementations of a AIL agent.""" from acme.agents.jax.ail import losses from acme.agents.jax.ail import rewards from acme.agents.jax.ail.agents import AIL from acme.agents.jax.ail.agents import DistributedAIL from acme.agents.jax.ail.builder import AILBuilder from acme.agents.jax.ail.config import AILConfig from acme.agents.jax.ail.dac_agents import DAC from acme.agents.jax.ail.dac_agents import DACConfig from acme.agents.jax.ail.dac_agents import DistributedDAC from acme.agents.jax.ail.gail_agents import DistributedGAIL from acme.agents.jax.ail.gail_agents import GAIL from acme.agents.jax.ail.gail_agents import GAILConfig from acme.agents.jax.ail.learning import AILLearner from acme.agents.jax.ail.networks import AILNetworks from acme.agents.jax.ail.networks import AIRLModule from acme.agents.jax.ail.networks import compute_ail_reward from acme.agents.jax.ail.networks import DiscriminatorMLP from acme.agents.jax.ail.networks import DiscriminatorModule from acme.agents.jax.ail.networks import make_discriminator
1,649
553
from __future__ import print_function from cineapp import app, db, lm from flask_login import login_required from flask import jsonify, session, g, url_for, request from pywebpush import webpush, WebPushException from cineapp.models import PushNotification import json, traceback, sys, datetime, time from cineapp.auth import guest_control @app.route('/notifications/subscribe', methods=['POST']) @login_required @guest_control def notification_subscribe(): app.logger.info('New user subscription !!') subscription = request.get_json() app.logger.info('User id: %s, Subscription data: %s' % (g.user.id,subscription)) # Let's register the subscription message into the database push_notification = PushNotification(endpoint_id=subscription["endpoint"], public_key=subscription["keys"]["p256dh"], auth_token=subscription["keys"]["auth"], session_id=session.sid, user_id=g.user.id) # Store the subscription data into database try: db.session.add(push_notification) db.session.commit() app.logger.info('User subscription correctly stored into database') except Exception as e: app.logger.error('Unable to store subscription user in database %s', repr(e)) return jsonify({ "status": "failure", "message": u"Unable to store subscription object into database" }) return jsonify({ "status": "success", "message": u"Endpoint enregistray" }) def notification_send(text,active_subscriptions): for cur_active_sub in active_subscriptions: try: expiration_date = time.mktime((datetime.datetime.now() + datetime.timedelta(hours=12)).timetuple()) webpush(cur_active_sub.serialize(), data=json.dumps({ "url":url_for('chat'), "message_title": "Message depuis le chat", "message": text }) , vapid_private_key=app.config["NOTIF_PRIVATE_KEY_PATH"], vapid_claims={ "sub": "mailto:ptitoliv@gmail.com", "exp": expiration_date } ) except WebPushException as ex: # If there is an error let's remove the subscription app.logger.error("Subscription for endpoint %s is incorrect ==> Delete it", cur_active_sub) print(traceback.print_exc(file=sys.stdout)); # Let's remove the notification notification_unsubscribe(cur_active_sub) print(("I'm sorry, Dave, but I can't do that: {}", repr(ex))) print(ex.response.json()) def notification_unsubscribe(sub): try: db.session.delete(sub) db.session.commit() app.logger.info('User subscription correctly delete from database') return True except Exception as e: app.logger.error('Unable to remove subscription user in database %s', repr(e)) return False
2,578
849
# generated by datamodel-codegen: # filename: openapi.yaml # timestamp: 2021-12-31T02:50:50+00:00 from __future__ import annotations from datetime import datetime from enum import Enum from typing import Annotated, Any, List, Optional from pydantic import BaseModel, Extra, Field class ResourceNotFoundException(BaseModel): __root__: Any class InvalidRequestException(ResourceNotFoundException): pass class InternalFailureException(ResourceNotFoundException): pass class ServiceUnavailableException(ResourceNotFoundException): pass class ThrottlingException(ResourceNotFoundException): pass class CancelPipelineReprocessingResponse(BaseModel): pass class ServiceManagedChannelS3Storage(CancelPipelineReprocessingResponse): """ Used to store channel data in an S3 bucket managed by IoT Analytics. You can't change the choice of S3 storage after the data store is created. """ pass class UnlimitedRetentionPeriod(BaseModel): __root__: bool class RetentionPeriodInDays(BaseModel): __root__: Annotated[int, Field(ge=1.0)] class ResourceAlreadyExistsException(ResourceNotFoundException): pass class LimitExceededException(ResourceNotFoundException): pass class UnlimitedVersioning(UnlimitedRetentionPeriod): pass class MaxVersions(BaseModel): __root__: Annotated[int, Field(ge=1.0, le=1000.0)] class ServiceManagedDatastoreS3Storage(CancelPipelineReprocessingResponse): """ Used to store data in an Amazon S3 bucket managed by IoT Analytics. You can't change the choice of Amazon S3 storage after your data store is created. """ pass class JsonConfiguration(CancelPipelineReprocessingResponse): """ Contains the configuration information of the JSON format. """ pass class RoleArn(BaseModel): __root__: Annotated[str, Field(max_length=2048, min_length=20)] class LoggingLevel(Enum): ERROR = 'ERROR' class LoggingEnabled(UnlimitedRetentionPeriod): pass class MessagePayload(BaseModel): __root__: str class TagResourceResponse(CancelPipelineReprocessingResponse): pass class UntagResourceResponse(CancelPipelineReprocessingResponse): pass class TagKey(BaseModel): __root__: Annotated[str, Field(max_length=256, min_length=1)] class ActivityBatchSize(MaxVersions): pass class ActivityName(BaseModel): __root__: Annotated[str, Field(max_length=128, min_length=1)] class AttributeNameMapping(BaseModel): pass class Config: extra = Extra.allow class AttributeName(TagKey): pass class AttributeNames(BaseModel): __root__: Annotated[List[AttributeName], Field(max_items=50, min_items=1)] class MessageId(BaseModel): __root__: Annotated[str, Field(max_length=128, min_length=1, regex='\\p{ASCII}*')] class ErrorCode(MessagePayload): pass class ErrorMessage(MessagePayload): pass class ChannelName(BaseModel): __root__: Annotated[ str, Field(max_length=128, min_length=1, regex='(^(?!_{2}))(^[a-zA-Z0-9_]+$)') ] class BucketKeyExpression(BaseModel): __root__: Annotated[ str, Field(max_length=255, min_length=1, regex="^[a-zA-Z0-9!_.*'()/{}:-]*$") ] class BucketName(BaseModel): __root__: Annotated[ str, Field(max_length=255, min_length=3, regex='^[a-zA-Z0-9.\\-_]*$') ] class PipelineName(ChannelName): pass class ReprocessingId(MessagePayload): pass class CancelPipelineReprocessingRequest(BaseModel): pass class ChannelArn(MessagePayload): pass class ChannelStatus(Enum): CREATING = 'CREATING' ACTIVE = 'ACTIVE' DELETING = 'DELETING' class RetentionPeriod(BaseModel): """ How long, in days, message data is kept. """ unlimited: Optional[UnlimitedRetentionPeriod] = None numberOfDays: Optional[RetentionPeriodInDays] = None class Timestamp(BaseModel): __root__: datetime class ServiceManagedChannelS3StorageSummary(CancelPipelineReprocessingResponse): """ Used to store channel data in an S3 bucket managed by IoT Analytics. """ pass class ColumnName(BaseModel): __root__: Annotated[ str, Field( max_length=255, min_length=1, regex='^[A-Za-z_]([A-Za-z0-9]*|[A-Za-z0-9][A-Za-z0-9_]*)$', ), ] class ColumnDataType(BaseModel): __root__: Annotated[str, Field(max_length=131072, min_length=1)] class Column(BaseModel): """ Contains information about a column that stores your data. """ name: ColumnName type: ColumnDataType class Columns(BaseModel): __root__: List[Column] class ComputeType(Enum): ACU_1 = 'ACU_1' ACU_2 = 'ACU_2' class Image(BaseModel): __root__: Annotated[str, Field(max_length=255)] class DatasetName(ChannelName): pass class DatasetContentVersion(BaseModel): __root__: Annotated[str, Field(max_length=36, min_length=7)] class CreateDatasetContentRequest(BaseModel): versionId: Optional[DatasetContentVersion] = None class VersioningConfiguration(BaseModel): """ Information about the versioning of dataset contents. """ unlimited: Optional[UnlimitedVersioning] = None maxVersions: Optional[MaxVersions] = None class DatasetArn(MessagePayload): pass class DatastoreName(ChannelName): pass class DatastoreArn(MessagePayload): pass class PipelineArn(MessagePayload): pass class S3KeyPrefix(BaseModel): __root__: Annotated[ str, Field(max_length=255, min_length=1, regex="^[a-zA-Z0-9!_.*'()/{}:-]*/$") ] class CustomerManagedDatastoreS3StorageSummary(BaseModel): """ Contains information about the data store that you manage. """ bucket: Optional[BucketName] = None keyPrefix: Optional[S3KeyPrefix] = None roleArn: Optional[RoleArn] = None class DatasetActionName(BaseModel): __root__: Annotated[ str, Field(max_length=128, min_length=1, regex='^[a-zA-Z0-9_]+$') ] class DatasetActionType(Enum): QUERY = 'QUERY' CONTAINER = 'CONTAINER' class EntryName(MessagePayload): pass class DatasetContentState(Enum): CREATING = 'CREATING' SUCCEEDED = 'SUCCEEDED' FAILED = 'FAILED' class Reason(MessagePayload): pass class DatasetContentStatus(BaseModel): """ The state of the dataset contents and the reason they are in this state. """ state: Optional[DatasetContentState] = None reason: Optional[Reason] = None class DatasetContentSummary(BaseModel): """ Summary information about dataset contents. """ version: Optional[DatasetContentVersion] = None status: Optional[DatasetContentStatus] = None creationTime: Optional[Timestamp] = None scheduleTime: Optional[Timestamp] = None completionTime: Optional[Timestamp] = None class DatasetContentSummaries(BaseModel): __root__: List[DatasetContentSummary] class DatasetContentVersionValue(BaseModel): """ The dataset whose latest contents are used as input to the notebook or application. """ datasetName: DatasetName class PresignedURI(MessagePayload): pass class TriggeringDataset(BaseModel): """ Information about the dataset whose content generation triggers the new dataset content generation. """ name: DatasetName class IotSiteWiseCustomerManagedDatastoreS3Storage(BaseModel): """ Used to store data used by IoT SiteWise in an Amazon S3 bucket that you manage. You can't change the choice of Amazon S3 storage after your data store is created. """ bucket: BucketName keyPrefix: Optional[S3KeyPrefix] = None class IotSiteWiseCustomerManagedDatastoreS3StorageSummary(BaseModel): """ Contains information about the data store that you manage, which stores data used by IoT SiteWise. """ bucket: Optional[BucketName] = None keyPrefix: Optional[S3KeyPrefix] = None class DatastoreIotSiteWiseMultiLayerStorageSummary(BaseModel): """ Contains information about the data store that you manage, which stores data used by IoT SiteWise. """ customerManagedS3Storage: Optional[ IotSiteWiseCustomerManagedDatastoreS3StorageSummary ] = None class ServiceManagedDatastoreS3StorageSummary(CancelPipelineReprocessingResponse): """ Contains information about the data store that is managed by IoT Analytics. """ pass class DatastoreStorageSummary(BaseModel): """ Contains information about your data store. """ serviceManagedS3: Optional[ServiceManagedDatastoreS3StorageSummary] = None customerManagedS3: Optional[CustomerManagedDatastoreS3StorageSummary] = None iotSiteWiseMultiLayerStorage: Optional[ DatastoreIotSiteWiseMultiLayerStorageSummary ] = None class FileFormatType(Enum): JSON = 'JSON' PARQUET = 'PARQUET' class DeleteChannelRequest(BaseModel): pass class DeleteDatasetContentRequest(BaseModel): pass class DeleteDatasetRequest(BaseModel): pass class DeleteDatastoreRequest(BaseModel): pass class DeletePipelineRequest(BaseModel): pass class OffsetSeconds(BaseModel): __root__: int class TimeExpression(MessagePayload): pass class DeltaTime(BaseModel): """ Used to limit data to that which has arrived since the last execution of the action. """ offsetSeconds: OffsetSeconds timeExpression: TimeExpression class SessionTimeoutInMinutes(BaseModel): __root__: Annotated[int, Field(ge=1.0, le=60.0)] class DeltaTimeSessionWindowConfiguration(BaseModel): """ <p>A structure that contains the configuration information of a delta time session window.</p> <p> <a href="https://docs.aws.amazon.com/iotanalytics/latest/APIReference/API_DeltaTime.html"> <code>DeltaTime</code> </a> specifies a time interval. You can use <code>DeltaTime</code> to create dataset contents with data that has arrived in the data store since the last execution. For an example of <code>DeltaTime</code>, see <a href="https://docs.aws.amazon.com/iotanalytics/latest/userguide/automate-create-dataset.html#automate-example6"> Creating a SQL dataset with a delta window (CLI)</a> in the <i>IoT Analytics User Guide</i>.</p> """ timeoutInMinutes: SessionTimeoutInMinutes class IncludeStatisticsFlag(UnlimitedRetentionPeriod): pass class DescribeChannelRequest(BaseModel): pass class DescribeDatasetRequest(BaseModel): pass class DescribeDatastoreRequest(BaseModel): pass class DescribeLoggingOptionsRequest(BaseModel): pass class LoggingOptions(BaseModel): """ Information about logging options. """ roleArn: RoleArn level: LoggingLevel enabled: LoggingEnabled class DescribePipelineRequest(BaseModel): pass class DoubleValue(BaseModel): __root__: float class EndTime(Timestamp): pass class SizeInBytes(DoubleValue): pass class FilterExpression(TagKey): pass class GetDatasetContentRequest(BaseModel): pass class GlueTableName(BaseModel): __root__: Annotated[str, Field(max_length=150, min_length=1)] class GlueDatabaseName(GlueTableName): pass class GlueConfiguration(BaseModel): """ Configuration information for coordination with Glue, a fully managed extract, transform and load (ETL) service. """ tableName: GlueTableName databaseName: GlueDatabaseName class IotEventsInputName(BaseModel): __root__: Annotated[ str, Field(max_length=128, min_length=1, regex='^[a-zA-Z][a-zA-Z0-9_]*$') ] class LambdaName(BaseModel): __root__: Annotated[ str, Field(max_length=64, min_length=1, regex='^[a-zA-Z0-9_-]+$') ] class LateDataRuleName(DatasetActionName): pass class LateDataRuleConfiguration(BaseModel): """ The information needed to configure a delta time session window. """ deltaTimeSessionWindowConfiguration: Optional[ DeltaTimeSessionWindowConfiguration ] = None class NextToken(MessagePayload): pass class MaxResults(BaseModel): __root__: Annotated[int, Field(ge=1.0, le=250.0)] class ListChannelsRequest(BaseModel): pass class ListDatasetContentsRequest(BaseModel): pass class ListDatasetsRequest(BaseModel): pass class ListDatastoresRequest(BaseModel): pass class ListPipelinesRequest(BaseModel): pass class ResourceArn(RoleArn): pass class ListTagsForResourceRequest(BaseModel): pass class LogResult(MessagePayload): pass class MathExpression(TagKey): pass class MaxMessages(BaseModel): __root__: Annotated[int, Field(ge=1.0, le=10.0)] class MessagePayloads(BaseModel): __root__: Annotated[List[MessagePayload], Field(max_items=10, min_items=1)] class OutputFileName(BaseModel): __root__: Annotated[str, Field(regex='[\\w\\.-]{1,255}')] class OutputFileUriValue(BaseModel): """ The value of the variable as a structure that specifies an output file URI. """ fileName: OutputFileName class SchemaDefinition(BaseModel): """ Information needed to define a schema. """ columns: Optional[Columns] = None class PartitionAttributeName(DatasetActionName): pass class PutLoggingOptionsRequest(BaseModel): loggingOptions: LoggingOptions class QueryFilter(BaseModel): """ Information that is used to filter message data, to segregate it according to the timeframe in which it arrives. """ deltaTime: Optional[DeltaTime] = None class QueryFilters(BaseModel): __root__: Annotated[List[QueryFilter], Field(max_items=1, min_items=0)] class ReprocessingStatus(Enum): RUNNING = 'RUNNING' SUCCEEDED = 'SUCCEEDED' CANCELLED = 'CANCELLED' FAILED = 'FAILED' class ReprocessingSummary(BaseModel): """ Information about pipeline reprocessing. """ id: Optional[ReprocessingId] = None status: Optional[ReprocessingStatus] = None creationTime: Optional[Timestamp] = None class VolumeSizeInGB(BaseModel): __root__: Annotated[int, Field(ge=1.0, le=50.0)] class S3PathChannelMessage(BaseModel): __root__: Annotated[ str, Field( max_length=1024, min_length=1, regex="^[a-zA-Z0-9/_!'(){}\\*\\s\\.\\-\\=\\:]+$", ), ] class StartTime(Timestamp): pass class SampleChannelDataRequest(BaseModel): pass class ScheduleExpression(MessagePayload): pass class SqlQuery(MessagePayload): pass class StringValue(BaseModel): __root__: Annotated[str, Field(max_length=1024, min_length=0)] class TagValue(TagKey): pass class TagKeyList(BaseModel): __root__: Annotated[List[TagKey], Field(max_items=50, min_items=1)] class TimestampFormat(BaseModel): __root__: Annotated[ str, Field(max_length=50, min_length=1, regex="^[a-zA-Z0-9\\s\\[\\]_,.'/:-]*$") ] class UntagResourceRequest(BaseModel): pass class VariableName(TagKey): pass class Variable(BaseModel): """ An instance of a variable to be passed to the <code>containerAction</code> execution. Each variable must have a name and a value given by one of <code>stringValue</code>, <code>datasetContentVersionValue</code>, or <code>outputFileUriValue</code>. """ name: VariableName stringValue: Optional[StringValue] = None doubleValue: Optional[DoubleValue] = None datasetContentVersionValue: Optional[DatasetContentVersionValue] = None outputFileUriValue: Optional[OutputFileUriValue] = None class Message(BaseModel): """ Information about a message. """ messageId: MessageId payload: MessagePayload class CreateChannelResponse(BaseModel): channelName: Optional[ChannelName] = None channelArn: Optional[ChannelArn] = None retentionPeriod: Optional[RetentionPeriod] = None class CustomerManagedChannelS3Storage(BaseModel): """ Used to store channel data in an S3 bucket that you manage. If customer-managed storage is selected, the <code>retentionPeriod</code> parameter is ignored. You can't change the choice of S3 storage after the data store is created. """ bucket: BucketName keyPrefix: Optional[S3KeyPrefix] = None roleArn: RoleArn class Tag(BaseModel): """ A set of key-value pairs that are used to manage the resource. """ key: TagKey value: TagValue class CreateDatasetResponse(BaseModel): datasetName: Optional[DatasetName] = None datasetArn: Optional[DatasetArn] = None retentionPeriod: Optional[RetentionPeriod] = None class LateDataRule(BaseModel): """ A structure that contains the name and configuration information of a late data rule. """ ruleName: Optional[LateDataRuleName] = None ruleConfiguration: LateDataRuleConfiguration class CreateDatasetContentResponse(BaseModel): versionId: Optional[DatasetContentVersion] = None class CreateDatastoreResponse(BaseModel): datastoreName: Optional[DatastoreName] = None datastoreArn: Optional[DatastoreArn] = None retentionPeriod: Optional[RetentionPeriod] = None class CustomerManagedDatastoreS3Storage(CustomerManagedChannelS3Storage): """ S3-customer-managed; When you choose customer-managed storage, the <code>retentionPeriod</code> parameter is ignored. You can't change the choice of Amazon S3 storage after your data store is created. """ pass class DatastoreIotSiteWiseMultiLayerStorage(BaseModel): """ Used to store data used by IoT SiteWise in an Amazon S3 bucket that you manage. You can't change the choice of Amazon S3 storage after your data store is created. """ customerManagedS3Storage: IotSiteWiseCustomerManagedDatastoreS3Storage class ParquetConfiguration(BaseModel): """ Contains the configuration information of the Parquet format. """ schemaDefinition: Optional[SchemaDefinition] = None class CreatePipelineResponse(BaseModel): pipelineName: Optional[PipelineName] = None pipelineArn: Optional[PipelineArn] = None class DescribeLoggingOptionsResponse(BaseModel): loggingOptions: Optional[LoggingOptions] = None class ListDatasetContentsResponse(BaseModel): datasetContentSummaries: Optional[DatasetContentSummaries] = None nextToken: Optional[NextToken] = None class RunPipelineActivityResponse(BaseModel): payloads: Optional[MessagePayloads] = None logResult: Optional[LogResult] = None class ChannelActivity(BaseModel): """ The activity that determines the source of the messages to be processed. """ name: ActivityName channelName: ChannelName next: Optional[ActivityName] = None class LambdaActivity(BaseModel): """ An activity that runs a Lambda function to modify the message. """ name: ActivityName lambdaName: LambdaName batchSize: ActivityBatchSize next: Optional[ActivityName] = None class DatastoreActivity(BaseModel): """ The datastore activity that specifies where to store the processed data. """ name: ActivityName datastoreName: DatastoreName class AddAttributesActivity(BaseModel): """ An activity that adds other attributes based on existing attributes in the message. """ name: ActivityName attributes: AttributeNameMapping next: Optional[ActivityName] = None class RemoveAttributesActivity(BaseModel): """ An activity that removes attributes from a message. """ name: ActivityName attributes: AttributeNames next: Optional[ActivityName] = None class SelectAttributesActivity(RemoveAttributesActivity): """ Used to create a new message using only the specified attributes from the original message. """ pass class FilterActivity(BaseModel): """ An activity that filters a message based on its attributes. """ name: ActivityName filter: FilterExpression next: Optional[ActivityName] = None class MathActivity(BaseModel): """ An activity that computes an arithmetic expression using the message's attributes. """ name: ActivityName attribute: AttributeName math: MathExpression next: Optional[ActivityName] = None class DeviceRegistryEnrichActivity(BaseModel): """ An activity that adds data from the IoT device registry to your message. """ name: ActivityName attribute: AttributeName thingName: AttributeName roleArn: RoleArn next: Optional[ActivityName] = None class DeviceShadowEnrichActivity(DeviceRegistryEnrichActivity): """ An activity that adds information from the IoT Device Shadow service to a message. """ pass class SampleChannelDataResponse(BaseModel): payloads: Optional[MessagePayloads] = None class StartPipelineReprocessingResponse(BaseModel): reprocessingId: Optional[ReprocessingId] = None class S3PathChannelMessages(BaseModel): __root__: Annotated[List[S3PathChannelMessage], Field(max_items=100, min_items=1)] class BatchPutMessageErrorEntry(BaseModel): """ Contains informations about errors. """ messageId: Optional[MessageId] = None errorCode: Optional[ErrorCode] = None errorMessage: Optional[ErrorMessage] = None class BatchPutMessageErrorEntries(BaseModel): __root__: List[BatchPutMessageErrorEntry] class Messages(BaseModel): __root__: List[Message] class BatchPutMessageRequest(BaseModel): channelName: ChannelName messages: Messages class ChannelStorage(BaseModel): """ Where channel data is stored. You may choose one of <code>serviceManagedS3</code>, <code>customerManagedS3</code> storage. If not specified, the default is <code>serviceManagedS3</code>. This can't be changed after creation of the channel. """ serviceManagedS3: Optional[ServiceManagedChannelS3Storage] = None customerManagedS3: Optional[CustomerManagedChannelS3Storage] = None class Channel(BaseModel): """ A collection of data from an MQTT topic. Channels archive the raw, unprocessed messages before publishing the data to a pipeline. """ name: Optional[ChannelName] = None storage: Optional[ChannelStorage] = None arn: Optional[ChannelArn] = None status: Optional[ChannelStatus] = None retentionPeriod: Optional[RetentionPeriod] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None lastMessageArrivalTime: Optional[Timestamp] = None class ChannelMessages(BaseModel): """ Specifies one or more sets of channel messages. """ s3Paths: Optional[S3PathChannelMessages] = None class EstimatedResourceSize(BaseModel): """ The estimated size of the resource. """ estimatedSizeInBytes: Optional[SizeInBytes] = None estimatedOn: Optional[Timestamp] = None class ChannelStatistics(BaseModel): """ Statistics information about the channel. """ size: Optional[EstimatedResourceSize] = None class CustomerManagedChannelS3StorageSummary(CustomerManagedDatastoreS3StorageSummary): """ Used to store channel data in an S3 bucket that you manage. """ pass class ChannelStorageSummary(BaseModel): """ Where channel data is stored. """ serviceManagedS3: Optional[ServiceManagedChannelS3StorageSummary] = None customerManagedS3: Optional[CustomerManagedChannelS3StorageSummary] = None class ChannelSummary(BaseModel): """ A summary of information about a channel. """ channelName: Optional[ChannelName] = None channelStorage: Optional[ChannelStorageSummary] = None status: Optional[ChannelStatus] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None lastMessageArrivalTime: Optional[Timestamp] = None class ChannelSummaries(BaseModel): __root__: List[ChannelSummary] class ResourceConfiguration(BaseModel): """ The configuration of the resource used to execute the <code>containerAction</code>. """ computeType: ComputeType volumeSizeInGB: VolumeSizeInGB class Variables(BaseModel): __root__: Annotated[List[Variable], Field(max_items=50, min_items=0)] class ContainerDatasetAction(BaseModel): """ Information required to run the <code>containerAction</code> to produce dataset contents. """ image: Image executionRoleArn: RoleArn resourceConfiguration: ResourceConfiguration variables: Optional[Variables] = None class TagList(BaseModel): __root__: Annotated[List[Tag], Field(max_items=50, min_items=1)] class CreateChannelRequest(BaseModel): channelName: ChannelName channelStorage: Optional[ChannelStorage] = None retentionPeriod: Optional[RetentionPeriod] = None tags: Optional[TagList] = None class LateDataRules(BaseModel): __root__: Annotated[List[LateDataRule], Field(max_items=1, min_items=1)] class DatastoreStorage(BaseModel): """ Where data in a data store is stored.. You can choose <code>serviceManagedS3</code> storage, <code>customerManagedS3</code> storage, or <code>iotSiteWiseMultiLayerStorage</code> storage. The default is <code>serviceManagedS3</code>. You can't change the choice of Amazon S3 storage after your data store is created. """ serviceManagedS3: Optional[ServiceManagedDatastoreS3Storage] = None customerManagedS3: Optional[CustomerManagedDatastoreS3Storage] = None iotSiteWiseMultiLayerStorage: Optional[DatastoreIotSiteWiseMultiLayerStorage] = None class FileFormatConfiguration(BaseModel): """ <p>Contains the configuration information of file formats. IoT Analytics data stores support JSON and <a href="https://parquet.apache.org/">Parquet</a>.</p> <p>The default file format is JSON. You can specify only one format.</p> <p>You can't change the file format after you create the data store.</p> """ jsonConfiguration: Optional[JsonConfiguration] = None parquetConfiguration: Optional[ParquetConfiguration] = None class SqlQueryDatasetAction(BaseModel): """ The SQL query to modify the message. """ sqlQuery: SqlQuery filters: Optional[QueryFilters] = None class DatasetActionSummary(BaseModel): """ Information about the action that automatically creates the dataset's contents. """ actionName: Optional[DatasetActionName] = None actionType: Optional[DatasetActionType] = None class DatasetActionSummaries(BaseModel): __root__: Annotated[List[DatasetActionSummary], Field(max_items=1, min_items=1)] class IotEventsDestinationConfiguration(BaseModel): """ Configuration information for delivery of dataset contents to IoT Events. """ inputName: IotEventsInputName roleArn: RoleArn class S3DestinationConfiguration(BaseModel): """ Configuration information for delivery of dataset contents to Amazon Simple Storage Service (Amazon S3). """ bucket: BucketName key: BucketKeyExpression glueConfiguration: Optional[GlueConfiguration] = None roleArn: RoleArn class DatasetContentDeliveryDestination(BaseModel): """ The destination to which dataset contents are delivered. """ iotEventsDestinationConfiguration: Optional[ IotEventsDestinationConfiguration ] = None s3DestinationConfiguration: Optional[S3DestinationConfiguration] = None class DatasetEntry(BaseModel): """ The reference to a dataset entry. """ entryName: Optional[EntryName] = None dataURI: Optional[PresignedURI] = None class DatasetEntries(BaseModel): __root__: List[DatasetEntry] class Schedule(BaseModel): """ The schedule for when to trigger an update. """ expression: Optional[ScheduleExpression] = None class Partition(BaseModel): """ A partition dimension defined by an attribute. """ attributeName: PartitionAttributeName class TimestampPartition(BaseModel): """ A partition dimension defined by a timestamp attribute. """ attributeName: PartitionAttributeName timestampFormat: Optional[TimestampFormat] = None class DatastorePartition(BaseModel): """ A single dimension to partition a data store. The dimension must be an <code>AttributePartition</code> or a <code>TimestampPartition</code>. """ attributePartition: Optional[Partition] = None timestampPartition: Optional[TimestampPartition] = None class DatastoreStatistics(ChannelStatistics): """ Statistical information about the data store. """ pass class ReprocessingSummaries(BaseModel): __root__: List[ReprocessingSummary] class PipelineSummary(BaseModel): """ A summary of information about a pipeline. """ pipelineName: Optional[PipelineName] = None reprocessingSummaries: Optional[ReprocessingSummaries] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None class StartPipelineReprocessingRequest(BaseModel): startTime: Optional[StartTime] = None endTime: Optional[EndTime] = None channelMessages: Optional[ChannelMessages] = None class TagResourceRequest(BaseModel): tags: TagList class UpdateChannelRequest(BaseModel): channelStorage: Optional[ChannelStorage] = None retentionPeriod: Optional[RetentionPeriod] = None class UpdateDatastoreRequest(BaseModel): retentionPeriod: Optional[RetentionPeriod] = None datastoreStorage: Optional[DatastoreStorage] = None fileFormatConfiguration: Optional[FileFormatConfiguration] = None class BatchPutMessageResponse(BaseModel): batchPutMessageErrorEntries: Optional[BatchPutMessageErrorEntries] = None class DatasetAction(BaseModel): """ A <code>DatasetAction</code> object that specifies how dataset contents are automatically created. """ actionName: Optional[DatasetActionName] = None queryAction: Optional[SqlQueryDatasetAction] = None containerAction: Optional[ContainerDatasetAction] = None class DatasetTrigger(BaseModel): """ The <code>DatasetTrigger</code> that specifies when the dataset is automatically updated. """ schedule: Optional[Schedule] = None dataset: Optional[TriggeringDataset] = None class DatasetContentDeliveryRule(BaseModel): """ When dataset contents are created, they are delivered to destination specified here. """ entryName: Optional[EntryName] = None destination: DatasetContentDeliveryDestination class Partitions(BaseModel): __root__: Annotated[List[DatastorePartition], Field(max_items=25, min_items=0)] class PipelineActivity(BaseModel): """ An activity that performs a transformation on a message. """ channel: Optional[ChannelActivity] = None lambda_: Annotated[Optional[LambdaActivity], Field(alias='lambda')] = None datastore: Optional[DatastoreActivity] = None addAttributes: Optional[AddAttributesActivity] = None removeAttributes: Optional[RemoveAttributesActivity] = None selectAttributes: Optional[SelectAttributesActivity] = None filter: Optional[FilterActivity] = None math: Optional[MathActivity] = None deviceRegistryEnrich: Optional[DeviceRegistryEnrichActivity] = None deviceShadowEnrich: Optional[DeviceShadowEnrichActivity] = None class DescribeChannelResponse(BaseModel): channel: Optional[Channel] = None statistics: Optional[ChannelStatistics] = None class GetDatasetContentResponse(BaseModel): entries: Optional[DatasetEntries] = None timestamp: Optional[Timestamp] = None status: Optional[DatasetContentStatus] = None class ListChannelsResponse(BaseModel): channelSummaries: Optional[ChannelSummaries] = None nextToken: Optional[NextToken] = None class ListTagsForResourceResponse(BaseModel): tags: Optional[TagList] = None class DatasetActions(BaseModel): __root__: Annotated[List[DatasetAction], Field(max_items=1, min_items=1)] class DatasetTriggers(BaseModel): __root__: Annotated[List[DatasetTrigger], Field(max_items=5, min_items=0)] class DatasetContentDeliveryRules(BaseModel): __root__: Annotated[ List[DatasetContentDeliveryRule], Field(max_items=20, min_items=0) ] class CreateDatasetRequest(BaseModel): datasetName: DatasetName actions: DatasetActions triggers: Optional[DatasetTriggers] = None contentDeliveryRules: Optional[DatasetContentDeliveryRules] = None retentionPeriod: Optional[RetentionPeriod] = None versioningConfiguration: Optional[VersioningConfiguration] = None tags: Optional[TagList] = None lateDataRules: Optional[LateDataRules] = None class DatastorePartitions(BaseModel): """ Contains information about the partition dimensions in a data store. """ partitions: Optional[Partitions] = None class CreateDatastoreRequest(BaseModel): datastoreName: DatastoreName datastoreStorage: Optional[DatastoreStorage] = None retentionPeriod: Optional[RetentionPeriod] = None tags: Optional[TagList] = None fileFormatConfiguration: Optional[FileFormatConfiguration] = None datastorePartitions: Optional[DatastorePartitions] = None class PipelineActivities(BaseModel): __root__: Annotated[List[PipelineActivity], Field(max_items=25, min_items=1)] class CreatePipelineRequest(BaseModel): pipelineName: PipelineName pipelineActivities: PipelineActivities tags: Optional[TagList] = None class Dataset(BaseModel): """ Information about a dataset. """ name: Optional[DatasetName] = None arn: Optional[DatasetArn] = None actions: Optional[DatasetActions] = None triggers: Optional[DatasetTriggers] = None contentDeliveryRules: Optional[DatasetContentDeliveryRules] = None status: Optional[ChannelStatus] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None retentionPeriod: Optional[RetentionPeriod] = None versioningConfiguration: Optional[VersioningConfiguration] = None lateDataRules: Optional[LateDataRules] = None class DatasetSummary(BaseModel): """ A summary of information about a dataset. """ datasetName: Optional[DatasetName] = None status: Optional[ChannelStatus] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None triggers: Optional[DatasetTriggers] = None actions: Optional[DatasetActionSummaries] = None class DatasetSummaries(BaseModel): __root__: List[DatasetSummary] class Datastore(BaseModel): """ Information about a data store. """ name: Optional[DatastoreName] = None storage: Optional[DatastoreStorage] = None arn: Optional[DatastoreArn] = None status: Optional[ChannelStatus] = None retentionPeriod: Optional[RetentionPeriod] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None lastMessageArrivalTime: Optional[Timestamp] = None fileFormatConfiguration: Optional[FileFormatConfiguration] = None datastorePartitions: Optional[DatastorePartitions] = None class DatastoreSummary(BaseModel): """ A summary of information about a data store. """ datastoreName: Optional[DatastoreName] = None datastoreStorage: Optional[DatastoreStorageSummary] = None status: Optional[ChannelStatus] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None lastMessageArrivalTime: Optional[Timestamp] = None fileFormatType: Optional[FileFormatType] = None datastorePartitions: Optional[DatastorePartitions] = None class DatastoreSummaries(BaseModel): __root__: List[DatastoreSummary] class Pipeline(BaseModel): """ Contains information about a pipeline. """ name: Optional[PipelineName] = None arn: Optional[PipelineArn] = None activities: Optional[PipelineActivities] = None reprocessingSummaries: Optional[ReprocessingSummaries] = None creationTime: Optional[Timestamp] = None lastUpdateTime: Optional[Timestamp] = None class PipelineSummaries(BaseModel): __root__: List[PipelineSummary] class RunPipelineActivityRequest(BaseModel): pipelineActivity: PipelineActivity payloads: MessagePayloads class UpdateDatasetRequest(BaseModel): actions: DatasetActions triggers: Optional[DatasetTriggers] = None contentDeliveryRules: Optional[DatasetContentDeliveryRules] = None retentionPeriod: Optional[RetentionPeriod] = None versioningConfiguration: Optional[VersioningConfiguration] = None lateDataRules: Optional[LateDataRules] = None class UpdatePipelineRequest(BaseModel): pipelineActivities: PipelineActivities class DescribeDatasetResponse(BaseModel): dataset: Optional[Dataset] = None class DescribeDatastoreResponse(BaseModel): datastore: Optional[Datastore] = None statistics: Optional[DatastoreStatistics] = None class DescribePipelineResponse(BaseModel): pipeline: Optional[Pipeline] = None class ListDatasetsResponse(BaseModel): datasetSummaries: Optional[DatasetSummaries] = None nextToken: Optional[NextToken] = None class ListDatastoresResponse(BaseModel): datastoreSummaries: Optional[DatastoreSummaries] = None nextToken: Optional[NextToken] = None class ListPipelinesResponse(BaseModel): pipelineSummaries: Optional[PipelineSummaries] = None nextToken: Optional[NextToken] = None
37,380
10,987
# -*- coding: utf-8 -*- """ Created on Sun Jan 25 15:48:52 2015 @author: Sofia """ import csv import json import os sourceEncoding = "iso-8859-1" targetEncoding = "utf-8" # encode files to utf8 (source: http://stackoverflow.com/questions/191359/how-to-convert-a-file-to-utf-8-in-python) csvfile = open('..\Data\AMFI.csv',"r") csvfile_encoded = open("..\Data\AMFI_encoded.csv", "w") csvfile_encoded.write(unicode(csvfile.read(), sourceEncoding).encode(targetEncoding)) csvfile_encoded.close() csvfile = open('..\Data\AMFI_categories.csv',"r") csvfile_encoded = open("..\Data\AMFIcategories_encoded.csv", "w") csvfile_encoded.write(unicode(csvfile.read(), sourceEncoding).encode(targetEncoding)) csvfile_encoded.close() csvfile = open('..\Data\AMFI_domains.csv',"r") csvfile_encoded = open("..\Data\AMFIdomains_encoded.csv", "w") csvfile_encoded.write(unicode(csvfile.read(), sourceEncoding).encode(targetEncoding)) csvfile_encoded.close() # open files AMFI_books = open("..\Data\AMFI_encoded.csv","r") AMFI_categories = open("..\Data\AMFIcategories_encoded.csv","r") AMFI_domains = open("..\Data\AMFIdomains_encoded.csv","r") # define fieldnames fieldnames_books = ("Callnumber","Barcode","Title","Year","Location") fieldnames_categories = ("Barcode","Category") # put data in reader reader_books = csv.DictReader(AMFI_books, fieldnames_books, delimiter=';') reader_categories = csv.DictReader(AMFI_categories, fieldnames_categories, delimiter = ';') reader_domains = csv.DictReader(AMFI_domains, delimiter = ';') output = {"name": "Library of the University of Applied Sciences", "type":"parent", "total":5605, "value":50, "children": []} # get data from reader_books barcode_books = [] names_books = [] tags_books = [] copies = [] for books in reader_books: barcode_books.append(books["Callnumber"]) names_books.append(books["Title"]) tags_books.append(books["Barcode"]) tags = [] size_books = len(barcode_books) # Modify data books for k in range(0, len(names_books), 1): # count copies count = names_books.count(names_books[k]) copies.append(count) # collect unique ids indeces = [i for i, x in enumerate(names_books) if x == names_books[k]] if len(indeces) == 1: tags.append(tags_books[indeces[0]]) else: list_tags = [] for w in range(0,len(indeces),1): tag = tags_books[indeces[w]] list_tags.append(tag) tags.append(list_tags) # set copies to NaN for t in range(1,len(indeces),1): names_books[indeces[t]] = "NaN" # Enter domains barcode_domain = [] for domain in reader_domains: output["children"].append({ "type": "domain", "name": domain["Domain"], "barcode": domain["Barcode"], "value": 6, "children": [] }) barcode_domain.append(domain["Barcode"]) # get category data barcode_category = [] names_category = [] for category in reader_categories: barcode_category.append(category["Barcode"]) names_category.append(category["Category"]) # Enter categories for i in range(0,len(barcode_domain),1): barcode_domain_values = output["children"][i]["barcode"] for j in range(0,len(barcode_category),1): if barcode_category[j] < barcode_domain_values: if names_category[j] != "NaN": output["children"][i]["children"].append({ "type":"category", "barcode": barcode_category[j], "value": 5, "name": names_category[j], "children": [] }) names_category[j] = "NaN" # append data to output lengths = [] codes_categories =[] for i in range(0,len(barcode_domain),1): lengths.append(len(output["children"][i]["children"])) for k in range(0, lengths[i], 1): #counter = 0 codes_categories = output["children"][i]["children"][k]["barcode"] for j in range(0,len(names_books),1): if barcode_books[j] < codes_categories: if names_books[j] != "NaN": output["children"][i]["children"][k]["children"].append({ "type":"book", "barcode": barcode_books[j], "tags": tags[j], "value": 2, "name": names_books[j], "copies": copies[j] }) names_books[j] = "NaN" # write data to file with open('../Data/tree.json', 'w') as f: json.dump(output, f, indent=True)
4,577
1,496
import datetime import os import unittest from io import StringIO from tempfile import mkdtemp from unittest.mock import patch from django.conf import settings from django.core.exceptions import ImproperlyConfigured from django.core.management import call_command as real_call_command from django.core.management.base import CommandError from django.test import TestCase from whoosh.qparser import QueryParser from haystack import connections, constants, indexes from haystack.utils.loading import UnifiedIndex from ..core.models import MockModel from .test_whoosh_backend import WhooshMockSearchIndex from .testcases import WhooshTestCase def call_command(*args, **kwargs): kwargs["using"] = ["whoosh"] print(args, kwargs) real_call_command(*args, **kwargs) class ManagementCommandTestCase(WhooshTestCase): fixtures = ["bulk_data"] def setUp(self): super().setUp() self.old_ui = connections["whoosh"].get_unified_index() self.ui = UnifiedIndex() self.wmmi = WhooshMockSearchIndex() self.ui.build(indexes=[self.wmmi]) self.sb = connections["whoosh"].get_backend() connections["whoosh"]._index = self.ui self.sb.setup() self.raw_whoosh = self.sb.index self.parser = QueryParser(self.sb.content_field_name, schema=self.sb.schema) self.sb.delete_index() self.sample_objs = MockModel.objects.all() def tearDown(self): connections["whoosh"]._index = self.old_ui super().tearDown() def verify_indexed_document_count(self, expected): with self.raw_whoosh.searcher() as searcher: count = searcher.doc_count() self.assertEqual(count, expected) def verify_indexed_documents(self): """Confirm that the documents in the search index match the database""" with self.raw_whoosh.searcher() as searcher: count = searcher.doc_count() self.assertEqual(count, 23) indexed_doc_ids = set(i["id"] for i in searcher.documents()) expected_doc_ids = set( "core.mockmodel.%d" % i for i in MockModel.objects.values_list("pk", flat=True) ) self.assertSetEqual(indexed_doc_ids, expected_doc_ids) def test_basic_commands(self): call_command("clear_index", interactive=False, verbosity=0) self.verify_indexed_document_count(0) call_command("update_index", verbosity=0) self.verify_indexed_documents() call_command("clear_index", interactive=False, verbosity=0) self.verify_indexed_document_count(0) call_command("rebuild_index", interactive=False, verbosity=0) self.verify_indexed_documents() def test_remove(self): call_command("clear_index", interactive=False, verbosity=0) self.verify_indexed_document_count(0) call_command("update_index", verbosity=0) self.verify_indexed_documents() # Remove several instances. MockModel.objects.get(pk=1).delete() MockModel.objects.get(pk=2).delete() MockModel.objects.get(pk=8).delete() self.verify_indexed_document_count(23) # Plain ``update_index`` doesn't fix it. call_command("update_index", verbosity=0) self.verify_indexed_document_count(23) # … but remove does: call_command("update_index", remove=True, verbosity=0) self.verify_indexed_document_count(20) def test_multiprocessing(self): call_command("clear_index", interactive=False, verbosity=0) self.verify_indexed_document_count(0) call_command("update_index", verbosity=2, workers=2, batchsize=5) self.verify_indexed_documents()
3,751
1,139
from http.server import HTTPServer, BaseHTTPRequestHandler from urllib.parse import urlparse, parse_qs from cowpy import cow import json import sys class SimpleHTTPRequestHandler(BaseHTTPRequestHandler): def do_GET(self): parsed_path = urlparse(self.path) parsed_qs = parse_qs(parsed_path.query) # import pdb; pdb.set_trace() if parsed_path.path == '/': self.send_response(200) self.end_headers() self.wfile.write(return_html_string()) return elif parsed_path.path == '/cowsay': self.send_response(200) self.end_headers() self.wfile.write(b'Helpful instructions about this application') return elif parsed_path.path == '/cow': try: # import pdb; pdb.set_trace() msg = parsed_qs['msg'][0] print(msg) except (KeyError, json.decoder.JSONDecodeError): self.send_response(400) self.end_headers() self.wfile.write(b'You did a bad thing') return cheese = cow.Moose(thoughts=True) message = cheese.milk(msg) self.send_response(200) self.end_headers() self.wfile.write(message.encode('utf8')) return else: self.send_response(404) self.end_headers() self.wfile.write(b'Not Found') def do_POST(self): parsed_path = urlparse(self.path) parsed_qs = parse_qs(parsed_path.query) if parsed_path.path == '/cow': try: msg = parsed_qs['msg'][0] cheese = cow.Moose(thoughts=True) message = cheese.milk(msg) post_dict = {} post_dict['content'] = message self.send_response(200) self.end_headers() self.wfile.write(json.dumps(post_dict).encode('utf8')) return except (KeyError, json.decoder.JSONDecodeError): self.send_response(400) self.end_headers() self.wfile.write(b'You did a bad thing') return def create_server(): return HTTPServer(('127.0.0.1', 3000), SimpleHTTPRequestHandler) def run_forever(): server = create_server() try: print('Starting server on port 3000') server.serve_forever() except KeyboardInterrupt: server.shutdown() server.server_close() # sys.exit() def return_html_string(): return b'''<!DOCTYPE html> <html> <head> <title> cowsay </title> </head> <body> <header> <nav> <ul> <li><a href="/cowsay">cowsay</a></li> </ul> </nav> <header> <main> <!-- project description --> </main> </body> </html>''' if __name__ == '__main__': run_forever()
2,977
912
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Oct 30 00:58:34 2018 @author: jeff """ '''************************************* #1. Import libraries and key varable values ''' import os import quandl import pandas as pd import numpy as np import keras from PIL import Image #folder path folder_path = os.path.dirname(__file__) #date range for full dataset str_dte = '2003-01-01' end_dte = '2018-7-31' date_dict = {'gte':str_dte, 'lte':end_dte} #Dates for back-testing start_dte = '2015-1-1' #Create list of dates datelist = pd.date_range(start_dte, periods=365*2).tolist() #API key for quandl quandl.ApiConfig.api_key = '[quandl id]' #Parameters for the image generation col_num_mid = 10 col_num_dte = 9 pixel_size = 100 window_size = 60 pred_window_size = 1 #model path model_path = "model2_2DCov.h5" model = keras.models.load_model(model_path) #number of channel for the image num_channel=1 #strategies parameters curr_pnl = 10000 curr_pnl_0=curr_pnl curr_pnl_1=curr_pnl curr_pnl_2=curr_pnl quant_trans_0 = 0 quant_trans_1 = 0 quant_trans_2 = 0 min_pnl = 0.0005 trading_cost = 0 trade_limit = 0.5 '''************************************* #2. Define functions ''' #input_X is a series of price #output_X is a series of price expressed in pixel def rescale(input_X, pixel, min_x,max_x): unit = (max_x - min_x)/pixel output_X = round((input_X-min_x)/unit,0) return output_X,unit '''************************************* #3. Running the test ''' #Get the data tkr = 'VTV' df =quandl.get_table('SHARADAR/SFP',date=date_dict,ticker=tkr) df = df.sort_values(by=['date']) df=df.reset_index(drop=True) #write header for the log of the strategy back-testing f = open('log.txt','w+') f.write('strategy\tBuySell\t' + 'dte' +'\t'+ 'cost' +'\t'+ 'T+1_actual' +'\t'+ 'T+1_pred'+'\t'+ 'Quantity'+'\t'+ 'PnL'+'\n') #loop through the dates for pred_dte in datelist: df_i = df.index[df['date']==pred_dte] #make sure both start and end dates are valid if df_i.empty: print('no data') continue df_i = df_i[0] print(pred_dte) df_start = df_i-(window_size) #starts at zero if df_start < 0: #in case the date inputted is not valid print('later date') continue #prepare the input data df['mid'] = (df['high'] + df['low'])/2 df_plot = df.iloc[df_start:df_i,:] min_p = min(df_plot['mid']) max_p = max(df_plot['mid']) output_pixel,unit = rescale(df_plot['mid'],pixel_size,min_p,max_p) #if no trend, then drop this data point if min_p ==max_p: print('no trend') continue #stack up for a numpy for Image Recognition #print the historical data img_ar = np.zeros((1,pixel_size,window_size,num_channel)) img_display = np.zeros((pixel_size,window_size,num_channel)) k=0 pix_p=0 for pix in output_pixel: y_pos = int(pix)-1 img_ar[0][y_pos][k][num_channel-1] = 255 img_display[y_pos][k][num_channel-1] = 255 pix_p=y_pos k+=1 img_row = img_ar/255 last_actual_p = pix_p * unit + min_p #make prediction pred_y = model.predict(img_row) max_y_val = max(pred_y[0]) pred_y_img = np.zeros((pixel_size,1)) #Obtain predicted price pred_pixel = 0 expected_p = 0 #calculate expected values for i in range(pixel_size): expected_p += pred_y_img[i,0] * i if pred_y[0,i] == max_y_val: pred_y_img[i,0] = 255 pred_pixel = i pred_p = pred_pixel * unit + min_p print('cost at ' + str(last_actual_p)) print('predict p be ' + str(pred_p) + ' and probability of ' + str(max_y_val)) pred_exp_p = expected_p * unit + min_p print('expected predict p be ' + str(pred_exp_p)) y_actual_p = df.iloc[df_i+1,:]['mid'] print('actual p be '+str(y_actual_p)) #Strategy Back-Testing #Benchmark - Strategy 0 - buy and hold if quant_trans_0 == 0: quant_trans_0 = curr_pnl/y_actual_p pnl = 0-trading_cost else: pnl = (y_actual_p/last_actual_p-1) * quant_trans_0 curr_pnl_0 += pnl f.write('B0\tNA\t' + str(pred_dte) +'\t'+ str(last_actual_p) +'\t'+ str(y_actual_p) +'\t'+ str(y_actual_p)+'\t'+ str(1)+'\t'+ str(last_actual_p-y_actual_p)+'\n') #Testing of strategy1 order_type = "" quant_trans_1 = int(curr_pnl_1/last_actual_p*0.5) if abs(pred_exp_p/last_actual_p-1)>min_pnl: if pred_exp_p>last_actual_p: #buy one now / long one unit #stock_unit_1+=quant_trans_1 pnl = (y_actual_p-last_actual_p) * quant_trans_1-trading_cost order_type = "B" curr_pnl_1 += pnl f.write('S1\tBuy\t' + str(pred_dte) +'\t'+ str(last_actual_p) +'\t'+ str(y_actual_p) +'\t'+ str(pred_exp_p)+'\t'+ str(quant_trans_1)+'\t'+ str(y_actual_p-last_actual_p)+'\n') elif pred_exp_p<last_actual_p: #sell one now / short one unit #stock_unit_1-=quant_trans_1 pnl = (last_actual_p-y_actual_p) * quant_trans_1-trading_cost order_type = "S" curr_pnl_1 += pnl f.write('S1\tSell\t' + str(pred_dte) +'\t'+ str(last_actual_p) +'\t'+ str(y_actual_p) +'\t'+ str(pred_exp_p)+'\t'+ str(quant_trans_1)+'\t'+ str(last_actual_p-y_actual_p)+'\n') else: #no trade if order_type == "B": pnl = (y_actual_p-last_actual_p) * quant_trans_1 else: pnl = (last_actual_p-y_actual_p) * quant_trans_1 curr_pnl_1 += pnl #Testing of strategy2 if max_y_val > 0.99 and abs(pred_p/last_actual_p-1)>min_pnl: quant_trans_2 = int(curr_pnl_2/last_actual_p*0.5) if pred_p>last_actual_p: #buy one now / long one unit #stock_unit_2+=quant_trans_2 order_type = "B" curr_pnl_2 += (y_actual_p-last_actual_p) * quant_trans_2-trading_cost f.write('S2\tBuy\t' + str(pred_dte) +'\t'+ str(last_actual_p) +'\t'+ str(y_actual_p) +'\t'+ str(pred_p) +'\t'+str(quant_trans_2)+'\t'+ str(y_actual_p-last_actual_p)+'\n') elif pred_p<last_actual_p: #sell one now / short one unit #stock_unit_2-=quant_trans_2 order_type = "S" curr_pnl_2 += (last_actual_p-y_actual_p) * quant_trans_2-trading_cost f.write('S2\tSell\t' + str(pred_dte) +'\t'+ str(last_actual_p) +'\t'+ str(y_actual_p) +'\t'+ str(pred_p)+'\t'+ str(quant_trans_2)+'\t'+ str(last_actual_p-y_actual_p)+'\n') else: #no trade if order_type == "B": pnl = (y_actual_p-last_actual_p) * quant_trans_2 else: pnl = (last_actual_p-y_actual_p) * quant_trans_2 curr_pnl_2 += pnl #print the final result of the strategies print(curr_pnl_0) print(curr_pnl_1) print(curr_pnl_2) f.close() ''' export CUDA_VISIBLE_DEVICES='' tensorboard --logdir AI_Finance_book/6B_TrendFollowings/Graph/ --host localhost --port 6006 '''
7,033
2,860
# -*- coding: utf-8 -*- from lichee import plugin from .field_parser_base import BaseFieldParser import os from PIL import Image from torchvision import transforms import torch from lichee.utils import storage @plugin.register_plugin(plugin.PluginType.FIELD_PARSER, "image_local_path") class ImgDataFieldParser(BaseFieldParser): """The field parser for local image. Read the image data from the path provided, transforms through ToSensor, Resize and Normalize. Attributes ---------- transformer: transforms.Compose compose the transforms(ToSensor, Resize and Normalize) """ def __init__(self): super().__init__() self.transformer = None def init(self, cfg): self.cfg = cfg resolution = [int(x) for x in self.global_config.DATASET.CONFIG.IMAGE_RESOLUTION] self.transformer = transforms.Compose([ transforms.ToTensor(), transforms.Resize(resolution), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) ]) def parse(self, row, training=False): """Parse the row and obtain the path of image, invoke prepare_img_data to transform the image data to tensor. Parameters ---------- row: memoryview Object contained in a single record training: bool inherited from parent, not used here. Returns ------- record: torch.Tensor the tensor of image data """ record = {} if self.key not in row: raise Exception("Cannot find key %s in row by image_local_path" % self.key) img_path = bytes(row[self.key]).decode("utf-8") if img_path[0] != "/": img_path = os.path.join(self.global_config.DATASET.DATA_BASE_DIR, img_path) record[self.alias] = self.prepare_img_data(img_path) return record def prepare_img_data(self, img_path): """Read and process the image from image_path Parameters ---------- img_path: str path of image Returns ------ torch.Tensor the tensor transformed from image data. """ with open(storage.get_storage_file(img_path), 'rb') as f: img = Image.open(f) img = img.convert('RGB') return self.transformer(img) def collate(self, batch): record = {} imgs = [instance[self.alias] for instance in batch] imgs = torch.stack(imgs) record[self.alias] = imgs return record
2,589
767
import torch from torch import nn import torch.nn.functional as F import torchvision from torchvision.models.inception import BasicConv2d, InceptionAux import pretrainedmodels from common import num_class class InceptionV3(nn.Module): def __init__(self, pre=True): super().__init__() self.encoder = torchvision.models.inception_v3(pretrained=pre) conv1 = BasicConv2d(4, 32, kernel_size=3, stride=2) if pre: w = self.encoder.Conv2d_1a_3x3.conv.weight conv1.conv.weight = nn.Parameter(torch.cat((w, 0.5 * (w[:, :1, :, :] + w[:, 2:, :, :])), dim=1)) self.encoder.Conv2d_1a_3x3 = conv1 self.encoder.AuxLogits = InceptionAux(768, num_class()) self.encoder.fc = nn.Linear(2048, num_class()) def forward(self, x): x = torch.nn.functional.interpolate(x, size=(299, 299), mode='bilinear') # resize if self.training: x, x_aux, feat = self.encoder(x) x = (torch.sigmoid(x) + torch.sigmoid(x_aux)) * 0.5 else: x, feat = self.encoder(x) x = torch.sigmoid(x) return {'logit': x, 'feat': feat} class InceptionV4(nn.Module): def __init__(self, pre=True): super().__init__() self.encoder = pretrainedmodels.__dict__['inceptionv4'](num_classes=1000, pretrained='imagenet') conv1 = BasicConv2d(4, 32, kernel_size=3, stride=2) if pre: w = self.encoder.features[0].conv.weight conv1.conv.weight = nn.Parameter(torch.cat((w, 0.5 * (w[:, :1, :, :] + w[:, 2:, :, :])), dim=1)) self.encoder.features[0].conv = conv1 self.last_linear = nn.Linear(1536, num_class()) pass def forward(self, x): # x = torch.nn.functional.interpolate(x, size=(299, 299), mode='bilinear') x = self.encoder.features(x) x = F.adaptive_avg_pool2d(x, (1, 1)) x = x.view(x.size(0), -1) feat = x x = self.last_linear(x) x = torch.sigmoid(x) return {'logit': x, 'feat': feat}
2,054
816
name = "Saeed" cordinates = (10.0, 20.0) names = ["Saeed", "Bob", "Mousa"]
74
40
#!/usr/bin/python ############################################## ###Python template ###Author: Elizabeth Lee ###Date: 9/2/14 ###Function: mean peak-based retro zOR metric vs. total attack rate ###Import data: SQL_export/OR_allweeks_outpatient.csv, SQL_export/OR_allweeks.csv ###Command Line: python Supp_zOR_totalAR.py ############################################## ### notes ### ### packages/modules ### import csv import matplotlib.pyplot as plt import numpy as np ## local modules ## import functions as fxn ### data structures ### ### functions ### ### data files ### zORin = open('/home/elee/Dropbox/Elizabeth_Bansal_Lab/SDI_Data/explore/Py_export/SDI_national_classifications.csv','r') zORin.readline() # rm header zOR = csv.reader(zORin, delimiter=',') allincidin = open('/home/elee/Dropbox/Elizabeth_Bansal_Lab/SDI_Data/explore/SQL_export/OR_allweeks.csv','r') allincid = csv.reader(allincidin, delimiter=',') popin = open('/home/elee/Dropbox/Elizabeth_Bansal_Lab/SDI_Data/explore/SQL_export/totalpop_age.csv', 'r') pop = csv.reader(popin, delimiter=',') ### called/local plotting parameters ### ps = fxn.pseasons sl = fxn.gp_seasonlabels fs = 24 fssml = 16 ### program ### ## import severity index ## # d_nat_classif[season] = (mean retro zOR, mean early zOR) d_nat_classif = fxn.readNationalClassifFile(zOR) ## import attack rate ## # dict_wk[week] = seasonnum, dict_incid[week] = ILI cases per 10,000 in US population in second calendar year of flu season, dict_OR[week] = OR d_wk, d_incid = fxn.week_incidCA_processing(allincid, pop) # dict_tot_attack[seasonnum] = total attack rate for weeks 40 to 20 by 100,000 _, d_tot_attack = fxn.contributions_CAO_to_attack(d_wk, d_incid) # plot values AR = [d_tot_attack[s] for s in ps] retrozOR = [d_nat_classif[s][0] for s in ps] earlyzOR = [d_nat_classif[s][1] for s in ps] print 'retro corr coef', np.corrcoef(AR, retrozOR) print 'early corr coef', np.corrcoef(AR, earlyzOR) # draw plots fig1 = plt.figure() ax1 = fig1.add_subplot(1,1,1) # mean retro zOR vs. attack rate ax1.plot(AR, retrozOR, marker = 'o', color = 'black', linestyle = 'None') for s, x, y in zip(sl, AR, retrozOR): ax1.annotate(s, xy=(x,y), xytext=(-10,5), textcoords='offset points', fontsize=fssml) ax1.set_ylabel(fxn.gp_sigma_r, fontsize=fs) ax1.set_xlabel(fxn.gp_attackrate, fontsize=fs) ax1.tick_params(axis='both', labelsize=fssml) ax1.set_ylim([-10,20]) ax1.invert_yaxis() plt.savefig('/home/elee/Dropbox/Elizabeth_Bansal_Lab/Manuscripts/Age_Severity/fluseverity_figs/Supp/zOR_totalAR.png', transparent=False, bbox_inches='tight', pad_inches=0) plt.close() # plt.show()
2,622
1,063
"""This is the solution to Problem 1 of Project Euler.""" """Copyright 2021 Dimitris Mantas""" import time def compute_all_multiples(of_number, below_number): """Compute all natural numbers, which are multiples of a natural number below a predefined number.""" # Register the list of said multiples. multiples = [] for i in range(1, below_number): if not i % of_number: multiples.append(i) return multiples # These lines are for debugging purposes. # print(compute_all_multiples(3,10)) # print(compute_all_multiples(5,10)) if __name__ == "__main__": # This line is for debugging purposes. # Start measuring the program runtime. runtime = time.time() # The resulting list is not sorted and contains the unique values the lists involved in the calculation. # This is because the multiples of 15 are contained on both said lists. ans = set([i for i in (compute_all_multiples(3, 1000) + compute_all_multiples(5, 1000))]) print(ans) print(sum(ans)) # These lines are for debugging purposes. # Compute the program runtime. print("This problem was solved in {0} seconds.".format(time.time() - runtime))
1,238
396
#!/usr/bin/env python # coding: utf-8 # # Colecciones - Listas # In[1]: l = [22, True, "una lista", [1, 2]] mi_var = l[0] # mi_var vale 22 mi_var1 = l[3] # mi_var1 vale [1, 2] print (mi_var) print (mi_var1) # In[50]: # Si queremos acceder a un elemento de una lista incluida dentro de otra lista tendremos que # utilizar dos veces este operador l = ["una lista", [1, 2]] mi_var = l[1][0] # mi_var vale 1 print (mi_var) # In[51]: # También podemos utilizar este operador para modificar un elemento de la lista si lo colocamos en la parte izquierda # de una asignación l = [22, True] l[0] = 99 # Con esto l valdrá [99, True] print (l) # In[52]: # Una curiosidad sobre el operador [] de Python es que podemos utilizar también números negativos. Si se utiliza un número # negativo como índice, esto se traduce en que el índice empieza a contar desde el final, hacia la izquierda; es decir, # con [-1] accederíamos al último elemento de la lista, con [-2] al penúltimo, con [-3], al antepenúltimo, y así # sucesivamente. l = [22, True, "una lista", [1, 2], "antepenúltimo", "penúltimo", "último"] print (l[-1]) # último print (l[-2]) # penúltimo print (l[-3]) # antepenúltimo # In[53]: # Otra cosa inusual es lo que en Python se conoce como slicing o particionado, y que consiste en ampliar este mecanismo # para permitir seleccionar porciones de la lista. Si en lugar de un número escribimos dos números inicio y fin separados # por dos puntos (inicio:fin) Python interpretará que queremos una lista que vaya desde la posición inicio a la posición # fin, sin incluir este último. Si escribimos tres números (inicio:fin:salto) en lugar de dos, el tercero se utiliza para # determinar cada cuantas posiciones añadir un elemento a la lista. l = [99, True, "una lista", [1, 2]] mi_var = l[0:2] # mi_var vale [99, True] print (mi_var) mi_var = l[0:4:2] # mi_var vale [99, "una lista"] print (mi_var) # In[54]: # Hay que mencionar así mismo que no es necesario indicar el principio y el final del slicing, sino que, si estos se # omiten, se usarán por defecto las posiciones de inicio y fin de la lista, respectivamente: l = [99, True, "una lista"] mi_var = l[1:] # mi_var vale [True, "una lista"] print (mi_var) mi_var = l[:2] # mi_var vale [99, True] print (mi_var) mi_var = l[:] # mi_var vale [99, True, "una lista"] print (mi_var) mi_var = l[::2] # mi_var vale [99, "una lista"] print (mi_var) # In[55]: # También podemos utilizar este mecanismo para modificar la lista: l = [99, True, "una lista", [1, 2]] l[0:2] = [0, 1] # l vale [0, 1, "una lista", [1, 2]] print (l) # # Colecciones - Tuplas # In[57]: # Todo lo que hemos explicado sobre las listas se aplica también a las tuplas, a excepción de la forma de definirla, # para lo que se utilizan paréntesis en lugar de corchetes. t = (1, 2, True, "python") print (t) # # Colecciones - Diccionarios # In[59]: # Los diccionarios, también llamados matrices asociativas, deben su nombre a que son colecciones que relacionan una clave y # un valor. Por ejemplo, veamos un diccionario de películas y directores: d = {"Love Actually": "Richard Curtis", "Kill Bill": "Tarantino", "Amélie": "Jean-Pierre Jeunet"} # La diferencia principal entre los diccionarios y las listas o las tuplas es que a los valores almacenados en un # diccionario se les accede no por su índice, porque de hecho no tienen orden, sino por su clave, utilizando de nuevo # el operador []. mi_var = d["Love Actually"] # devuelve "Richard Curtis" print (mi_var) # Al igual que en listas y tuplas también se puede utilizar este operador para reasignar valores. d["Kill Bill"] = "Quentin Tarantino" mi_var = d["Kill Bill"] # devuelve "Quentin Tarantino" print (mi_var)
3,731
1,395
import pytest from itertools import product import torch from torch import nn from torch.nn.utils import parameters_to_vector from torch.utils.data import DataLoader, TensorDataset from torchvision.models import wide_resnet50_2 from laplace import Laplace, SubnetLaplace, FullSubnetLaplace, DiagSubnetLaplace from laplace.baselaplace import DiagLaplace from laplace.utils import (SubnetMask, RandomSubnetMask, LargestMagnitudeSubnetMask, LargestVarianceDiagLaplaceSubnetMask, LargestVarianceSWAGSubnetMask, ParamNameSubnetMask, ModuleNameSubnetMask, LastLayerSubnetMask) torch.manual_seed(240) torch.set_default_tensor_type(torch.DoubleTensor) score_based_subnet_masks = [RandomSubnetMask, LargestMagnitudeSubnetMask, LargestVarianceDiagLaplaceSubnetMask, LargestVarianceSWAGSubnetMask] layer_subnet_masks = [ParamNameSubnetMask, ModuleNameSubnetMask, LastLayerSubnetMask] all_subnet_masks = score_based_subnet_masks + layer_subnet_masks likelihoods = ['classification', 'regression'] hessian_structures = ['full', 'diag'] @pytest.fixture def model(): model = torch.nn.Sequential(nn.Linear(3, 20), nn.Linear(20, 2)) model_params = list(model.parameters()) setattr(model, 'n_params', len(parameters_to_vector(model_params))) return model @pytest.fixture def large_model(): model = wide_resnet50_2() return model @pytest.fixture def class_loader(): X = torch.randn(10, 3) y = torch.randint(2, (10,)) return DataLoader(TensorDataset(X, y), batch_size=3) @pytest.fixture def reg_loader(): X = torch.randn(10, 3) y = torch.randn(10, 2) return DataLoader(TensorDataset(X, y), batch_size=3) @pytest.mark.parametrize('likelihood', likelihoods) def test_subnet_laplace_init(model, likelihood): # use random subnet mask for this test subnetwork_mask = RandomSubnetMask subnetmask_kwargs = dict(model=model, n_params_subnet=10) subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select() # subnet Laplace with full Hessian should work hessian_structure = 'full' lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert isinstance(lap, FullSubnetLaplace) # subnet Laplace with diagonal Hessian should work hessian_structure = 'diag' lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert isinstance(lap, DiagSubnetLaplace) # subnet Laplace without specifying subnetwork indices should raise an error with pytest.raises(TypeError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', hessian_structure=hessian_structure) # subnet Laplace with kron or lowrank Hessians should raise errors hessian_structure = 'kron' with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) hessian_structure = 'lowrank' with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) @pytest.mark.parametrize('likelihood,hessian_structure', product(likelihoods, hessian_structures)) def test_subnet_laplace_large_init(large_model, likelihood, hessian_structure): # use random subnet mask for this test subnetwork_mask = RandomSubnetMask n_param_subnet = 10 subnetmask_kwargs = dict(model=large_model, n_params_subnet=n_param_subnet) subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select() lap = Laplace(large_model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert lap.n_params_subnet == n_param_subnet if hessian_structure == 'full': assert lap.H.shape == (lap.n_params_subnet, lap.n_params_subnet) else: assert lap.H.shape == (lap.n_params_subnet,) H = lap.H.clone() lap._init_H() assert torch.allclose(H, lap.H) @pytest.mark.parametrize('likelihood,hessian_structure', product(likelihoods, hessian_structures)) def test_custom_subnetwork_indices(model, likelihood, class_loader, reg_loader, hessian_structure): loader = class_loader if likelihood == 'classification' else reg_loader # subnetwork indices that are None should raise an error subnetwork_indices = None with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are not PyTorch tensors should raise an error subnetwork_indices = [0, 5, 11, 42] with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are empty tensors should raise an error subnetwork_indices = torch.LongTensor([]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are scalar tensors should raise an error subnetwork_indices = torch.LongTensor(11) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are not 1D PyTorch tensors should raise an error subnetwork_indices = torch.LongTensor([[0, 5], [11, 42]]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are double tensors should raise an error subnetwork_indices = torch.DoubleTensor([0.0, 5.0, 11.0, 42.0]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are float tensors should raise an error subnetwork_indices = torch.FloatTensor([0.0, 5.0, 11.0, 42.0]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are half tensors should raise an error subnetwork_indices = torch.HalfTensor([0.0, 5.0, 11.0, 42.0]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are int tensors should raise an error subnetwork_indices = torch.IntTensor([0, 5, 11, 42]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are short tensors should raise an error subnetwork_indices = torch.ShortTensor([0, 5, 11, 42]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are char tensors should raise an error subnetwork_indices = torch.CharTensor([0, 5, 11, 42]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that are bool tensors should raise an error subnetwork_indices = torch.BoolTensor([0, 5, 11, 42]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that contain elements smaller than zero should raise an error subnetwork_indices = torch.LongTensor([0, -1, -11]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that contain elements larger than n_params should raise an error subnetwork_indices = torch.LongTensor([model.n_params + 1, model.n_params + 42]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # subnetwork indices that contain duplicate entries should raise an error subnetwork_indices = torch.LongTensor([0, 0, 5, 11, 11, 42]) with pytest.raises(ValueError): lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) # Non-empty, 1-dimensional torch.LongTensor with valid entries should work subnetwork_indices = torch.LongTensor([0, 5, 11, 42]) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetwork_indices, hessian_structure=hessian_structure) lap.fit(loader) assert isinstance(lap, SubnetLaplace) assert lap.n_params_subnet == 4 if hessian_structure == 'full': assert lap.H.shape == (4, 4) else: assert lap.H.shape == (4,) assert lap.backend.subnetwork_indices.equal(subnetwork_indices) @pytest.mark.parametrize('subnetwork_mask,likelihood,hessian_structure', product(score_based_subnet_masks, likelihoods, hessian_structures)) def test_score_based_subnet_masks(model, likelihood, subnetwork_mask, class_loader, reg_loader, hessian_structure): loader = class_loader if likelihood == 'classification' else reg_loader model_params = parameters_to_vector(model.parameters()) # set subnetwork mask arguments if subnetwork_mask == LargestVarianceDiagLaplaceSubnetMask: diag_laplace_model = DiagLaplace(model, likelihood) subnetmask_kwargs = dict(model=model, diag_laplace_model=diag_laplace_model) elif subnetwork_mask == LargestVarianceSWAGSubnetMask: subnetmask_kwargs = dict(model=model, likelihood=likelihood) else: subnetmask_kwargs = dict(model=model) # should raise error if we don't pass number of subnet parameters within the subnetmask_kwargs with pytest.raises(TypeError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # should raise error if we set number of subnet parameters to None subnetmask_kwargs.update(n_params_subnet=None) with pytest.raises(ValueError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # should raise error if number of subnet parameters is larger than number of model parameters subnetmask_kwargs.update(n_params_subnet=99999) with pytest.raises(ValueError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # define subnetwork mask n_params_subnet = 32 subnetmask_kwargs.update(n_params_subnet=n_params_subnet) subnetmask = subnetwork_mask(**subnetmask_kwargs) # should raise error if we try to access the subnet indices before the subnet has been selected with pytest.raises(AttributeError): subnetmask.indices # select subnet mask subnetmask.select(loader) # should raise error if we try to select the subnet again with pytest.raises(ValueError): subnetmask.select(loader) # define valid subnet Laplace model lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert isinstance(lap, SubnetLaplace) # fit Laplace model lap.fit(loader) # check some parameters assert subnetmask.indices.equal(lap.backend.subnetwork_indices) assert subnetmask.n_params_subnet == n_params_subnet assert lap.n_params_subnet == n_params_subnet assert parameters_to_vector(model.parameters()).equal(model_params) # check that Hessian and prior precision is of correct shape if hessian_structure == 'full': assert lap.H.shape == (n_params_subnet, n_params_subnet) else: assert lap.H.shape == (n_params_subnet,) assert lap.prior_precision_diag.shape == (n_params_subnet,) @pytest.mark.parametrize('subnetwork_mask,likelihood,hessian_structure', product(layer_subnet_masks, likelihoods, hessian_structures)) def test_layer_subnet_masks(model, likelihood, subnetwork_mask, class_loader, reg_loader, hessian_structure): loader = class_loader if likelihood == 'classification' else reg_loader subnetmask_kwargs = dict(model=model) # fit last-layer Laplace model lllap = Laplace(model, likelihood=likelihood, subset_of_weights='last_layer', hessian_structure=hessian_structure) lllap.fit(loader) # should raise error if we pass number of subnet parameters subnetmask_kwargs.update(n_params_subnet=32) with pytest.raises(TypeError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) subnetmask_kwargs = dict(model=model) if subnetwork_mask == ParamNameSubnetMask: # should raise error if we pass no parameter name list subnetmask_kwargs.update() with pytest.raises(TypeError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # should raise error if we pass an empty parameter name list subnetmask_kwargs.update(parameter_names=[]) with pytest.raises(ValueError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # should raise error if we pass a parameter name list with invalid parameter names subnetmask_kwargs.update(parameter_names=['123']) with pytest.raises(ValueError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # define last-layer Laplace model by parameter names and check that # Hessian is identical to that of a full LLLaplace model subnetmask_kwargs.update(parameter_names=['1.weight', '1.bias']) subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert lllap.H.equal(lap.H) # define valid parameter name subnet mask subnetmask_kwargs.update(parameter_names=['0.weight', '1.bias']) subnetmask = subnetwork_mask(**subnetmask_kwargs) # should raise error if we access number of subnet parameters before selecting the subnet n_params_subnet = 62 with pytest.raises(AttributeError): n_params_subnet = subnetmask.n_params_subnet # select subnet mask and fit Laplace model subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert isinstance(lap, SubnetLaplace) elif subnetwork_mask == ModuleNameSubnetMask: # should raise error if we pass no module name list subnetmask_kwargs.update() with pytest.raises(TypeError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # should raise error if we pass an empty module name list subnetmask_kwargs.update(module_names=[]) with pytest.raises(ValueError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # should raise error if we pass a module name list with invalid module names subnetmask_kwargs.update(module_names=['123']) with pytest.raises(ValueError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # define last-layer Laplace model by module name and check that # Hessian is identical to that of a full LLLaplace model subnetmask_kwargs.update(module_names=['1']) subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert lllap.H.equal(lap.H) # define valid parameter name subnet mask subnetmask_kwargs.update(module_names=['0']) subnetmask = subnetwork_mask(**subnetmask_kwargs) # should raise error if we access number of subnet parameters before selecting the subnet n_params_subnet = 80 with pytest.raises(AttributeError): n_params_subnet = subnetmask.n_params_subnet # select subnet mask and fit Laplace model subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert isinstance(lap, SubnetLaplace) elif subnetwork_mask == LastLayerSubnetMask: # should raise error if we pass invalid last-layer name subnetmask_kwargs.update(last_layer_name='123') with pytest.raises(KeyError): subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(loader) # define valid last-layer subnet mask (without passing the last-layer name) subnetmask_kwargs = dict(model=model) subnetmask = subnetwork_mask(**subnetmask_kwargs) # should raise error if we access number of subnet parameters before selecting the subnet with pytest.raises(AttributeError): n_params_subnet = subnetmask.n_params_subnet # select subnet mask and fit Laplace model subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert isinstance(lap, SubnetLaplace) # check that Hessian is identical to that of a full LLLaplace model assert lllap.H.equal(lap.H) # define valid last-layer subnet mask (with passing the last-layer name) subnetmask_kwargs.update(last_layer_name='1') subnetmask = subnetwork_mask(**subnetmask_kwargs) # should raise error if we access number of subnet parameters before selecting the subnet n_params_subnet = 42 with pytest.raises(AttributeError): n_params_subnet = subnetmask.n_params_subnet # select subnet mask and fit Laplace model subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert isinstance(lap, SubnetLaplace) # check that Hessian is identical to that of a full LLLaplace model assert lllap.H.equal(lap.H) # check some parameters assert subnetmask.indices.equal(lap.backend.subnetwork_indices) assert subnetmask.n_params_subnet == n_params_subnet assert lap.n_params_subnet == n_params_subnet # check that Hessian and prior precision is of correct shape if hessian_structure == 'full': assert lap.H.shape == (n_params_subnet, n_params_subnet) else: assert lap.H.shape == (n_params_subnet,) assert lap.prior_precision_diag.shape == (n_params_subnet,) @pytest.mark.parametrize('likelihood,hessian_structure', product(likelihoods, hessian_structures)) def test_full_subnet_mask(model, likelihood, class_loader, reg_loader, hessian_structure): loader = class_loader if likelihood == 'classification' else reg_loader # define full model 'subnet' mask class (i.e. where all parameters are part of the subnet) class FullSubnetMask(SubnetMask): def get_subnet_mask(self, train_loader): return torch.ones(model.n_params).byte() # define and fit valid subnet Laplace model over all weights subnetwork_mask = FullSubnetMask subnetmask = subnetwork_mask(model=model) subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) lap.fit(loader) assert isinstance(lap, SubnetLaplace) # check some parameters assert subnetmask.indices.equal(torch.tensor(list(range(model.n_params)))) assert subnetmask.n_params_subnet == model.n_params assert lap.n_params_subnet == model.n_params # check that the Hessian is identical to that of an all-weights Laplace model full_lap = Laplace(model, likelihood=likelihood, subset_of_weights='all', hessian_structure=hessian_structure) full_lap.fit(loader) assert full_lap.H.equal(lap.H) @pytest.mark.parametrize('subnetwork_mask,hessian_structure', product(all_subnet_masks, hessian_structures)) def test_regression_predictive(model, reg_loader, subnetwork_mask, hessian_structure): subnetmask_kwargs = dict(model=model) if subnetwork_mask in score_based_subnet_masks: subnetmask_kwargs.update(n_params_subnet=32) if subnetwork_mask == LargestVarianceSWAGSubnetMask: subnetmask_kwargs.update(likelihood='regression') elif subnetwork_mask == LargestVarianceDiagLaplaceSubnetMask: diag_laplace_model = DiagLaplace(model, 'regression') subnetmask_kwargs.update(diag_laplace_model=diag_laplace_model) elif subnetwork_mask == ParamNameSubnetMask: subnetmask_kwargs.update(parameter_names=['0.weight', '1.bias']) elif subnetwork_mask == ModuleNameSubnetMask: subnetmask_kwargs.update(module_names=['0']) subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(reg_loader) lap = Laplace(model, likelihood='regression', subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert isinstance(lap, SubnetLaplace) lap.fit(reg_loader) X, _ = reg_loader.dataset.tensors f = model(X) # error with pytest.raises(ValueError): lap(X, pred_type='linear') # GLM predictive f_mu, f_var = lap(X, pred_type='glm') assert torch.allclose(f_mu, f) assert f_var.shape == torch.Size([f_mu.shape[0], f_mu.shape[1], f_mu.shape[1]]) assert len(f_mu) == len(X) # NN predictive (only diagonal variance estimation) f_mu, f_var = lap(X, pred_type='nn') assert f_mu.shape == f_var.shape assert f_var.shape == torch.Size([f_mu.shape[0], f_mu.shape[1]]) assert len(f_mu) == len(X) @pytest.mark.parametrize('subnetwork_mask,hessian_structure', product(all_subnet_masks, hessian_structures)) def test_classification_predictive(model, class_loader, subnetwork_mask, hessian_structure): subnetmask_kwargs = dict(model=model) if subnetwork_mask in score_based_subnet_masks: subnetmask_kwargs.update(n_params_subnet=32) if subnetwork_mask == LargestVarianceSWAGSubnetMask: subnetmask_kwargs.update(likelihood='classification') elif subnetwork_mask == LargestVarianceDiagLaplaceSubnetMask: diag_laplace_model = DiagLaplace(model, 'classification') subnetmask_kwargs.update(diag_laplace_model=diag_laplace_model) elif subnetwork_mask == ParamNameSubnetMask: subnetmask_kwargs.update(parameter_names=['0.weight', '1.bias']) elif subnetwork_mask == ModuleNameSubnetMask: subnetmask_kwargs.update(module_names=['0']) subnetmask = subnetwork_mask(**subnetmask_kwargs) subnetmask.select(class_loader) lap = Laplace(model, likelihood='classification', subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert isinstance(lap, SubnetLaplace) lap.fit(class_loader) X, _ = class_loader.dataset.tensors f = torch.softmax(model(X), dim=-1) # error with pytest.raises(ValueError): lap(X, pred_type='linear') # GLM predictive f_pred = lap(X, pred_type='glm', link_approx='mc', n_samples=100) assert f_pred.shape == f.shape assert torch.allclose(f_pred.sum(), torch.tensor(len(f_pred), dtype=torch.double)) # sum up to 1 f_pred = lap(X, pred_type='glm', link_approx='probit') assert f_pred.shape == f.shape assert torch.allclose(f_pred.sum(), torch.tensor(len(f_pred), dtype=torch.double)) # sum up to 1 f_pred = lap(X, pred_type='glm', link_approx='bridge') assert f_pred.shape == f.shape assert torch.allclose(f_pred.sum(), torch.tensor(len(f_pred), dtype=torch.double)) # sum up to 1 # NN predictive f_pred = lap(X, pred_type='nn', n_samples=100) assert f_pred.shape == f.shape assert torch.allclose(f_pred.sum(), torch.tensor(len(f_pred), dtype=torch.double)) # sum up to 1 @pytest.mark.parametrize('subnetwork_mask,likelihood,hessian_structure', product(all_subnet_masks, likelihoods, hessian_structures)) def test_subnet_marginal_likelihood(model, subnetwork_mask, likelihood, hessian_structure, class_loader, reg_loader): subnetmask_kwargs = dict(model=model) if subnetwork_mask in score_based_subnet_masks: subnetmask_kwargs.update(n_params_subnet=32) if subnetwork_mask == LargestVarianceSWAGSubnetMask: subnetmask_kwargs.update(likelihood=likelihood) elif subnetwork_mask == LargestVarianceDiagLaplaceSubnetMask: diag_laplace_model = DiagLaplace(model, likelihood) subnetmask_kwargs.update(diag_laplace_model=diag_laplace_model) elif subnetwork_mask == ParamNameSubnetMask: subnetmask_kwargs.update(parameter_names=['0.weight', '1.bias']) elif subnetwork_mask == ModuleNameSubnetMask: subnetmask_kwargs.update(module_names=['0']) subnetmask = subnetwork_mask(**subnetmask_kwargs) loader = class_loader if likelihood == 'classification' else reg_loader subnetmask.select(loader) lap = Laplace(model, likelihood=likelihood, subset_of_weights='subnetwork', subnetwork_indices=subnetmask.indices, hessian_structure=hessian_structure) assert isinstance(lap, SubnetLaplace) lap.fit(loader) lap.log_marginal_likelihood()
28,045
8,858
""" rylog Logging happening in a 3-dimensional Cartesian product of: 1. The logging level: [debug, info, warn, error] 2. The logging category: e.g. software event, action, output 3. The detected function/method: e.g. my_class.class_method or foo """ from .misc import * from .server import * from .client import *
332
105
class ConstraintSyntaxError(SyntaxError): """A generic error indicating an improperly defined constraint.""" pass class ConstraintValueError(ValueError): """A generic error indicating a value violates a constraint.""" pass
243
63
#!/usr/bin/python # -*- coding: UTF-8 -*- import timeit import numpy as np import sys import random as rand class Queue_array: """ 顺序队列 """ def __init__(self,capacity): self._items = [None]*(capacity+1) #最后一个位置 空置 self._capacity = capacity self._head = 0 self._tail = 0 def enqueue(self,item): """ 入队 :param item: :return: """ if self._tail== self._capacity: # 队列的最后一个位置为空置,队尾指针指在此处 if self._head!=0: # 进行数据的搬移 ,在头部腾出空间,插入新的元素 self._items[0:self._tail-self._head]=self._items[self._head:self._tail] self._tail= self._tail-self._head self._head=0 else: # self._head==0 并且 self._tail== self._capacity 表示 队列已满 print('the Queue is full!') return False self._items[self._tail]=item self._tail+=1 return True def dequeue(self): """ 出队 :return: """ if self._head==self._tail: # 队列为空 print('the Queue is empty!') return None res=self._items[self._head] self._items[self._head]=None self._head += 1 return res def __repr__(self): return ','.join(self._items[self._head : self._tail]) class CircularQueue: """ 循环队列 """ def __init__(self,capacity): self._items = [None]*(capacity) self._capacity = capacity self._head = 0 self._tail = 0 def enqueue(self,item): """ 入队 循环队列 省略了 数据搬移的 开销 :param item: :return: """ if (self._tail+1) % self._capacity==self._head: # (tail+1)% n=head 表示 队列已满 print('the Queue is full!') return False self._items[self._tail]=item self._tail=(self._tail+1)%self._capacity return True def dequeue(self): """ 出队 :return: """ if self._head==self._tail: # 队列为空 print('the Queue is empty!') return None res=self._items[self._head] self._items[self._head]=None self._head = (self._head+1)%self._capacity return res class BlockingQueue: """ 阻塞队列 """ def __init__(self, capacity): self._items = [] self._capacity = capacity def producer(self,item): #TODO:多线程调用,然后给队列加锁 if len(self._items)<=self._capacity: self._items.append(item) return True else: print('the Queue is full!') return False def consumer(self): if len(self._items)>0: res=self._items.pop() return res else: print('the Queue is empty!') return None if __name__ == '__main__': # 1. 顺序队列 # queue=Queue_array(8) # string_list=['a','b','c','d','e','f','g','h'] # # for ele in string_list: # queue.enqueue(ele) # # print(queue._items) # # queue.enqueue('i') # # print('pop:',queue.dequeue()) # print('pop:', queue.dequeue()) # print('pop:', queue.dequeue()) # print(queue._items) # # queue.enqueue('i') # print(queue) #2. 循环队列 queue = CircularQueue(8) string_list=['e','f','g','h','i','j'] for ele in string_list: queue.enqueue(ele) print(queue._items) for i in range(3): print('pop:',queue.dequeue()) print(queue._items) queue.enqueue('a') queue.enqueue('b') print(queue._items) queue.enqueue('c') queue.enqueue('d') print(queue._items) queue.enqueue('e')
3,701
1,390
import numpy as np from matplotlib import pyplot as plt import math MAX_SPEED = 2 ACCELERATION = 0.5 DRAG = 0.3 TURN_SPEED=5 IMAGE = np.array([ [0,0,0,1,0,0,0], [0,0,1,1,1,0,0], [0,1,1,1,1,1,0], [1,1,1,1,1,1,1], [0,1,1,1,1,1,0], [0,0,1,1,1,0,0], [0,0,0,1,0,0,0]]) def main(): position=(42 ,42) speed=0 bearing=0 acc=0 turn=0 plt.ion() fig, ax = plt.subplots() img = np.zeros((420,420)) img[207:214,207:214]=IMAGE im = ax.imshow(img) for i in range(1000): acc+=np.random.rand()-0.5 turn+=np.random.rand()-0.5 acc=np.clip(acc,-1,1) turn=np.clip(turn,-1,1) (position,bearing,speed) = update_coords(position,bearing,speed,acc,turn) print(acc,turn) print(position) render(ax,im,position,bearing,speed) def update_coords(position,bearing,speed,acceleration,turning): (x_pos,y_pos) = position speed = update_speed(speed,acceleration) bearing = (bearing + TURN_SPEED*turning) % 360 x_pos += speed * math.sin(bearing*2*math.pi/360) y_pos += speed * math.cos(bearing*2*math.pi/360) return ((x_pos,y_pos),bearing,speed) def update_speed(speed,acceleration): speed *= DRAG speed += acceleration*ACCELERATION speed = min(speed,MAX_SPEED) if speed > 0 else max(speed,-MAX_SPEED) return speed def render(ax,im,position,bearing,speed): x_pos,y_pos = position img = np.zeros((420,420)) x = int(x_pos*5) y = int(y_pos*5) img[x:x+7,y:y+7]=IMAGE # plt.scatter(x,y) # plt.show() im.set_data(img) ax.set_title(f"bearing : {bearing}, speed: {speed}") plt.pause(0.001) plt.draw() if __name__ == "__main__": main()
1,703
798
# coding: utf-8 # 给定 n 个非负整数 a1,a2,...,an,每个数代表坐标中的一个点 (i, ai) 。 # 在坐标内画 n 条垂直线,垂直线 i 的两个端点分别为 (i, ai) 和 (i, 0)。 # 找出其中的两条线,使得它们与 x 轴共同构成的容器可以容纳最多的水。 # # 来源:力扣(LeetCode) # 链接:https://leetcode-cn.com/problems/container-with-most-water class Solution(object): # 递归做法 def maxArea1(self, height): """ :type height: List[int] :rtype: int """ n = len(height) if n == 2: return height[0] if height[0] < height[1] else height[1] maxArea = self.maxArea(height[0:n - 1]) for j, item in enumerate(height): x = n - 1 - j # x轴距离 y = min(item, height[n-1]) # y轴距离 area = x * y if (maxArea < area): maxArea = area return maxArea # 去除尾递归 def maxArea2(self, height): """ :type height: List[int] :rtype: int """ n = len(height) maxArea = 0 j = 1 while j < n: i = 0 while i < j: x = j - i # x轴距离 y = min(height[i], height[j]) # y轴距离 area = x * y if (maxArea < area): maxArea = area i = i + 1 j = j + 1 return maxArea # 双指针法 def maxArea(self, height): """ :type height: List[int] :rtype: int """ n = len(height) maxArea = 0 i = 0 j = n - 1 while i < j: x = j - i # x轴距离 y = min(height[i], height[j]) # y轴距离 area = x * y if (maxArea < area): maxArea = area if height[j] > height[i]: i = i + 1 else: j = j - 1 return maxArea list = [1,8,6,2,5,4,8,3,7] print Solution().maxArea(list)
1,868
810
# Generated by Django 2.1.3 on 2018-12-06 10:42 from django.db import migrations, models import system.models.system_settings class Migration(migrations.Migration): dependencies = [ ('system', '0002_systemsettings_contract_vot_manager_address'), ] operations = [ migrations.AlterField( model_name='systemsettings', name='contract_vot_manager_address', field=models.CharField(blank=True, max_length=42, null=True, validators=[system.models.system_settings.validate_address]), ), ]
561
179
from __future__ import division, print_function from .trajgroup import TrajectoryGroup from .hypath import HyPath from .hygroup import HyGroup def print_clusterprocedure(): """Print clustering guide.""" print(""" In ``PySPLIT`` 1. Create ``TrajectoryGroup`` with desired set of trajectories 2. ``TrajectoryGroup.make_infile()`` In ``HYSPLIT`` 3. Trajectory --> Special Runs --> Clustering --> Standard 4. Adjust clustering parameters and working folder (where output will be stored, where INFILE lives) 5. ``Run cluster analysis`` 6. Determine and set appropriate number of clusters 7. Assign trajectories to clusters (``Run``) 8. ``Display Means``, ``Quit`` In ``PySPLIT`` 9. ``spawn_clusters()``""") class Cluster(HyPath, HyGroup): """ A special :subclass: of both ``HyGroup`` and ``HyPath``. Clusters contain both trajectories and mean path information. The mean path and the trajectory composition is determined by ``HySPLIT``. """ def __init__(self, clusterdata, pathdata, datetime, clusterheader, trajectories, cluster_number): """ Initialize ``Cluster`` object. Parameters ---------- trajectories : list of ``Trajectory`` objects Trajectories that belong in the cluster. cluster_number : int The ``Cluster`` identification number. Distinguishes ``Cluster`` from other ``Clusters`` in its ``ClusterGroup`` """ HyPath.__init__(self, clusterdata, pathdata, datetime, clusterheader) HyGroup.__init__(self, trajectories) self.start_longitude = self.trajectories[0].data.loc[0, 'geometry'].x self.clusternumber = cluster_number self.multitraj = False def __getitem__(self, index): """ Get ``Trajectory`` or ``TrajectoryGroup``. Parameters ---------- index : int or slice Returns ------- ``Trajectory`` or ``TrajectoryGroup`` depending if indexed or sliced. Won't return a ``Cluster`` because those are specially defined. """ newthing = self.trajectories[index] if isinstance(newthing, list): newthing = TrajectoryGroup(newthing) return newthing def __add__(self, other): """ Add a ``HyGroup`` to this ``Cluster`` instance. Parameters ---------- other : ``HyGroup`` Another ``TrajectoryGroup`` or ``Cluster``. May or may not contain some of the same ``Trajectory`` instances. Returns ------- A new ``TrajectoryGroup`` containing the union of the sets of ``Trajectory`` instances. """ return TrajectoryGroup(HyGroup.__add__(self, other)) def __sub__(self, other): """ Subtract a ``HyGroup`` from this ``Cluster`` instance. Parameters ---------- other : ``HyGroup`` Another ``Cluster`` or ``TrajectoryGroup`` Returns ------- A new ``TrajectoryGroup`` containing the set difference betwee the sets of ``Trajectory`` instances. """ return TrajectoryGroup(HyGroup.__sub__(self, other)) class ClusterGroup(object): """ Group of ``Cluster`` instances. Contains all the ``Cluster``s produced in one ``HYSPLIT`` cluster analysis. """ def __init__(self, clusters): """ Initialize ``ClusterGroup`` object. Parameters ---------- clusters : list of ``Cluster`` instances ``Cluster`` instances from the same HYSPLIT clustering run. """ self.clusters = clusters self.clustercount = len(clusters) self.trajcount = sum([c.trajcount for c in self.clusters]) def __getitem__(self, index): """ Get ``Cluster`` or ``ClusterGroup``. Index or slice ``self.clusters`` to get a ``Cluster`` or ``ClusterGroup``, respectively. """ newthing = self.clusters[index] try: newthing = ClusterGroup(newthing) except: pass return newthing
4,338
1,199
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import argparse import os import re import requests import sys sys.path.append(os.path.join(sys.path[0], "../", "lib")) import lkft_squad_client # noqa: E402 def get_branch_from_make_kernelversion(make_kernelversion): """ IN: "4.4.118" OUT: "4.4" IN: "4.9.118-rc1" OUT: "4.9" """ pattern = re.compile(r"^(\d+\.\d+).*$") match = pattern.match(make_kernelversion) return match.group(1) def get_most_recent_release(builds_url): """ Given a list of builds that is sorted with the newest first, return the most recent finished build. """ first_build = None for build in lkft_squad_client.Builds(builds_url): if not first_build: first_build = build if build["finished"]: return build # If none found, return first build return first_build def get_build_report(build_url): build = lkft_squad_client.Build(build_url) baseline_branch = get_branch_from_make_kernelversion( build.build_metadata["make_kernelversion"] ) # Get baseline baseline_project_url = lkft_squad_client.get_projects_by_branch()[baseline_branch] baseline_builds_url = baseline_project_url + "builds" baseline_build = get_most_recent_release(baseline_builds_url) template_url = build_url + "email" parameters = {"baseline": baseline_build["id"], "template": "9"} result = requests.get(template_url, parameters) email = build.build_metadata.get("email-notification", "") if "No regressions" in result.text: subject = "{}: no regressions found".format(build.build["version"]) else: subject = "{}: regressions detected".format(build.build["version"]) return (email, subject, result.text) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("build_url", help="API URL to developer build") args = parser.parse_args() (email_destination, email_subject, email_body) = get_build_report(args.build_url) with open("email.to", "w") as f: f.write(email_destination) with open("email.subject", "w") as f: f.write(email_subject) with open("email.body", "w") as f: f.write(email_body) print("TO: {}".format(email_destination)) print("SUBJECT: {}".format(email_subject)) print("\n{}\n".format(email_body))
2,424
817
from math import ceil, sqrt def part_one(input): circle_index = get_circle_index(input) circle_zero = get_circle_zero(circle_index) cardinal_points = get_cardinal_points(circle_index, circle_zero) distance_to_closest_cardinal_point = compute_distance_to_closest_cardinal_point(input, cardinal_points) return circle_index + distance_to_closest_cardinal_point def get_circle_index(input): return ceil(sqrt(input)) // 2 def get_circle_zero(circle_index): return pow(circle_index * 2 - 1, 2) def get_cardinal_points(circle_index, circle_zero): return [circle_zero + x * circle_index for x in [1, 3, 5, 7]] def compute_distance_to_closest_cardinal_point(input, cardinal_points): return min([abs(input - x) for x in cardinal_points]) def part_two(input): spiral = {} x = 0 y = 0 spiral[(0, 0)] = 1 while spiral[(x, y)] < input: x, y = get_next_coordinates(x, y) coordinates_offsets = [-1, 0, 1] spiral[(x, y)] = sum([spiral.get((x + i, y + j), 0) for i in coordinates_offsets for j in coordinates_offsets]) return spiral[(x, y)] def get_next_coordinates(x, y): if x == y == 0: return (1, 0) if y > -x and x > y: return (x, y + 1) if y > -x and y >= x: return (x - 1, y) if y <= -x and x < y: return (x, y - 1) if y <= -x and x >= y: return (x + 1, y) if __name__ == "__main__": input = 325489 print(part_one(input)) print(part_two(input))
1,507
581
#022: Crie um programa que leia o nome completo de uma pessoa e mostre: # - O nome com todas as letras maiúsculas e minúsculas. # - Quantas letras ao tdo (sem considerar espaços). # - Quantas letras tem o primeiro nome. nome = input("Qual é o seu nome? ") print(">>",nome.upper()) print(">>",nome.lower()) jnome = nome.strip() v = jnome.count(" ") print(">>",len(jnome)- v) pnome = nome.split() print(">>",len(pnome[0]))
423
156
from typing import Optional from pysqldump.domain.formatters import ( CSVFormatter, DictFormatter, JsonFormatter, ConsoleFormatter, ) from pysqldump.settings.base import get_config config = get_config() class File: def __init__(self, filename): self.filename = filename def get_extension(self): try: return self.filename.split(".")[1] except (IndexError, AttributeError): return None def get_filename(self): if self.filename is None: return "" return self.filename class OutputManager: formats = {"csv": "csv", "json": "json", "console": "console"} def __init__(self, data: tuple, headers: list, export_to: Optional[str] = None): self._filename = File(filename=export_to) self.headers = headers self.data = data @property def filename(self): return self._filename.get_filename() @property def formatter(self): extension = self._filename.get_extension() return self.formats.get(extension, "console") def run(self, pprint: bool = False, json: bool = False): if self.formatter == "csv": return self.__to_csv(pprint=pprint) elif self.formatter == "console" and json or self.formatter == "json": return self.__to_json(pprint=pprint, json=json) elif self.formatter == "console": return self.__to_console(pprint=pprint) def __to_console(self, pprint: bool = False): if pprint: return ConsoleFormatter( headers=self.headers, data=self.data, export_to=self.filename ).print() return DictFormatter( headers=self.headers, data=self.data, export_to=self.filename ).export() def __to_csv(self, pprint: bool = False): formatter = CSVFormatter( headers=self.headers, data=self.data, export_to=self.filename ) if pprint: return formatter.print() return formatter.export() def __to_json(self, pprint: bool = False, json: bool = False): formatter = JsonFormatter( headers=self.headers, data=self.data, export_to=self.filename ) if pprint: formatter.print() if self._filename and not json: return formatter.export() return formatter.use()
2,395
684
"""**Mortgage Calculator** - Calculate the monthly payments of a fixed term mortgage over given Nth terms at a given interest rate. Also figure out how long it will take the user to pay back the loan.""" months = int(input("Enter mortgage term (in months): ")) rate = float(input("Enter interest rate (in %): ")) loan = float(input("Enter loan value: ")) monthly_rate = rate / 100 / 12 payment = (monthly_rate / (1 - (1 + monthly_rate)**(-months))) * loan print("Monthly payment for a $%.2f %s year mortgage at %.2f%% interest rate is: $%.2f" % (loan, (months / 12), rate, payment))
605
218
import numpy as np def div(a, b): if b == 0: print("denominator iz zero!!!") return np.inf else: return a/b def add(a,b): return (a+b)
172
66
import cv2 class StereoSGBMConfig: min_disparity = 0 num_disparities = 16*10 sad_window_size = 3 uniqueness_ratio = 5 p1 = 16*sad_window_size*sad_window_size p2 = 96*sad_window_size*sad_window_size pre_filter_cap = 63 speckle_window_size = 0 speckle_range = 0 disp_max_diff = 1 mode = cv2.STEREO_SGBM_MODE_SGBM class StereoSGBMConfig2: pre_filter_cap = 63 sad_window_size = 3 p1 = sad_window_size * sad_window_size * 4 p2 = sad_window_size * sad_window_size * 32 min_disparity = 0 num_disparities = 128 uniqueness_ratio = 10 speckle_window_size = 100 speckle_range = 32 disp_max_diff = 1 full_dp = 1 mode = cv2.STEREO_SGBM_MODE_SGBM_3WAY class MatcherConfig: ransac = { 'max_iterations': 5, 'error_threshold': 50, 'min_consensus': 5 } hough = { 'dxbin': 100, 'dangbin': 50, 'umbralvotos': 10 }
956
435
from brnolm.runtime.tensor_reorganization import TensorReorganizer import torch from torch.autograd import Variable from .common import TestCase class Dummy_lstm(): def __init__(self, nb_hidden): self._nb_hidden = nb_hidden def init_hidden(self, batch_size): return ( torch.FloatTensor([[[0.0] * self._nb_hidden] * batch_size] * 2), torch.FloatTensor([[[0.0] * self._nb_hidden] * batch_size] * 2) ) class TensorReorganizerTests(TestCase): def setUp(self): self.lm = Dummy_lstm(nb_hidden=2) self.reorganizer = TensorReorganizer(self.lm.init_hidden) def test_passing(self): last_h = ( torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [2, 2], [3, 3]]]*2), ) mask = torch.LongTensor([0, 1, 2]) bsz = 3 new_h = self.reorganizer(last_h, mask, bsz) self.assertEqual(new_h, last_h) def test_shrinks(self): last_h = ( torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [2, 2], [3, 3]]]*2), ) mask = torch.LongTensor([0, 2]) bsz = 2 new_h = self.reorganizer(last_h, mask, bsz) expected = ( torch.FloatTensor([[[0.1, 0.1], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [3, 3]]]*2), ) self.assertEqual(new_h, expected) def test_requires_bsz_greater_than_mask(self): last_h = ( torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [2, 2], [3, 3]]]*2), ) mask = torch.LongTensor([0, 1, 2]) bsz = 2 self.assertRaises(ValueError, self.reorganizer, last_h, mask, bsz) def test_on_empty_mask_zeros(self): last_h = ( torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [2, 2], [3, 3]]]*2), ) mask = torch.LongTensor([]) bsz = 2 new_h = self.reorganizer(last_h, mask, bsz) expected = self.lm.init_hidden(bsz) self.assertEqual(new_h, expected) def test_completion_by_zeros(self): last_h = ( torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [2, 2], [3, 3]]]*2), ) mask = torch.LongTensor([1]) bsz = 2 new_h = self.reorganizer(last_h, mask, bsz) expected = ( torch.FloatTensor([[[0.2, 0.2], [0.0, 0.0]]]*2), torch.FloatTensor([[[2.0, 2.0], [0.0, 0.0]]]*2), ) self.assertEqual(new_h, expected) def test_bug_regression_single_addition(self): last_h = ( torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]*2), torch.FloatTensor([[[1, 1], [2, 2], [3, 3]]]*2), ) mask = torch.LongTensor([1, 2]) bsz = 3 new_h = self.reorganizer(last_h, mask, bsz) expected = ( torch.FloatTensor([[[0.2, 0.2], [0.3, 0.3], [0.0, 0.0]]]*2), torch.FloatTensor([[[2.0, 2.0], [3.0, 3.0], [0.0, 0.0]]]*2), ) self.assertEqual(new_h, expected) class Dummy_srn(): def __init__(self, nb_hidden): self._nb_hidden = nb_hidden self._nb_layers = 1 def init_hidden(self, batch_size): return torch.FloatTensor(self._nb_layers, batch_size, self._nb_hidden).zero_() class TensorReorganizerTests_SRN(TestCase): def setUp(self): lm = Dummy_srn(nb_hidden=2) self.reorganizer = TensorReorganizer(lm.init_hidden) def test_passing(self): last_h = torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]) mask = torch.LongTensor([0, 1, 2]) bsz = 3 new_h = self.reorganizer(last_h, mask, bsz) self.assertEqual(new_h, last_h) def test_passing_variables(self): last_h = Variable(torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]])) mask = Variable(torch.LongTensor([0, 1, 2])) bsz = 3 new_h = self.reorganizer(last_h, mask, bsz) self.assertEqual(new_h, last_h) def test_shrinks(self): last_h = torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]) mask = torch.LongTensor([0, 2]) bsz = 2 new_h = self.reorganizer(last_h, mask, bsz) expected = torch.FloatTensor([[[0.1, 0.1], [0.3, 0.3]]]) self.assertEqual(new_h, expected) def test_requires_bsz_greater_than_mask(self): last_h = torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]) mask = torch.LongTensor([0, 1, 2]) bsz = 2 self.assertRaises(ValueError, self.reorganizer, last_h, mask, bsz) def test_on_empty_mask_zeros(self): last_h = torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]) mask = torch.LongTensor([]) bsz = 2 new_h = self.reorganizer(last_h, mask, bsz) expected = torch.FloatTensor([[[0.0, 0.0], [0.0, 0.0]]]) self.assertEqual(new_h, expected) def test_completion_by_zeros(self): last_h = torch.FloatTensor([[[0.1, 0.1], [0.2, 0.2], [0.3, 0.3]]]) mask = torch.LongTensor([1]) bsz = 2 new_h = self.reorganizer(last_h, mask, bsz) expected = torch.FloatTensor([[[0.2, 0.2], [0.0, 0.0]]]) self.assertEqual(new_h, expected)
5,460
2,359
jobname="BachelorThesis"
25
11
from eventsourcing.infrastructure.event_sourced_repo import EventSourcedRepository from quantdsl.domain.model.contract_specification import ContractSpecification, ContractSpecificationRepository class ContractSpecificationRepo(ContractSpecificationRepository, EventSourcedRepository): domain_class = ContractSpecification
331
85
from requests_html import HTMLSession import os import sys writeFileName = "courseLinks.out" writeFileStream = open(writeFileName,'w',encoding='utf-8') session = HTMLSession() url = 'https://www.ji.sjtu.edu.cn/academics/courses/courses-by-number/' r = session.get(url) for i in range(2,100): sel = '#Faculty-information > li:nth-child(' + str(i) + ') > a' # print(sel) results = r.html.find(sel) if len(results) == 0: break; else: for result in results: writeFileStream.write(result.absolute_links.pop()+'\n') writeFileStream.close() # #Faculty-information > li:nth-child(3) > a
660
247
import logging import discord from datetime import datetime from discord.ext import tasks, commands from discord.ext.commands import Cog from cogs.utils.utils import json_io_dump, json_io_load log = logging.getLogger(__name__) STATUS = 'status' TIME_STARTED = 'time_started' NAME = 'name' GAMES = 'games' NONE = 'none' # Reference JSON # { # "player_id1": { # "status": "a string of status", # "time_started": "time_started_current_status", # "games":{ # "COD MW2": "time_played", # "Poop": "time_played" # } # }, # "player_id2": { # "status": "a string of status", # "time_started": "time_started_current_status", # "games":{ # "COD MW2": "time_played", # "Poop": "time_played" # } # } # } class TimePlayed(Cog): """ Tracks your time played for each status you have had """ def __init__(self, bot): self.bot = bot self.log = logging.getLogger() self.gametime_file = 'gametime.json' self.gametime = None self.update_time.start() async def game_load(self): """ Loads games from JSON """ self.gametime = json_io_load(self.gametime_file) async def game_dump(self): """ Dumps games to JSON """ if not json_io_dump(self.gametime_file, self.gametime): self.log.critical('Unable to dump JSON file for TimePlayed!') def calculate_addition(self, time_started): """ Returns whether to add 2 minutes (in seconds) or something less than that Time_started is a datetime.datetime string """ converted_time = datetime.strptime(time_started, '%Y-%m-%d %H:%M:%S') delta = (datetime.utcnow().replace(microsecond=0) - converted_time).total_seconds() return int(delta) if delta < 120 else 120 async def get_current_gametime(self): """ Returns the dictionary of the current players and what they are playing """ current_gametime = dict() for member in set(self.bot.get_all_members()): # Initialize the dictionary for this member and set everything to None current_gametime[str(member.id)] = dict() current_gametime[str(member.id)][NAME] = member.name current_gametime[str(member.id)][STATUS] = NONE current_gametime[str(member.id)][TIME_STARTED] = NONE current_gametime[str(member.id)][GAMES] = dict() # If the member is not doing anything, continue if not member.activities: continue # If the member is playing something, then take note of this for activity in member.activities: if activity.type == discord.ActivityType.playing: # If for some reason this is not None, then we have 2 gaming activities on this member if current_gametime[str(member.id)][STATUS] != NONE: self.log.critical('There are multiple games playing right now in Gametime for single member!') self.log.critical('{} had status {} instead of none as expected.'.format( current_gametime[str(member.id)][NAME], current_gametime[str(member.id)][STATUS])) current_gametime[str(member.id)][STATUS] = activity.name date = member.activity.start.replace(microsecond=0) if member.activity.start else datetime.utcnow().replace(microsecond=0) current_gametime[str(member.id)][TIME_STARTED] = str(date) current_gametime[str(member.id)][GAMES][activity.name] = 0 return current_gametime async def compare_and_update(self, current_gametime): """ Compares and updates the playing list """ for id in current_gametime: if id not in self.gametime: self.gametime[id] = current_gametime[id] current_status = current_gametime[id][STATUS] # If the current gametime is not None, then update the time on the game currently played if current_status != NONE: if current_status not in self.gametime[id][GAMES]: self.gametime[id][GAMES][current_status] = 0 result = self.calculate_addition(current_gametime[id][TIME_STARTED]) self.gametime[id][GAMES][current_status] += result # If the current game is different from last game, add 2 minutes to the last game if current_status != self.gametime[id][STATUS] and self.gametime[id][STATUS] != NONE: self.gametime[id][GAMES][self.gametime[id][STATUS]] += 120 # Update the game status regardless of what's going on self.gametime[id][STATUS] = current_gametime[id][STATUS] self.gametime[id][TIME_STARTED] = current_gametime[id][TIME_STARTED] def calculate_days_minutes_seconds(self, seconds): """ Returns the days hours minutes seconds from seconds """ # years, seconds = seconds // 31556952, seconds % 31556952 # months, seconds = seconds // 2629746, seconds % 2629746 days, seconds = seconds // 86400, seconds % 86400 hours, seconds = seconds // 3600, seconds % 3600 minutes, seconds = seconds // 60, seconds % 60 msg = '{:02d} Days, {:02d} Hours, {:02d} Minutes, {:02d} Seconds'.format(days, hours, minutes, seconds) if days > 9000: msg += ' ITS OVER 9000!' if days == 69: msg += ' Hah, nice... 69' return msg @commands.command() async def played(self, ctx, *, member: discord.Member = None): """Returns the amount of time played for every game If Member is not specified, then returns the played information for member that sent command """ if member is None: member = ctx.author if str(member.id) not in self.gametime: await ctx.send('ERROR!: Unable to find {} in gametime list... looks like a bug'.format(member.mention)) msg = 'Time played for {}\n'.format(member.mention) if not self.gametime[str(member.id)][GAMES]: msg += '`Looks like {} hasn\'t played any games!`'.format(member.display_name) for game in self.gametime[str(member.id)][GAMES]: msg += '`{:<30}: {}`\n'.format(game, self.calculate_days_minutes_seconds(self.gametime[str(member.id)][GAMES][game])) await ctx.send('{}'.format(msg)) @tasks.loop(minutes=2) async def update_time(self): """ Loop that updates the time played of the current game for each member Steps: 1. Load list 2. Get Current List of people playing 3. Compare new with old list of people playing and update old gametime list as needed 4. Write list """ self.log.debug('Starting gametime save loop') await self.game_load() current_gametime = await self.get_current_gametime() await self.compare_and_update(current_gametime) await self.game_dump() self.log.debug('End gametime save loop') @update_time.before_loop async def before_update_time(self): """ We want to wait until the bot is ready before going into the loop """ await self.bot.wait_until_ready() @update_time.after_loop async def after_update_time(self): """ If anything is happening after the loop, we want to store all the information before any exits """ await self.game_dump() def setup(bot): bot.add_cog(TimePlayed(bot))
7,655
2,313
import numpy as np from .converter import TransformConverter from .. import linear class PyqtgraphTransformConverter(TransformConverter): name = 'pyqtgraph' def __init__(self): try: import pyqtgraph self._import_error = None except ImportError as exc: self._import_error = str(exc) return self._to_classes = { linear.STTransform: self._STTransform_to_pg, linear.AffineTransform: self._AffineTransform_to_pg, } self._from_classes = { # pyqtgraph.SRTTransform: self._from_SRTTransform, # pyqtgraph.SRTTransform3D: self._from_SRTTransform, pyqtgraph.QtGui.QTransform: self._from_QTransform, pyqtgraph.QtGui.QMatrix4x4: self._from_QMatrix4x4, pyqtgraph.Transform3D: self._from_QMatrix4x4, } def _STTransform_to_pg(self, tr): import pyqtgraph if tr.dims == (2, 2): ptr = pyqtgraph.SRTTransform() ptr.setScale(tr.scale) ptr.setTranslate(tr.offset) return ptr elif tr.dims == (3, 3): ptr = pyqtgraph.SRTTransform3D() ptr.setScale(tr.scale) ptr.setTranslate(tr.offset) return ptr else: raise TypeError("Converting STTransform of dimension %r to pyqtgraph is not supported." % tr.dims) def _AffineTransform_to_pg(self, tr): import pyqtgraph if tr.dims == (2, 2): m = tr.matrix o = tr.offset ptr = pyqtgraph.QtGui.QTransform(m[0,0], m[1,0], 0.0, m[0,1], m[1,1], 0.0, o[0], o[1], 1.0) return ptr elif tr.dims == (3, 3): m = np.eye(4) m[:3, :3] = tr.matrix m[:3, 3] = tr.offset ptr = pyqtgraph.Transform3D(m) return ptr else: raise TypeError("Converting AffineTransform of dimension %r to pyqtgraph is not supported." % tr.dims) def _from_SRTTransform(self, tr): return linear.STTransform(offset=tr.getTranslation(), scale=tr.getScale()) def _from_QTransform(self, tr): m = np.array([ [tr.m11(), tr.m21()], [tr.m12(), tr.m22()], ]) o = np.array([tr.m31(), tr.m32()]) return linear.AffineTransform(matrix=m, offset=o) def _from_QMatrix4x4(self, tr): m = np.array(tr.copyDataTo()).reshape(4,4) return linear.AffineTransform(matrix=m[:3, :3], offset=m[:3, 3])
2,578
842
from pyqt.utils import time
28
9
class BadIngredient(Exception): """ Something is wrong with an ingredient """ class BadRecipe(Exception): """ Something is wrong with a recipe """ class InvalidColumnError(Exception): def __init__(self, *args, **kwargs): self.column_name = kwargs.pop("column_name", None) if not args: # default exception message args = ['Invalid column "{}"'.format(self.column_name)] super(InvalidColumnError, self).__init__(*args, **kwargs)
491
136
import os def readAnn(textfolder="../data/SemEval2017Task10/"): ''' Read .ann files and look up corresponding spans in .txt files Args: textfolder: ''' flist = os.listdir(textfolder) for f in flist: if not f.endswith(".ann"): continue f_anno = open(os.path.join(textfolder, f), "rU") f_text = open(os.path.join(textfolder, f.replace(".ann", ".txt")), "rU") # there's only one line, as each .ann file is one text paragraph for l in f_text: text = l #@TODO: collect all keyphrase and relation annotations, create pairs of all keyphrase that appear in same sentence for USchema style RE for l in f_anno: anno_inst = l.strip().split("\t") if len(anno_inst) == 3: keytype, start, end = anno_inst[1].split(" ") if not keytype.endswith("-of"): # look up span in text and print error message if it doesn't match the .ann span text keyphr_text_lookup = text[int(start):int(end)] keyphr_ann = anno_inst[2] if keyphr_text_lookup != keyphr_ann: print("Spans don't match for anno " + l.strip() + " in file " + f) #if keytype.endswith("-of"): if __name__ == '__main__': readAnn()
1,373
429
from collections import namedtuple import os import sympy import numpy as np from means.core.model import Model _Reaction = namedtuple('_REACTION', ['id', 'reactants', 'products', 'propensity', 'parameters']) def _sbml_like_piecewise(*args): if len(args) % 2 == 1: # Add a final True element you can skip in SBML args += (True,) sympy_args = [] for i in range(len(args)/2): # We need to group args into tuples of form # (value, condition) # SBML usually outputs them in form (value, condition, value, condition, value ...) sympy_args.append((args[i*2], args[i*2+1])) return sympy.Piecewise(*sympy_args) def _sympify_kinetic_law_formula(formula): # We need to define some namespace hints for sympy to deal with certain functions in SBML formulae # For instance, `eq` in formula should map to `sympy.Eq` namespace = {'eq': sympy.Eq, 'neq': sympy.Ne, 'floor': sympy.floor, 'ceiling': sympy.ceiling, 'gt': sympy.Gt, 'lt': sympy.Lt, 'geq': sympy.Ge, 'leq': sympy.Le, 'pow': sympy.Pow, 'piecewise': _sbml_like_piecewise} return sympy.sympify(formula, locals=namespace) def _parse_reaction(libsbml_reaction): id_ = libsbml_reaction.getId() reactants = {sympy.Symbol(r.getSpecies()): r.getStoichiometry() for r in libsbml_reaction.getListOfReactants()} products = {sympy.Symbol(p.getSpecies()): p.getStoichiometry() for p in libsbml_reaction.getListOfProducts()} kinetic_law = _sympify_kinetic_law_formula(libsbml_reaction.getKineticLaw().getFormula()) # This would only work for SBML Level 3, prior levels do not have parameters within kinetic law parameters = [(sympy.Symbol(p.getId()), p.getValue()) for p in libsbml_reaction.getKineticLaw().getListOfParameters()] return _Reaction(id_, reactants, products, kinetic_law, parameters) def read_sbml(filename): """ Read the model from a SBML file. :param filename: SBML filename to read the model from :return: A tuple, consisting of :class:`~means.core.model.Model` instance, set of parameter values, and set of initial conditions variables. """ import libsbml if not os.path.exists(filename): raise IOError('File {0!r} does not exist'.format(filename)) reader = libsbml.SBMLReader() document = reader.readSBML(filename) sbml_model = document.getModel() if not sbml_model: raise ValueError('Cannot parse SBML model from {0!r}'.format(filename)) species = sympy.symbols([s.getId() for s in sbml_model.getListOfSpecies()]) initial_conditions = [s.getInitialConcentration() for s in sbml_model.getListOfSpecies()] compartments = sympy.symbols([s.getId() for s in sbml_model.getListOfCompartments()]) compartment_sizes = [s.getSize() for s in sbml_model.getListOfCompartments()] reactions = map(_parse_reaction, sbml_model.getListOfReactions()) # getListOfParameters is an attribute of the model for SBML Level 1&2 parameters_with_values = [(sympy.Symbol(p.getId()), p.getValue()) for p in sbml_model.getListOfParameters()] parameter_values = dict(parameters_with_values) parameters = map(lambda x: x[0], parameters_with_values) if not parameters: track_local_parameters = True parameters = set() parameter_values = {} else: track_local_parameters = False stoichiometry_matrix = np.zeros((len(species), len(reactions)), dtype=int) propensities = [] for reaction_index, reaction in enumerate(reactions): if track_local_parameters: for param, value in reaction.parameters: parameters.add(param) parameter_values[param] = value reactants = reaction.reactants products = reaction.products propensities.append(reaction.propensity) for species_index, species_id in enumerate(species): net_stoichiometry = products.get(species_id, 0) - reactants.get(species_id, 0) stoichiometry_matrix[species_index, reaction_index] = net_stoichiometry if track_local_parameters: # sympy does not allow sorting its parameter lists by default, # explicitly tell to sort by str representation sorted_parameters = sorted(parameters, key=str) else: sorted_parameters = parameters parameter_values_list = [parameter_values[p] for p in sorted_parameters] # We need to concatenate compartment names and parameters as in our framework we cannot differentiate the two compartments_and_parameters = compartments + sorted_parameters parameter_values_list = compartment_sizes + parameter_values_list model = Model(species, compartments_and_parameters, propensities, stoichiometry_matrix) return model, parameter_values_list, initial_conditions
5,018
1,507
import pytest import clrenv @pytest.fixture(autouse=True) def clear_overlay_path(monkeypatch): monkeypatch.setenv("CLRENV_OVERLAY_PATH", "") def test_custom_base(tmp_path, monkeypatch): custom_path = tmp_path / "custom/path" custom_path.parent.mkdir() custom_path.write_text("data") monkeypatch.setenv("CLRENV_PATH", str(custom_path)) assert clrenv.path.environment_paths() == (custom_path,) def test_missing_base(tmp_path, monkeypatch): monkeypatch.setenv("CLRENV_PATH", str(tmp_path / "aaa")) with pytest.raises(ValueError): clrenv.path.environment_paths() def test_overlay(tmp_path, monkeypatch): env_path = tmp_path / "env" monkeypatch.setenv("CLRENV_PATH", str(env_path)) env_path.write_text("") overlay_path1 = tmp_path / "overlay1" overlay_path2 = tmp_path / "overlay2" overlay_path1.write_text("data") overlay_path2.write_text("data2") monkeypatch.setenv("CLRENV_OVERLAY_PATH", f"{overlay_path1}:{overlay_path2}") assert clrenv.path.environment_paths() == ( overlay_path1, overlay_path2, env_path, )
1,123
428
import random import pandas as pd import numpy as np import json from tqdm import * def split(full_list, shuffle=False, ratio=0.2): n_total = len(full_list) offset = int(n_total * ratio) if n_total == 0 or offset < 1: return [], full_list if shuffle: random.shuffle(full_list) sublist_1 = full_list[:offset] sublist_2 = full_list[offset:2 * offset] sublist_3 = full_list[2 * offset:] return sublist_1, sublist_2, sublist_3 def convert_to_codah_statement(input_file: str, output_file1: str): print(f'converting {input_file} to entailment dataset...') tsv_file = pd.read_csv(input_file) qa_list = tsv_file.to_numpy() nrow = sum(1 for _ in qa_list) id = 0 with open(output_file1, 'w') as output_handle1: # print("Writing to {} from {}".format(output_file, qa_file)) for sample in tqdm(qa_list, total=nrow): output_dict = convert_sample_to_entailment(sample, id) output_handle1.write(json.dumps(output_dict)) output_handle1.write("\n") id += 1 print(f'converted statements saved to {output_file1}') print() # Convert the QA file json to output dictionary containing premise and hypothesis def convert_sample_to_entailment(sample: list, id: int): question_text = sample[1] choices = sample[3:7] # left close right open single_qa_dict = {'id': id, 'question': {'stem': sample[1]}, 'answer_key': 0} choice_list = [] choice_count = 0 for choice in choices: statement = question_text + ' ' + choice create_output_dict(single_qa_dict, statement, choice_count == 0) choice_list.append({'text': choice, 'label': choice_count}) choice_count += 1 single_qa_dict['question']['choices'] = choice_list return single_qa_dict # Create the output json dictionary from the input json, premise and hypothesis statement def create_output_dict(input_json: dict, statement: str, label: bool) -> dict: if "statements" not in input_json: input_json["statements"] = [] input_json["statements"].append({"label": label, "statement": statement}) return input_json if __name__ == "__main__": convert_to_codah_statement('../data/codah/fold_0/train.csv', './data/codah/fold_0/train.jsonl') # train, dev, test = split(full_list, shuffle=True, ratio=0.2) # convert_to_codah_statement(train, 'train.statement.jsonl') # convert_to_codah_statement(dev, 'train.statement.jsonl') # convert_to_codah_statement(test, 'train.statement.jsonl') print('Hey, there!')
2,583
867
''' pip install Scrapy pip install selenium In a folder: scrapy startproject imgt when running: scrapy crawl new_imgt -o out.json when using scrapy shell: scrapy shell 'url' in Ipython, you can use response.xpath or response.css to try out object: 1. selectorlist if css('a') and there are a lot of 'a' 2. selector it will have css and xpath method 3. reponse conda activate selenium remember make change to the python scirpt under spider folder ''' ''' If encounter robot blockage error: open setting.py and change the robot setting to False you can specify hla in __init__, and then when call: scrapy crawl new_imgt -a hla="HLA-A*0101" -o out.json When encounter dynamic page, use selenium to get the page and pass it to scrapy response object Double check using both 'inspect' and 'see source code' in a webpage, they can be different ''' ''' cat inventory_compliant.txt | while read line; do scrapy crawl new_imgt -a hla="$line" -o "./hla_paratope/$line.json"; done ''' import scrapy from scrapy.crawler import CrawlerProcess from selenium import webdriver from selenium.webdriver.support.ui import WebDriverWait class imgtSpider(scrapy.Spider): name = 'new_imgt' start_urls = ['http://www.imgt.org/3Dstructure-DB/'] def __init__(self,hla): self.hla = hla path_to_chromedriver = '/Users/ligk2e/Downloads/chromedriver' self.driver = webdriver.Chrome(executable_path=path_to_chromedriver) self.driver.implicitly_wait(5) def get_selenium(self,url): self.driver.get(url) self.driver.find_element_by_xpath('//*[@id="species"]/option[27]').click() # choose Home Sapien (select drop down) self.driver.find_element_by_xpath('//*[@id="radio_pMH1"]').click() # choose pMHCI (input) self.driver.find_element_by_xpath('//*[@id="datas"]/p[2]/input[1]').click() # click submit (button) return self.driver.page_source.encode('utf-8') def parse(self,response): # for parsing 550 entry page response = scrapy.Selector(text=self.get_selenium(imgtSpider.start_urls[0])) for row in response.css('body#result div#data table.Results tbody tr')[1:]: #[Selector,Selector,Selector...] # don't need header mhc = row.css('td')[2].css('td::text').get() if self.hla in mhc: url_suffix = row.css('td')[1].css('a::attr(href)').get() # details.cgi?pdbcode=2CLR # what we need is: http://www.imgt.org/3Dstructure-DB/cgi/details.cgi?pdbcode=2CLR&Part=Epitope url_next = 'http://www.imgt.org/3Dstructure-DB/cgi/' + url_suffix + '&Part=Epitope' yield scrapy.Request(url_next,callback=self.parse_paratope) def parse_paratope(self,response): url_next = response.url paratope = '' for i in response.css('body#result div#mybody div#main table')[0].css('tr')[2].css('td')[1].css('span a'): aa = i.css('a::text').get() paratope += aa yield {'{}'.format(url_next):paratope} # if using process, you can just run a python new_imgt_spider.py # process = CrawlerProcess() # process.crawl(imgtSpider) # process.start()
3,237
1,094
# coding: utf-8 # # Copyright 2022 :Barry-Thomas-Paul: Moss # # Licensed under the Apache License, Version 2.0 (the "License") # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http: // www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Const Class # this is a auto generated file generated by Cheetah # Libre Office Version: 7.3 # Namespace: com.sun.star.script.provider class ScriptFrameworkErrorType(object): """ Const Class is a checked exception that represents an error encountered by the Scripting Framework whilst executing a script See Also: `API ScriptFrameworkErrorType <https://api.libreoffice.org/docs/idl/ref/namespacecom_1_1sun_1_1star_1_1script_1_1provider_1_1ScriptFrameworkErrorType.html>`_ """ __ooo_ns__: str = 'com.sun.star.script.provider' __ooo_full_ns__: str = 'com.sun.star.script.provider.ScriptFrameworkErrorType' __ooo_type_name__: str = 'const' UNKNOWN = 0 """ Unknown. """ NOTSUPPORTED = 1 """ ProviderNotSupported. """ NO_SUCH_SCRIPT = 2 """ the requested method, and/or with the requested signature, does not exist """ MALFORMED_URL = 3 """ the requested method, with the requested signature, does not exist """ __all__ = ['ScriptFrameworkErrorType']
1,667
529
#!/usr/bin/python3 # (c) Robert Muth - see LICENSE for more info from typing import List, Dict import enum from Util import cgen # maximum number of operands in an instruction MAX_OPERANDS = 5 # maximum number of function parameters (or results) MAX_PARAMETERS = 64 ############################################################ # Opcode Families [OF.] # # Each Opcode belongs to one of the families below. # Within each family the order and kind of the operands is similar ############################################################ @enum.unique class OPC_KIND(enum.Enum): INVALID = 0 ALU = 1 ALU1 = 2 MOV = 3 LEA = 4 LEA1 = 5 COND_BRA = 6 BRA = 7 BSR = 8 JSR = 9 SWITCH = 10 RET = 11 SYSCALL = 12 ST = 13 LD = 14 PUSHARG = 15 POPARG = 16 NOP = 17 NOP1 = 18 CONV = 19 CMP = 20 BCOPY = 21 BZERO = 22 DIRECTIVE = 23 # not a real instruction _OF_TO_PURPOSE = { OPC_KIND.ALU: ["dst", "src1", "src2"], OPC_KIND.ALU1: ["dst", "src"], OPC_KIND.COND_BRA: ["op1", "op2", "target_bbl"], OPC_KIND.SWITCH: ["index", "table"], OPC_KIND.BRA: ["target_bbl"], OPC_KIND.RET: [], OPC_KIND.BSR: ["target_fun"], OPC_KIND.JSR: ["target_fun_addr", "target_fun_sig"], OPC_KIND.SYSCALL: ["target_fun_sig", "syscall_no"], OPC_KIND.LEA: ["dst", "base", "offset"], OPC_KIND.LEA1: ["dst", "base"], OPC_KIND.LD: ["dst", "base", "offset"], OPC_KIND.ST: ["base", "offset", "src"], OPC_KIND.NOP: [], OPC_KIND.NOP1: ["src_and_dst"], OPC_KIND.BZERO: ["dst_addr", "width"], OPC_KIND.BCOPY: ["dst_addr", "src_addr", "width"], OPC_KIND.POPARG: ["dst"], OPC_KIND.PUSHARG: ["src"], OPC_KIND.CONV: ["dst", "src"], OPC_KIND.MOV: ["dst", "src"], OPC_KIND.CMP: ["dst", "src1", "src2", "cmp1", "cmp2"], } _OFS_CFG = {OPC_KIND.BSR, OPC_KIND.JSR, OPC_KIND.SYSCALL, OPC_KIND.SWITCH, OPC_KIND.BRA, OPC_KIND.COND_BRA, OPC_KIND.RET} # These instructions do not have a written register _OFS_NO_DEF = _OFS_CFG | {OPC_KIND.ST, OPC_KIND.BCOPY, OPC_KIND.BZERO, OPC_KIND.PUSHARG, OPC_KIND.NOP} # These instructions have a written register _OFS_WRITING_REGS = { OPC_KIND.LEA, OPC_KIND.LEA1, OPC_KIND.ALU, OPC_KIND.ALU1, OPC_KIND.CMP, OPC_KIND.MOV, OPC_KIND.CONV, OPC_KIND.LD, OPC_KIND.POPARG, OPC_KIND.NOP1} @enum.unique class OA(enum.Flag): """Opcode Attributes""" BBL_TERMINATOR = 1 << 0 NO_FALL_THROUGH = 1 << 1 CALL = 1 << 2 COMMUTATIVE = 1 << 3 MEM_RD = 1 << 4 MEM_WR = 1 << 5 SPECIAL = 1 << 6 OAS_CFG = OA.CALL | OA.BBL_TERMINATOR OAS_SIDE_EFFECT = OA.CALL | OA.BBL_TERMINATOR | OA.MEM_RD | OA.MEM_WR | OA.SPECIAL ############################################################ # Operand Kinds [OK.] # # Each instruction operates on a list of operands. Since we mimic a # three address machine, ALU instructions usually have 3 operands, # the destination being the first one. # There is a large variety of operands denoting registers or immediates # which enable some basic typing on a per operand basis. # Additional typing constraints across the operands are enforced by "rules". ############################################################ @enum.unique class OP_KIND(enum.Enum): INVALID = 0 REG = 1 CONST = 2 REG_OR_CONST = 3 # bbl immediates ref to a bbl in the current function # Note: bbls can be referred to before they are defined BBL = 4 # mem immediates refer to a global memory or stack region MEM = 5 # stk immediates refer to a stack region in the current function STK = 6 # fun immediates ref to a function in global function table # Note: funs can be referred to before they are defined FUN = 7 JTB = 8 TYPE_LIST = 20 DATA_KIND = 21 # one of the RK_ MEM_KIND = 23 # one of the MK_ FUN_KIND = 24 # one of the FK_ FIELD = 25 NAME = 26 NAME_LIST = 27 INT = 28 BBL_TAB = 29 BYTES = 30 ############################################################ # Type Constraints ############################################################ @enum.unique class TC(enum.Enum): INVALID = 0 ANY = 1 ADDR_NUM = 2 ADDR_INT = 3 NUM = 4 FLT = 5 INT = 6 ADDR = 7 CODE = 8 UINT = 9 SINT = 10 OFFSET = 11 # SAME_AS_PREV = 20 # for bitcast SAME_SIZE_AS_PREV = 22 ############################################################ # DataType Flavors ############################################################ DK_FLAVOR_S = 0x20 # signed int DK_FLAVOR_U = 0x40 # unsigned int DK_FLAVOR_F = 0x60 # ieee floating point DK_FLAVOR_A = 0x80 # data address DK_FLAVOR_C = 0xa0 # code address _DK_WIDTH_8 = 0 _DK_WIDTH_16 = 1 _DK_WIDTH_32 = 2 _DK_WIDTH_64 = 3 _DK_WIDTH_128 = 4 class DK(enum.Enum): """Data Kind - primarily used to associate a type with Const and Reg""" INVALID = 0 # signed S8 = DK_FLAVOR_S + _DK_WIDTH_8 S16 = DK_FLAVOR_S + _DK_WIDTH_16 S32 = DK_FLAVOR_S + _DK_WIDTH_32 S64 = DK_FLAVOR_S + _DK_WIDTH_64 # S128 = _RK_S + _RK_128 # unsigned U8 = DK_FLAVOR_U + _DK_WIDTH_8 U16 = DK_FLAVOR_U + _DK_WIDTH_16 U32 = DK_FLAVOR_U + _DK_WIDTH_32 U64 = DK_FLAVOR_U + _DK_WIDTH_64 # U128 = _RK_U + _RK_128 # float F8 = DK_FLAVOR_F + _DK_WIDTH_8 F16 = DK_FLAVOR_F + _DK_WIDTH_16 F32 = DK_FLAVOR_F + _DK_WIDTH_32 F64 = DK_FLAVOR_F + _DK_WIDTH_64 # F128 = _RK_F + _RK_128 # data address A32 = DK_FLAVOR_A + _DK_WIDTH_32 A64 = DK_FLAVOR_A + _DK_WIDTH_64 # code address C32 = DK_FLAVOR_C + _DK_WIDTH_32 C64 = DK_FLAVOR_C + _DK_WIDTH_64 def flavor(self) -> int: return self.value & 0xe0 def bitwidth(self) -> int: return 8 << (self.value & 0x7) SHORT_STR_TO_RK = {x.name: x for x in DK} # this does contain the aliases def RegIsAddrInt(rk: DK): return (DK.A32.value <= rk.value <= DK.A64.value or DK.S8.value <= rk.value <= DK.U64.value) def RegIsInt(rk: DK): return DK.S8.value <= rk.value <= DK.U64.value TC_TO_CHECKER = { TC.ANY: lambda x: True, TC.ADDR_NUM: lambda x: x.flavor() != DK_FLAVOR_C, TC.NUM: lambda x: x.flavor() in {DK_FLAVOR_U, DK_FLAVOR_S, DK_FLAVOR_F}, TC.INT: lambda x: x.flavor() in {DK_FLAVOR_U, DK_FLAVOR_S}, TC.ADDR: lambda x: x.flavor() == DK_FLAVOR_A, TC.CODE: lambda x: x.flavor() == DK_FLAVOR_C, TC.SINT: lambda x: x.flavor() == DK_FLAVOR_S, TC.UINT: lambda x: x.flavor() == DK_FLAVOR_U, TC.ADDR_INT: RegIsAddrInt, TC.FLT: lambda x: x.flavor() == DK_FLAVOR_F, TC.OFFSET: lambda x: x.flavor() in {DK_FLAVOR_U, DK_FLAVOR_S}, # maybe change this to just U or S } def CheckTypeConstraint(last_type: DK, constraint: TC, this_type: DK) -> bool: checker = TC_TO_CHECKER.get(constraint) if checker: return checker(this_type) if constraint == TC.SAME_AS_PREV: return last_type == this_type elif constraint == TC.SAME_SIZE_AS_PREV: return last_type.bitwidth() == this_type.bitwidth() else: assert False, f"unknown contraint {constraint.name}" @enum.unique class MEM_KIND(enum.Enum): """Represents Allocation Type of Global Memory """ INVALID = 0 RO = 1 RW = 2 TLS = 3 FIX = 4 # a fixed address provide via EXTERN = 5 # forward declaration must be defined before code emission BUILTIN = 6 # linker defined SHORT_STR_TO_MK = {x.name: x for x in MEM_KIND} @enum.unique class FUN_KIND(enum.Enum): """Function Kinds""" INVALID = 0 BUILTIN = 1 # linker defined EXTERN = 2 # forward declaration must be defined before code emission NORMAL = 3 SIGNATURE = 4 SHORT_STR_TO_FK = {x.name: x for x in FUN_KIND} ############################################################ # Operand Value Kind Sets ############################################################ OKS_LIST = {OP_KIND.BYTES, OP_KIND.NAME_LIST, OP_KIND.BBL_TAB, OP_KIND.TYPE_LIST} OKS_ALLOWED_FOR_INSTRUCTIONS = {OP_KIND.REG, OP_KIND.CONST, OP_KIND.REG_OR_CONST, OP_KIND.FUN, OP_KIND.BBL, OP_KIND.JTB, OP_KIND.MEM, OP_KIND.STK, OP_KIND.FIELD} # we do not want non-scalar operands in instructions as they # increase memory usage and complicate the code assert not (OKS_LIST & OKS_ALLOWED_FOR_INSTRUCTIONS) OKS_ALLOWED_FOR_DIRECTIVES = {OP_KIND.INT, OP_KIND.MEM_KIND, OP_KIND.BYTES, OP_KIND.NAME, OP_KIND.BBL_TAB, OP_KIND.FUN_KIND, OP_KIND.TYPE_LIST, OP_KIND.NAME_LIST, OP_KIND.DATA_KIND, OP_KIND.FUN, OP_KIND.MEM, OP_KIND.BBL } OKS_ALL = OKS_ALLOWED_FOR_INSTRUCTIONS | OKS_ALLOWED_FOR_DIRECTIVES ############################################################ # Opcode Groups ############################################################ @enum.unique class OPC_GENUS(enum.Enum): INVALID = 0 BASE = 1 TBD = 2 _DIR_TO_PURPOSE = { ".mem": ["name", "alignment", "mem_kind"], ".data": ["repeat", "data"], ".addr.fun": ["width", "fun"], ".addr.mem": ["width", "mem", "offset"], ".fun": ["name", "fun_kind", "out_params", "in_params"], ".bbl": ["name"], ".reg": ["reg_kind", "names"], ".stk": ["name", "alignment", "size"], ".jtb": ["name", "size", "default_bbl", "map"], ".struct": ["name"], ".field": ["name", "alignment", "size"], ".endstruct": [], ".stk.s": ["name", "name"], } ############################################################ # Opcode ############################################################ class Opcode: """Opcodes are templates for instructions similar to what you would find in assembly language manual for a processor. Note, the main purpose of instantiating an opcode instance is to populate the Table/TableByNo class member """ Table: Dict[str, "Opcode"] = {} TableByNo: Dict[int, "Opcode"] = {} def __init__(self, no, name: str, kind: OPC_KIND, operand_kinds: List[OP_KIND], constraints: List[TC], group: OPC_GENUS, desc, attributes=OA(0)): assert name not in Opcode.Table, f"duplicate opcode {name}" assert len(operand_kinds) <= MAX_OPERANDS, name Opcode.Table[name] = self assert no not in Opcode.TableByNo, f"duplicate no: {no} {name}" Opcode.TableByNo[no] = self self.no = no self.name = name self.kind: OPC_KIND = kind self.operand_kinds: List[OP_KIND] = operand_kinds self.constraints: List[TC] = constraints self.group = group self.desc = desc self.attributes = attributes assert kind != OPC_KIND.INVALID, f"unknown {kind}" is_directive = kind == OPC_KIND.DIRECTIVE if is_directive: assert name.startswith(".") self.purpose = _DIR_TO_PURPOSE[name] else: self.purpose = _OF_TO_PURPOSE[kind] assert len(self.purpose) == len( operand_kinds), f"{name} {operand_kinds}" assert len(operand_kinds) == len(constraints), f"{no} {name}" for ok, tc in zip(operand_kinds, constraints): # self.operands_tab[o] = op assert ok in OKS_ALL, f"unexpected operand: {ok}" if ok in {OP_KIND.REG, OP_KIND.CONST, OP_KIND.REG_OR_CONST}: assert tc != TC.INVALID, f"{no} {name}" else: assert tc == TC.INVALID, f"{no} {name}" if is_directive: assert ok in OKS_ALLOWED_FOR_DIRECTIVES, f"bad ins op [{ok}]" else: assert ok in OKS_ALLOWED_FOR_INSTRUCTIONS, f"bad ins op [{ok}]" def is_call(self): return OA.CALL in self.attributes def is_bbl_terminator(self): return OA.BBL_TERMINATOR in self.attributes def has_fallthrough(self): return OA.NO_FALL_THROUGH not in self.attributes def has_side_effect(self): return OAS_SIDE_EFFECT & self.attributes def def_ops_count(self): """How many of the leading operands write are register writes""" if self.kind in {OPC_KIND.INVALID, OPC_KIND.DIRECTIVE} or self.kind in _OFS_NO_DEF: return 0 else: return 1 @classmethod def Lookup(cls, name: str) -> "Opcode": return cls.Table[name] def __str__(self): return f"[OPCODE: {self.name}]" ############################################################ # ARITHMETIC ALU 0x10 # FLOAT + INT # note: limited address arithmetic allowed ADD = Opcode(0x10, "add", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.NUM, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Addition: dst := src1 + src2", OA.COMMUTATIVE) # note: limited address arithmetic allowed SUB = Opcode(0x11, "sub", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.NUM, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Subtraction: dst := src1 - src2 Note: `sub dst = 0 src` can be used to emulate `neg` for integers. (for floating point use `dat = mul src -1.0`) """) # needs more work wrt to size MUL = Opcode(0x12, "mul", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.NUM, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Multiplication: dst := src1 \\* src2", OA.COMMUTATIVE) DIV = Opcode(0x13, "div", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.NUM, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Division: dst := src1 / src2 Some day the operation might be more strictly defined as: dst := 0 if src2 == 0 else src1 / src2""") # cf.: # https://www.gingerbill.org/article/2020/01/25/a-reply-to-lets-stop-copying-c/ REM = Opcode(0x14, "rem", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Modulo: dst := a % b Some day the sign of the result might be more strictly defined. Note: does not apply to floating point numbers""") COPYSIGN = Opcode(0x15, "copysign", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Set the sign of src1 to match src2 (floating point only) Note: `copysign dst src1 0.0` can be used to emulate `abs`""") ############################################################ # LOGIC ALU 0x30 # INT ONLY (all regs are treated as unsigned except for shr/rshr XOR = Opcode(0x18, "xor", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Bitwise exclusive or: dst := src1 ^ src2 Note: `xor dst = src1 0b111...1` can be used to emulate `not`""", OA.COMMUTATIVE) # note: limited address arithmetic allowed AND = Opcode(0x19, "and", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Bitwise and: dst := src1 & src2", OA.COMMUTATIVE) # note: limited address arithmetic allowed OR = Opcode(0x1a, "or", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Bitwise or: dst := src1 | src2", OA.COMMUTATIVE) # shift amount is determined as follows: # use the log2(width(dst)) low order bits of src2 # e.g. for a dst of kind s8 the low order 3 bits of # src2 will be used. # src2 is treated as an unsigned register SHL = Opcode(0x1b, "shl", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Shift left: dst := src1 << src2 dst: = src1 << (src2 % bitwidth(src1))""") SHR = Opcode(0x1c, "shr", OPC_KIND.ALU, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Shift right: dst := src1 >> src2 dst: = src1 >> (src2 % bitwidth(src1))""") # do we need both directions, do we need a reverse version? # should we rather use a funnel shift? # ROTL = Opcode(0x1d, "rotl", OPC_KIND.ALU, # [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], # [TC.INT, TC.SAME_AS_PREV, TC.SAME_AS_PREV], OPC_GENUS.TBD, # "Rotation Left") ############################################################ # CONDITIONAL BRANCHES 0x20 # do we need unordered variants for floating point? # not beq/bne is the only operation for c_regs BEQ = Opcode(0x20, "beq", OPC_KIND.COND_BRA, [OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.BBL], [TC.ANY, TC.SAME_AS_PREV, TC.INVALID], OPC_GENUS.BASE, "Conditional branch if equal.", OA.COMMUTATIVE | OA.BBL_TERMINATOR) BNE = Opcode(0x21, "bne", OPC_KIND.COND_BRA, [OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.BBL], [TC.ANY, TC.SAME_AS_PREV, TC.INVALID], OPC_GENUS.BASE, "Conditional branch if not equal.", OA.COMMUTATIVE | OA.BBL_TERMINATOR) BLT = Opcode(0x22, "blt", OPC_KIND.COND_BRA, [OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.BBL], [TC.ADDR_NUM, TC.SAME_AS_PREV, TC.INVALID], OPC_GENUS.BASE, "Conditional branch if greater than.", OA.BBL_TERMINATOR) BLE = Opcode(0x23, "ble", OPC_KIND.COND_BRA, [OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.BBL], [TC.ADDR_NUM, TC.SAME_AS_PREV, TC.INVALID], OPC_GENUS.BASE, "Conditional branch if less or equal.", OA.BBL_TERMINATOR) ############################################################ # More Control Flow 0x28 SWITCH = Opcode(0x28, "switch", OPC_KIND.SWITCH, [OP_KIND.REG, OP_KIND.JTB], [TC.UINT, TC.INVALID], OPC_GENUS.BASE, """Multi target computed jump. The register argument must be less than the jtb `size`. The jtb symbol must have been previously defined with the `.jtb` directive. """, OA.BBL_TERMINATOR | OA.NO_FALL_THROUGH) BRA = Opcode(0x29, "bra", OPC_KIND.BRA, [OP_KIND.BBL], [TC.INVALID], OPC_GENUS.BASE, "Unconditional branch.", OA.BBL_TERMINATOR | OA.NO_FALL_THROUGH) RET = Opcode(0x2a, "ret", OPC_KIND.RET, [], [], OPC_GENUS.BASE, "Return from subroutine.", OA.BBL_TERMINATOR | OA.NO_FALL_THROUGH) BSR = Opcode(0x2b, "bsr", OPC_KIND.BSR, [OP_KIND.FUN], [TC.INVALID], OPC_GENUS.BASE, "Branch to subroutine fun", OA.CALL) JSR = Opcode(0x2c, "jsr", OPC_KIND.JSR, [OP_KIND.REG, OP_KIND.FUN], [TC.CODE, TC.INVALID], OPC_GENUS.BASE, """Jump indirectly to subroutine through register (fun describes the signature). The signature must have been previously defined with the `.fun` directive.""", OA.CALL) SYSCALL = Opcode(0x2d, "syscall", OPC_KIND.SYSCALL, [OP_KIND.FUN, OP_KIND.CONST], [TC.INVALID, TC.UINT], OPC_GENUS.BASE, """Syscall to `syscall_no`. (fun describes the signature). The signature must have been previously defined with the `.fun` directive.""", OA.CALL) TRAP = Opcode(0x2e, "trap", OPC_KIND.RET, [], [], OPC_GENUS.BASE, "Abort program.", OA.BBL_TERMINATOR | OA.NO_FALL_THROUGH) ############################################################ # Misc 0x30 PUSHARG = Opcode(0x30, "pusharg", OPC_KIND.PUSHARG, [OP_KIND.REG_OR_CONST], [TC.ANY], OPC_GENUS.BASE, "push a call or return arg - must immediately precede bsr/jsr or ret.", OA.SPECIAL) POPARG = Opcode(0x31, "poparg", OPC_KIND.POPARG, [OP_KIND.REG], [TC.ANY], OPC_GENUS.BASE, "pop a call or return arg - must immediately follow fun entry or bsr/jsr.", OA.SPECIAL) CONV = Opcode(0x32, "conv", OPC_KIND.CONV, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.NUM, TC.NUM], OPC_GENUS.BASE, # TODO: specify rounding and overflow for float <-> int conversions """Conversion of numerical regs which do not have to be of same size. Bits may change. If the conversion involves both a widening and a change of type, the widening is performed first. """) BITCAST = Opcode(0x33, "bitcast", OPC_KIND.CONV, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.ANY, TC.SAME_SIZE_AS_PREV], OPC_GENUS.BASE, """Cast between regs of same size. Bits will be re-interpreted but do not change. This is useful for manipulating addresses in unusual ways or looking at the binary representation of floats.""") MOV = Opcode(0x34, "mov", OPC_KIND.MOV, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.ANY, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Move between registers. While a mov can be emulated via a `add dst = src 0`, having a dedicated instruction makes some optimizations easier to implement when combined with a canonicalization.""") CMPEQ = Opcode(0x35, "cmpeq", OPC_KIND.CMP, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.ANY, TC.SAME_AS_PREV, TC.SAME_AS_PREV, TC.ANY, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Conditional move (compare equal). dst := (cmp1 == cmp2) ? src1 : src2 Note: dst/cmp1/cmp2 may be of a different type than src1/src2.""", OA.COMMUTATIVE) CMPLT = Opcode(0x36, "cmplt", OPC_KIND.CMP, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.ANY, TC.SAME_AS_PREV, TC.SAME_AS_PREV, TC.ADDR_NUM, TC.SAME_AS_PREV], OPC_GENUS.BASE, """Conditional move (compare less than). dst := (cmp1 < cmp2) ? src1 : src2 Note: dst/cmp1/cmp2 may be of a different type than src1/src2.""") # materialize addresses in a register LEA = Opcode(0x38, "lea", OPC_KIND.LEA, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.ADDR, TC.SAME_AS_PREV, TC.OFFSET], OPC_GENUS.BASE, """Load effective Address. dst := base + offset Note: dst and base are addresses but offset is not.""") LEA_MEM = Opcode(0x39, "lea.mem", OPC_KIND.LEA, [OP_KIND.REG, OP_KIND.MEM, OP_KIND.REG_OR_CONST], [TC.ADDR, TC.INVALID, TC.OFFSET], OPC_GENUS.BASE, "Load effective memory address with offset, dst := base + offset") LEA_STK = Opcode(0x3a, "lea.stk", OPC_KIND.LEA, [OP_KIND.REG, OP_KIND.STK, OP_KIND.REG_OR_CONST], [TC.ADDR, TC.INVALID, TC.OFFSET], OPC_GENUS.BASE, "Load effective stack address with offset. dst := base + offset") LEA_FUN = Opcode(0x3b, "lea.fun", OPC_KIND.LEA1, [OP_KIND.REG, OP_KIND.FUN], [TC.CODE, TC.INVALID], OPC_GENUS.BASE, "Load effective function address: dst := base (note: no offset).") ############################################################ # LOAD STORE 0x60 # ld/st base address is in register, offset is immediate # ld/st base address is register LD = Opcode(0x40, "ld", OPC_KIND.LD, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.ANY, TC.ADDR, TC.OFFSET], OPC_GENUS.BASE, "Load from register base with offset. dst := RAM[base + offset]", OA.MEM_RD) # note: signedness of offset may matter here LD_MEM = Opcode(0x41, "ld.mem", OPC_KIND.LD, [OP_KIND.REG, OP_KIND.MEM, OP_KIND.REG_OR_CONST], [TC.ANY, TC.INVALID, TC.OFFSET], OPC_GENUS.BASE, "Load from memory base with offset. dst := RAM[base + offset] ", OA.MEM_RD) LD_STK = Opcode(0x42, "ld.stk", OPC_KIND.LD, [OP_KIND.REG, OP_KIND.STK, OP_KIND.REG_OR_CONST], [TC.ANY, TC.INVALID, TC.OFFSET], OPC_GENUS.BASE, "Load from stack base with offset. dst := RAM[base + offset]", OA.MEM_RD) ST = Opcode(0x48, "st", OPC_KIND.ST, [OP_KIND.REG, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.ADDR, TC.OFFSET, TC.ANY], OPC_GENUS.BASE, "Store to register base with offset. RAM[base + offset] := src", OA.MEM_WR) ST_MEM = Opcode(0x49, "st.mem", OPC_KIND.ST, [OP_KIND.MEM, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INVALID, TC.OFFSET, TC.ANY], OPC_GENUS.BASE, "Store to memory base with offset. RAM[base + offset] := src", OA.MEM_WR) ST_STK = Opcode(0x4a, "st.stk", OPC_KIND.ST, [OP_KIND.STK, OP_KIND.REG_OR_CONST, OP_KIND.REG_OR_CONST], [TC.INVALID, TC.OFFSET, TC.ANY], OPC_GENUS.BASE, "Store to stack base with offset. RAM[base + offset] := src", OA.MEM_WR) ############################################################ # FLOAT ALU OPERAND: 0x50 CEIL = Opcode(0x50, "ceil", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Round float to integral, toward positive infinity") FLOOR = Opcode(0x51, "floor", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Round float to integral, toward negative infinity") ROUND = Opcode(0x52, "round", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Round float to integral, to nearest with ties to away") TRUNC = Opcode(0x53, "trunc", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.BASE, """ Round float to integral, toward zero. Note, frac(val) = val - trunc(val)""") SQRT = Opcode(0x54, "sqrt", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Compute the sqrt of floating point value") # do we need all these? Opcode(0x58, "sin", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x59, "cos", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x5a, "tan", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x5b, "asin", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x5c, "acos", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x5d, "atan", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x5e, "exp", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") Opcode(0x5f, "log", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.FLT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") ############################################################ # Advanced ALU ############################################################ CNTLZ = Opcode(0x60, "cntlz", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Count leading zeros.") CNTTZ = Opcode(0x61, "cnttz", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV], OPC_GENUS.BASE, "Count trailing zeros.") # INT SINGLE OPERAND 0xb0 # the src reg is treated as an unsigned reg Opcode(0x62, "cntpop", OPC_KIND.ALU1, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.INT, TC.SAME_AS_PREV], OPC_GENUS.TBD, "TBD") ############################################################ # Annotations ############################################################ NOP = Opcode(0x70, "nop", OPC_KIND.NOP, [], [], OPC_GENUS.BASE, "nop - internal use.") NOP1 = Opcode(0x71, "nop1", OPC_KIND.NOP1, [OP_KIND.REG], [TC.ANY], OPC_GENUS.BASE, "nop with one reg - internal use. Can be used to `reserve` a reg for code generation.", OA.SPECIAL) # LINE = Opcode(0x78, "line", OPC_KIND., [OP_KIND.NAME, OP_KIND.CONST], # [TC.ANY], OPC_GENUS.BASE, # "", # OA.SPECIAL) ############################################################ # Misc Experimental ############################################################ # Note, negative lengths copy downwards Opcode(0xb8, "bcopy", OPC_KIND.BCOPY, [OP_KIND.REG, OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.ADDR, TC.SAME_AS_PREV, TC.OFFSET], OPC_GENUS.TBD, "TBD", OA.MEM_WR | OA.MEM_RD) # Note, negative lengths copy downwards Opcode(0xba, "bzero", OPC_KIND.BZERO, [OP_KIND.REG, OP_KIND.REG_OR_CONST], [TC.ADDR, TC.OFFSET], OPC_GENUS.TBD, "TBD", OA.MEM_WR) ############################################################ # Directives 0xd # # do not correspond to instructions ############################################################ def Directive(no: int, name: str, operands, desc, group=OPC_GENUS.BASE): return Opcode(no, name, OPC_KIND.DIRECTIVE, operands, constraints=[TC.INVALID] * len(operands), desc=desc, group=group) Directive(0x01, ".mem", [OP_KIND.NAME, OP_KIND.INT, OP_KIND.MEM_KIND], "Add new memory region to unit") Directive(0x02, ".data", [OP_KIND.INT, OP_KIND.BYTES], "Add content to current memory region: multiple bytes") Directive(0x03, ".addr.fun", [OP_KIND.INT, OP_KIND.FUN], "Add content to current memory region: code address") Directive(0x04, ".addr.mem", [OP_KIND.INT, OP_KIND.MEM, OP_KIND.INT], "Add content to current memory region: " "memory address with offset") Directive(0x05, ".fun", [OP_KIND.NAME, OP_KIND.FUN_KIND, OP_KIND.TYPE_LIST, OP_KIND.TYPE_LIST], "Add new function to unit") Directive(0x06, ".bbl", [OP_KIND.NAME], "Add new basic block to current function") Directive(0x07, ".reg", [OP_KIND.DATA_KIND, OP_KIND.NAME_LIST], "Add new registers to current function") Directive(0x08, ".stk", [OP_KIND.NAME, OP_KIND.INT, OP_KIND.INT], "Add stack region to current function") Directive(0x09, ".jtb", [OP_KIND.NAME, OP_KIND.INT, OP_KIND.BBL, OP_KIND.BBL_TAB], "bbl jump table: <name> <size> <default-bbl> <sparse-table>") ############################################################ # experimental/unimplemented ############################################################ # add/sub/rotate with carry for legalizing say 64bit regs into pairs of 32bit regs # unreachable # swap # unordered comparison # https://stackoverflow.com/questions/8627331/what-does-ordered-unordered-comparison-mean # conv int - flt (urgent) # conv int - int (urgent) # extract (urgent) # insert (urgent) # ld_l, st_C, cmpxch, cmpswp # pow, pow2 powi # log # crc32c (supported by x86-64 and arm64 - using 0x1EDC6F41) # aes ??? # ld.scaled /st.scaled: base_reg + index_reg * scale imm + offset_imm # copysign # prefetch # other built-ins: cf.: # https://github.com/llvm-mirror/compiler-rt/tree/master/lib/builtins _GROUPS = { 0x01: "## Directives\n", 0x10: "## Basic ALU\n", 0x20: "## Conditional Branches\n", 0x28: "## Other Control Flow\n", 0x30: "## Move/Conversion\n", 0x38: "## Address Arithmetic\n", 0x40: "## Load\n", 0x48: "## Store\n", 0x50: "## Float ALU\n", 0x60: "## Advanced ALU\n", 0x70: "## Annotation\n", 0xf1: "## Misc\n", } def _render_operand_desc(purpose: str, kind: OP_KIND, constraint: TC, mod1="", mod2="") -> str: kind_str = kind.name.replace("REG_OR_CONST", "REG/CONST") if constraint == TC.INVALID: return f"*{purpose}* {mod1}{kind_str}{mod2}" else: return f"*{purpose}* {mod1}{kind_str}:{constraint.name}{mod2}" def _render_directive_doc(o: Opcode, fout): print_ops = [_render_operand_desc(*t, mod1="<sub>[", mod2="]</sub>") for t in zip(o.purpose, o.operand_kinds, o.constraints)] print(f"#### [{o.no:02x}] {o.name} {' '.join(print_ops)}", file=fout) print(o.desc, file=fout) def _render_opcode_doc(o: Opcode, fout): print_ops = [_render_operand_desc(*t, mod1="<sub>[", mod2="]</sub>") for t in zip(o.purpose, o.operand_kinds, o.constraints)] if o.kind in _OFS_WRITING_REGS: print_ops.insert(1, "=") if o.kind in {OPC_KIND.ST}: print_ops.insert(-1, "=") print(f"#### [{o.no:02x}] {o.name} {' '.join(print_ops)}", file=fout) print(o.desc, file=fout) # print("* constraints:", ' '.join(ops)) # print(f"{name:15.15}, // {' '.join(ops)} [{' # '.join(cons)}]" def _render_documentation(fout): for opc in Opcode.Table.values(): if opc.group != OPC_GENUS.BASE: continue if opc.no in _GROUPS: print(_GROUPS[opc.no], file=fout) if opc.kind == OPC_KIND.DIRECTIVE: _render_directive_doc(opc, fout) else: _render_opcode_doc(opc, fout) print() def _render_h(fout): print("enum class OPC : uint8_t {", file=fout) last = 0 print(f" INVALID = 0x00,", file=fout) for opc in Opcode.Table.values(): if opc.group != OPC_GENUS.BASE: continue if (opc.no & 0xff0) != last & 0xff0: print("", file=fout) last = opc.no name = opc.name.upper().replace(".", "_") if opc.kind == OPC_KIND.DIRECTIVE: name = "DIR_" + name[1:] print(f" {name} = 0x{opc.no:02x},", file=fout) print("};", file=fout) # _render_enum("OpcodeFamily", ["OF.INVALID", "OF.DIRECTIVE"] + # list(OFS_ALL)) # _render_enum("OperandKind", ["OK.INVALID"] + # [x.upper() for x in OKS_ALL]) for cls in [OPC_GENUS, FUN_KIND, MEM_KIND, TC, OPC_KIND, DK, OP_KIND]: cgen.RenderEnum(cgen.NameValues(cls), f"class {cls.__name__} : uint8_t", fout) cgen.RenderEnum(cgen.NameValues(OA), f"{OA.__name__} : uint16_t", fout) def _render_c(fout): def render(cls, both_ways=True): cgen.RenderEnumToStringMap(cgen.NameValues(cls), cls.__name__, fout) cgen.RenderEnumToStringFun(cls.__name__, fout) if both_ways: cgen.RenderStringToEnumMap(cgen.NameValues(cls), cls.__name__ + "FromStringMap", cls.__name__ + "Jumper", fout) render(OPC_GENUS) render(FUN_KIND) render(MEM_KIND) render(TC) render(DK) render(OP_KIND, False) alpha = [(opc.name, opc.no) for opc in Opcode.Table.values()] cgen.RenderStringToEnumMap(alpha, "OPCFromStringMap", "OPCJumper", fout) print("const Opcode GlobalOpcodes[256] = {") opcodes = sorted([(o.no, o) for o in Opcode.Table.values()]) last = -1 dummy_opc = Opcode(0, "", OPC_KIND.RET, [], [], OPC_GENUS.INVALID, "") dummy_opc.name = "" dummy_opc.kind = OPC_KIND.INVALID def emit_one(opc: Opcode): kinds_str = [f"OP_KIND::{x.name}" for x in opc.operand_kinds] constraints_str = [f"TC::{x.name}" for x in opc.constraints] attributes = [f"OA::{x.name}" for x in OA if x in opc.attributes] if not attributes: attributes = ["0"] print(" { // %2x %s" % (opc.no, opc.name)) print(' {%s}, ' % ", ".join(kinds_str)) print(' OPC_KIND::%s, OPC_GENUS::%s, %d, %d,' % (opc.kind.name, opc.group.name, len(opc.operand_kinds), opc.def_ops_count())) print(' {%s}, ' % ", ".join(constraints_str)) print(' "%s", %s },' % (opc.name, '|'.join(attributes))) for n, o in opcodes: if o.group != OPC_GENUS.BASE: continue last += 1 while last < n: dummy_opc.no = last emit_one(dummy_opc) last += 1 emit_one(o) print("};\n") def Dump(): last = None for opc in Opcode.Table.values(): if opc.kind != last: print() last = opc.kind ops = [_render_operand_desc(a, b, c) for a, b, c in zip(opc.purpose, opc.operand_kinds, opc.constraints)] print(f"{opc.kind.name} {opc.name} {' '.join(ops)}") print("total opcodes: %d" % len(Opcode.Table)) if __name__ == "__main__": import sys if len(sys.argv) > 1: if sys.argv[1] == "documentation": cgen.ReplaceContent(_render_documentation, sys.stdin, sys.stdout) elif sys.argv[1] == "gen_h": cgen.ReplaceContent(_render_h, sys.stdin, sys.stdout) elif sys.argv[1] == "gen_c": cgen.ReplaceContent(_render_c, sys.stdin, sys.stdout) else: Dump()
38,786
15,211
# Generated by Django 3.2.7 on 2021-10-27 09:40 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ("connector_airflow", "0012_remove_airflow_run_message"), ] operations = [ migrations.AlterModelOptions( name="dagrun", options={"ordering": ("-execution_date",)}, ), migrations.AlterField( model_name="dagrun", name="state", field=models.CharField( choices=[ ("success", "Success"), ("running", "Running"), ("failed", "Failed"), ("queued", "Queued"), ], max_length=200, ), ), ]
784
225
# Generated by the protocol buffer compiler. DO NOT EDIT! # source: LAMARCK_ML/data_util/TypeShape.proto import sys _b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1')) from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database from google.protobuf import descriptor_pb2 # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() from LAMARCK_ML.data_util import Shape_pb2 as LAMARCK__ML_dot_data__util_dot_Shape__pb2 from LAMARCK_ML.data_util import DType_pb2 as LAMARCK__ML_dot_data__util_dot_DType__pb2 DESCRIPTOR = _descriptor.FileDescriptor( name='LAMARCK_ML/data_util/TypeShape.proto', package='LAMARCK_ML', syntax='proto3', serialized_pb=_b('\n$LAMARCK_ML/data_util/TypeShape.proto\x12\nLAMARCK_ML\x1a LAMARCK_ML/data_util/Shape.proto\x1a LAMARCK_ML/data_util/DType.proto\"w\n\x0eTypeShapeProto\x12\x0f\n\x07id_name\x18\x01 \x01(\t\x12)\n\tdtype_val\x18\x02 \x01(\x0b\x32\x16.LAMARCK_ML.DTypeProto\x12)\n\tshape_val\x18\x03 \x01(\x0b\x32\x16.LAMARCK_ML.ShapeProtob\x06proto3') , dependencies=[LAMARCK__ML_dot_data__util_dot_Shape__pb2.DESCRIPTOR,LAMARCK__ML_dot_data__util_dot_DType__pb2.DESCRIPTOR,]) _sym_db.RegisterFileDescriptor(DESCRIPTOR) _TYPESHAPEPROTO = _descriptor.Descriptor( name='TypeShapeProto', full_name='LAMARCK_ML.TypeShapeProto', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='id_name', full_name='LAMARCK_ML.TypeShapeProto.id_name', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dtype_val', full_name='LAMARCK_ML.TypeShapeProto.dtype_val', index=1, number=2, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shape_val', full_name='LAMARCK_ML.TypeShapeProto.shape_val', index=2, number=3, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=120, serialized_end=239, ) _TYPESHAPEPROTO.fields_by_name['dtype_val'].message_type = LAMARCK__ML_dot_data__util_dot_DType__pb2._DTYPEPROTO _TYPESHAPEPROTO.fields_by_name['shape_val'].message_type = LAMARCK__ML_dot_data__util_dot_Shape__pb2._SHAPEPROTO DESCRIPTOR.message_types_by_name['TypeShapeProto'] = _TYPESHAPEPROTO TypeShapeProto = _reflection.GeneratedProtocolMessageType('TypeShapeProto', (_message.Message,), dict( DESCRIPTOR = _TYPESHAPEPROTO, __module__ = 'LAMARCK_ML.data_util.TypeShape_pb2' # @@protoc_insertion_point(class_scope:LAMARCK_ML.TypeShapeProto) )) _sym_db.RegisterMessage(TypeShapeProto) # @@protoc_insertion_point(module_scope)
3,447
1,387
from flask import Flask, render_template, request, session, redirect, url_for import pymysql.cursors import datetime from pyecharts import options as opts from pyecharts.charts import Pie,Bar from appdef import * #Get the airline the staff member works for def getStaffAirline(): username = session['username'] cursor = conn.cursor() #username is a primary key query = 'select airline_name from airline_staff where username = %s' cursor.execute(query, (username)) #fetchall returns an array, each element is a dictionary airline = cursor.fetchall()[0]['airline_name'] cursor.close() return airline #Make sure that the user is actually staff before performing any operations def authenticateStaff(): username = "" try: #could be that there is no user, make sure username = session['username'] except: return False cursor = conn.cursor() query = 'select * from airline_staff where username=%s' cursor.execute(query, (username)) data = cursor.fetchall() cursor.close() if data: return True else: #Logout before returning error message session.pop('username') return False @app.route('/staffHome') def staffHome(): if authenticateStaff(): username = session['username'] message = request.args.get('message') cursor = conn.cursor() queryGetairline = "SELECT airline_name FROM airline_staff WHERE username= %s" cursor.execute(queryGetairline, username) airline_name = cursor.fetchone()['airline_name'] # query top destination for the past 3 months query1 = "select count(ticket.ticket_id) as cnt, airport.airport_city as city\ from airport,flight,ticket,purchases\ where airport.airport_name = flight.arrival_airport\ and flight.flight_num = ticket.flight_num\ and flight.airline_name = %s\ and purchases.ticket_id = ticket.ticket_id\ and purchases.purchase_date between DATE_SUB(curdate(), INTERVAL 3 MONTH) and curdate()\ group by city \ order by cnt DESC limit 3" cursor.execute(query1,airline_name) data1 = cursor.fetchall() if len(data1)<3: num = len(data1) range1 = range(num) data1 = [data1[i]['city'] for i in range(num)] else: range1 = range(3) data1 = [data1[i]['city'] for i in range(3)] # query top destination for the past 1 year query2 = "select count(ticket.ticket_id) as cnt, airport.airport_city as city\ from airport,flight,ticket,purchases\ where airport.airport_name = flight.arrival_airport\ and flight.flight_num = ticket.flight_num\ and flight.airline_name = %s\ and purchases.ticket_id = ticket.ticket_id\ and purchases.purchase_date between DATE_SUB(curdate(), INTERVAL 1 YEAR) and curdate()\ group by city \ order by cnt DESC limit 3" cursor.execute(query2,airline_name) data2 = cursor.fetchall() if len(data2)<3: num = len(data2) range2 = range(num) data2 = [data2[i]['city'] for i in range(num)] else: range2 = range(3) data2 = [data2[i]['city'] for i in range(3)] cursor.close() return render_template('staff.html', username=username, message=message, destination1 = data1, destination2 = data2, range1 = range1, range2 = range2) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/searchFlights') def searchFlightsPage(): if authenticateStaff(): cursor = conn.cursor() airline = getStaffAirline() query = "select * from flight where airline_name = %s \ and ((departure_time between curdate() and date_add(curdate(), interval 30 day)) \ or (arrival_time between curdate() and date_add(curdate(), interval 30 day)))" cursor.execute(query, (airline)) data = cursor.fetchall() cursor.close() error = request.args.get('error') return render_template('searchStaff.html', error=error, results=data) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/searchFlights/city', methods=['POST']) def searchFlightsCity(): if authenticateStaff(): cursor = conn.cursor() city = request.form['citysearchbox'] airline = getStaffAirline() query = "select * from flight,airport \ where (airport.airport_name=flight.departure_airport or airport.airport_name=flight.arrival_airport) \ and airport.airport_city=%s and airline_name=%s" cursor.execute(query, (city, airline)) data = cursor.fetchall() cursor.close() error = None if data: return render_template('searchStaffResults.html', results=data) else: #returns an error message to the html page error = 'No results found' return redirect(url_for('searchFlightsPage', error=error)) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/searchFlights/airport', methods=['POST']) def searchFlightsAirport(): if authenticateStaff(): cursor = conn.cursor() airport = request.form['airportsearchbox'] airline = getStaffAirline() query = 'select * from flight where (departure_airport = %s or arrival_airport = %s) and airline_name=%s' cursor.execute(query, (airport, airport, airline)) data = cursor.fetchall() cursor.close() error = None if data: return render_template('searchStaffResults.html', results=data) else: #returns an error message to the html page error = 'No results found' return redirect(url_for('searchFlightsPage', error=error)) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/searchFlights/date', methods=['POST']) def searchFlightsDate(): if authenticateStaff(): begintime = request.form['begintime'] endtime = request.form['endtime'] if not validateDates(begintime, endtime): error = 'Invalid date range' return redirect(url_for('searchFlightsPage', error=error)) airline = getStaffAirline() cursor = conn.cursor() query = "select * from flight \ where ((departure_time between %s and %s) \ or (arrival_time between %s and %s)) and airline_name=%s" cursor.execute(query, (begintime, endtime, begintime, endtime, airline)) data = cursor.fetchall() cursor.close() error = None if data: return render_template('searchStaffResults.html', results=data) else: #returns an error message to the html page error = 'No results found' return redirect(url_for('searchFlightsPage', error=error)) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/searchFlights/customers', methods=['POST']) def searchFlightsCustomer(): if authenticateStaff(): flightnum = request.form['flightsearchbox'] airline = getStaffAirline() cursor = conn.cursor() query = "select customer_email from purchases natural join ticket\ where flight_num = %s and airline_name=%s" cursor.execute(query, (flightnum, airline)) data = cursor.fetchall() cursor.close() if data: return render_template('searchStaffResults.html', customerresults=data, flightnum=flightnum) else: #returns an error message to the html page error = 'No results found' return redirect(url_for('searchFlightsPage', error=error)) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/createFlight') def createFlightPage(): if authenticateStaff(): airline = getStaffAirline() cursor = conn.cursor() airline = getStaffAirline() query = "select * from flight where airline_name = %s \ and ((departure_time between curdate() and date_add(curdate(), interval 30 day)) \ or (arrival_time between curdate() and date_add(curdate(), interval 30 day)))" cursor.execute(query, (airline)) data = cursor.fetchall() cursor = conn.cursor() query = 'select distinct airport_name from airport' cursor.execute(query) airportdata = cursor.fetchall() query = 'select distinct airplane_id from airplane where airline_name=%s' cursor.execute(query, (airline)) airplanedata = cursor.fetchall() cursor.close() error = request.args.get('error') return render_template('createFlight.html', error = error, airportdata = airportdata, airplanedata = airplanedata, results = data) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/createFlight/Auth', methods=['POST']) def createFlight(): # prevent unauthorized users from doing this action if not authenticateStaff(): error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) username = session['username'] flightnum = request.form['flightnum'] departport = request.form['departport'] departtime = request.form['departtime'] arriveport = request.form['arriveport'] arrivetime = request.form['arrivetime'] price = request.form['price'] status = "Upcoming" airplaneid = request.form['airplanenum'] ########################################################################## airline = getStaffAirline() cursor = conn.cursor() query1 = 'select * from flight where airline_name = %s and flight_num = %s' cursor.execute(query1,(airline,flightnum)) data1 = cursor.fetchall() if data1: error = "The flight number already exists, please enter another one." return redirect(url_for('createFlightPage', error=error)) cursor.close() ############################################################################# ############################################################################# cursor = conn.cursor() query2 = 'select * from airport where airport_name = %s ' cursor.execute(query2,(departport)) data2 = cursor.fetchall() query3 = 'select * from airport where airport_name = %s ' cursor.execute(query3,(arriveport)) data3 = cursor.fetchall() if (not data2): error = "The Departure Airport does not exist, please add the airport first." return redirect(url_for('createFlightPage', error=error)) if (not data3): error = "The Arrival Airport does not exist, please add the airport first." return redirect(url_for('createFlightPage', error=error)) cursor.close() ############################################################################# if not validateDates(departtime, arrivetime): error = 'Invalid date range' return redirect(url_for('createFlightPage', error=error)) airline = getStaffAirline() #Check that airplane is valid cursor = conn.cursor() query = 'select * from airplane where airplane_id = %s' cursor.execute(query, (airplaneid)) data = cursor.fetchall() if not data: error = 'Invalid Airplane ID' return redirect(url_for('createFlightPage', error=error)) query = 'insert into flight values (%s, %s, %s, %s, %s, %s, %s, %s, %s)' cursor.execute(query, (airline, flightnum, departport, departtime, arriveport, arrivetime, price, status, airplaneid)) conn.commit() cursor.close() return redirect(url_for('staffHome', message="Operation Successful")) @app.route('/staffHome/changeFlight') def changeFlightStatusPage(): if authenticateStaff(): error = request.args.get('error') return render_template('changeFlight.html', error=error) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/changeFlight/Auth', methods=['POST']) def changeFlightStatus(): # prevent unauthorized users from doing this action if not authenticateStaff(): error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) username = session['username'] cursor = conn.cursor() flightnum = request.form['flightnum'] status = request.form['status'] if not status: error = 'Did not select new status' return redirect(url_for('changeFlightStatusPage', error=error)) airline = getStaffAirline() #Check that the flight is from the same airline as the staff query = 'select * from flight where flight_num = %s and airline_name = %s' cursor.execute(query, (flightnum, airline)) data = cursor.fetchall() ################################################################################## if not data: error = 'Incorrect enter - flight number is not in your airline ' return redirect(url_for('changeFlightStatusPage', error=error)) ################################################################################## #Update the specified flight query = 'update flight set status=%s where flight_num=%s and airline_name = %s' cursor.execute(query, (status, flightnum, airline)) conn.commit() cursor.close() return redirect(url_for('staffHome', message="Operation Successful")) @app.route('/staffHome/addAirplane') def addAirplanePage(): if authenticateStaff(): error = request.args.get('error') return render_template('addAirplane.html', error=error) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/addAirplane/confirm', methods=['POST']) def addAirplane(): # prevent unauthorized users from doing this action if not authenticateStaff(): error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) username = session['username'] planeid = request.form['id'] seats = request.form['seats'] airline = getStaffAirline() #Check if planeid is not taken cursor = conn.cursor() query = 'select * from airplane where airplane_id = %s' cursor.execute(query, (planeid)) data = cursor.fetchall() if data: error = "Airplane ID already taken" return redirect(url_for('addAirplanePage', error=error)) #Insert the airplane query = 'insert into airplane values (%s, %s, %s)' cursor.execute(query, (airline, planeid, seats)) conn.commit() #Get a full list of airplanes query = 'select * from airplane where airline_name = %s' cursor.execute(query, (airline)) data = cursor.fetchall() cursor.close() return render_template('addAirplaneConfirm.html', results=data) @app.route('/staffHome/addAirport') def addAirportPage(): if authenticateStaff(): error = request.args.get('error') return render_template('addAirport.html', error=error) else: error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/addAirport/Auth', methods=['POST']) def addAirport(): # prevent unauthorized users from doing this action if not authenticateStaff(): error = 'Invalid Credentials' return redirect(url_for('errorpage', error=error)) username = session['username'] name = request.form['name'] city = request.form['city'] ##################################################################### if len(name)>3: error = "Please enter the abbreviation of airport." return redirect(url_for('addAirportPage', error=error)) cursor = conn.cursor() query = "select * from airport where airport_name = %s and airport_city = %s" cursor.execute(query,(name,city)) data1 = cursor.fetchall() cursor.close() if data1: error = "Airport Already exits." return redirect(url_for('addAirportPage', error=error)) ##################################################################### cursor = conn.cursor() query = 'insert into airport values (%s, %s)' cursor.execute(query, (name, city)) conn.commit() cursor.close() return redirect(url_for('staffHome', message="Operation Successful")) @app.route('/staffHome/viewAgents') def viewAgentsPage(): if authenticateStaff(): error = request.args.get('error') return render_template('viewAgents.html', error=error) else: error = "Invalid Credentials" return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/viewAgents/sales', methods=['POST']) def viewAgentsSales(): if authenticateStaff(): daterange = request.form['range'] airline = getStaffAirline() #datrange specify the past 1 month or year cursor = conn.cursor() query = 'select email,count(ticket_id) as sales \ from booking_agent natural join purchases natural join ticket \ where purchase_date >= date_sub(curdate(), interval 1 ' + daterange + ') \ and airline_name=%s group by email order by sales DESC limit 5' cursor.execute(query, (airline)) data = cursor.fetchall() cursor.close() #Use only the top 5 sellers #Python will not break if we try to access a range that extends beyond the end of the array return render_template('viewAgentsSales.html', results = data[0:5], date=daterange) else: error = "Invalid Credentials" return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/viewAgents/commission') def viewAgentsCommission(): if authenticateStaff(): airline = getStaffAirline() cursor = conn.cursor() query = "select email,sum(flight.price)*0.1 as commission \ from booking_agent natural join purchases natural join ticket natural join flight \ where purchase_date >= date_sub(curdate(), interval 1 year) and airline_name=%s\ group by email order by commission DESC limit 5" cursor.execute(query, (airline)) data = cursor.fetchall() cursor.close() #Use only the top 5 sellers #Python will not break if we try to access a range that extends beyond the end of the array return render_template('viewAgentsCommission.html', results = data[0:5]) else: error = "Invalid Credentials" return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/viewCustomers') def viewCustomersPage(): if authenticateStaff(): airline = getStaffAirline() cursor = conn.cursor() query = 'select customer_email, count(ticket_id) as customerpurchases \ from purchases natural join ticket \ where airline_name= %s \ and purchase_date >= date_sub(curdate(), interval 1 year) group by customer_email \ having customerpurchases \ >= all (select count(ticket_id) \ from purchases natural join ticket \ where airline_name = %s \ and purchase_date >= date_sub(curdate(), interval 1 year) GROUP by customer_email)' cursor.execute(query, (airline, airline)) data = cursor.fetchall() cursor.close() error = request.args.get('error') return render_template('viewCustomers.html', error=error, results=data) else: error = "Invalid Credentials" return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/viewCustomers/results', methods=['POST']) def viewCustomers(): if authenticateStaff(): airline = getStaffAirline() customer = request.form['email'] cursor = conn.cursor() query1 = "select * from customer where email = %s" cursor.execute(query1,customer) data1 = cursor.fetchone() error = request.args.get('error') cursor.close() if not data1: error = "Not a customer email, please enter a customer email." return redirect(url_for('viewCustomersPage',error = error)) else: cursor = conn.cursor() query = 'select distinct flight_num from purchases natural join ticket where airline_name = %s and customer_email=%s' cursor.execute(query, (airline, customer)) data = cursor.fetchall() cursor.close() return render_template('viewCustomersResults.html', results=data, customer=customer) else: error = "Invalid Credentials" return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/viewReports') def viewReportsPage(): if authenticateStaff(): airline = getStaffAirline() currentmonth = datetime.datetime.now().month monthtickets = [] cursor = conn.cursor() for i in range(0, 12): query = 'select count(ticket_id) as sales \ from purchases natural join ticket \ where year(purchase_date) = year(curdate() - interval ' + str(i) + ' month) \ and month(purchase_date) = month(curdate() - interval ' + str(i) + ' month) \ and airline_name=%s' cursor.execute(query, (airline)) data = cursor.fetchall() salemonth = ((currentmonth - (i+1)) % 12) + 1 # print (data[0]['sales']) monthtickets.append([data[0]['sales'], salemonth]) cursor.close() c1 = ( Bar() .add_xaxis([d[1] for d in monthtickets]) .add_yaxis('total ticket number',[d[0] for d in monthtickets]) .set_global_opts(xaxis_opts=opts.AxisOpts(axislabel_opts=opts.LabelOpts(rotate=0)), title_opts=opts.TitleOpts(title="Ticket Amount in the Past", subtitle= "In the past 1 year"), legend_opts=opts.LegendOpts(pos_right="15%")) ) error = request.args.get('error') return render_template('viewReports.html', bar_options1=c1.dump_options(),error = error) else: error = "Invalid Credentials" return redirect(url_for('errorpage', error=error)) @app.route('/staffHome/viewReports/dates', methods=['POST']) def viewReportsDates(): if authenticateStaff(): airline = getStaffAirline() begintime = request.form['begintime'] endtime = request.form['endtime'] if not validateDates(begintime, endtime): error = 'Invalid date range' return redirect(url_for('viewReportsPage', error=error)) cursor = conn.cursor() query = 'select count(ticket_id) as sales \ from purchases natural join ticket where airline_name=%s\ and purchase_date between %s and %s' cursor.execute(query, (airline, begintime, endtime)) data = cursor.fetchall() cursor.close() return render_template('viewReportsDate.html', sales=data[0]['sales'], begintime=begintime, endtime=endtime) else: error = "Invalid Credentials" return render_template('error.html',error=error) @app.route('/staffHome/viewReports/past', methods=['POST']) def viewReportsPast(): if authenticateStaff(): airline = getStaffAirline() daterange = request.form['range'] cursor = conn.cursor() query = 'select count(ticket_id) as sales \ from purchases natural join ticket where airline_name=%s \ and purchase_date >= date_sub(curdate(), interval 1 ' + daterange + ')' cursor.execute(query, (airline)) data = cursor.fetchall() cursor.close() return render_template('viewReportsPast.html', sales=data[0]['sales'], datetime=daterange) else: error = "Invalid Credentials" return render_template('error.html',error=error) @app.route('/staffHome/ComparisonRevenue') def ComparisonRevenue(): if authenticateStaff(): username = session['username'] error = None # query for airline_name the staff works for cursor = conn.cursor() queryGetairline = "SELECT airline_name FROM airline_staff WHERE username= %s" cursor.execute(queryGetairline, username) airline_name = cursor.fetchone()['airline_name'] # query for direct purchase revenue (last month) query1 = "select sum(flight.price) as rev\ from purchases, ticket, flight\ where purchases.ticket_id = ticket.ticket_id \ and ticket.flight_num = flight.flight_num\ and ticket.airline_name = flight.airline_name\ and flight.airline_name = %s\ and purchases.purchase_date between DATE_SUB(curdate(), INTERVAL 1 MONTH) and curdate()\ and purchases.booking_agent_id is null" cursor.execute(query1,str(airline_name)) direct_revenue = cursor.fetchone()['rev'] # query for indirect purchase revenue (last month) query2 = "select sum(flight.price) as rev\ from purchases, ticket, flight\ where purchases.ticket_id = ticket.ticket_id \ and ticket.flight_num = flight.flight_num\ and ticket.airline_name = flight.airline_name\ and flight.airline_name = %s\ and purchases.purchase_date between DATE_SUB(curdate(), INTERVAL 1 MONTH) and curdate()\ and purchases.booking_agent_id is not null" cursor.execute(query2,str(airline_name)) indirect_revenue = cursor.fetchone()['rev'] #draw the pie chart (last month) x_data = ['Direct Revenue','Indirect Revenue'] y_data = [direct_revenue,indirect_revenue] data_pair = [list(z) for z in zip(x_data, y_data)] c1 = ( Pie() .add('',[d for d in data_pair]) .set_global_opts(title_opts=opts.TitleOpts(title="Revenue Comparison", subtitle = "Last Month"), legend_opts=opts.LegendOpts(pos_right="15%")) .set_series_opts(label_opts=opts.LabelOpts(formatter="{b}: {c}")) ) #Customized pie (a fancier version pie chart) # c1 = ( # Pie() # .add( # series_name="Revenue Source", # data_pair=data_pair, # rosetype="radius", # radius="55%", # center=["50%", "50%"], # label_opts=opts.LabelOpts(is_show=False, position="center"), # ) # .set_global_opts( # title_opts=opts.TitleOpts( # title="Revenue Source (last month)", # pos_left="center", # pos_top="20", # title_textstyle_opts=opts.TextStyleOpts(color="black"), # ), # legend_opts=opts.LegendOpts(is_show=False), # ) # .set_series_opts( # tooltip_opts=opts.TooltipOpts( # trigger="item", formatter="{a} <br/>{b}: {c} ({d}%)" # ), # label_opts=opts.LabelOpts(color="rgba(0,0,0,255)"), # ) # ) # query for direct purchase revenue (last year) query1_ = "select sum(flight.price) as rev\ from purchases, ticket, flight\ where purchases.ticket_id = ticket.ticket_id \ and ticket.flight_num = flight.flight_num\ and ticket.airline_name = flight.airline_name\ and flight.airline_name = %s\ and purchases.purchase_date between DATE_SUB(curdate(), INTERVAL 1 YEAR) and curdate()\ and purchases.booking_agent_id is null" cursor.execute(query1_,str(airline_name)) direct_revenue_ = cursor.fetchone()['rev'] # query for indirect purchase revenue (last month) query2_ = "select sum(flight.price) as rev\ from purchases, ticket, flight\ where purchases.ticket_id = ticket.ticket_id \ and ticket.flight_num = flight.flight_num\ and ticket.airline_name = flight.airline_name\ and flight.airline_name = %s\ and purchases.purchase_date between DATE_SUB(curdate(), INTERVAL 1 YEAR) and curdate()\ and purchases.booking_agent_id is not null" cursor.execute(query2_,str(airline_name)) indirect_revenue_ = cursor.fetchone()['rev'] cursor.close() #draw the pie chart (last month) x_data_ = ['Direct Revenue','Indirect Revenue'] y_data_ = [direct_revenue_,indirect_revenue_] data_pair_ = [list(z) for z in zip(x_data_, y_data_)] c2 = ( Pie() .add('',[d for d in data_pair_]) .set_global_opts(title_opts=opts.TitleOpts(title="Revenue Comparison", subtitle = "Last Year"), legend_opts=opts.LegendOpts(pos_right="15%")) .set_series_opts(label_opts=opts.LabelOpts(formatter="{b}: {c}")) ) if direct_revenue and indirect_revenue: return render_template('ComparisonRevenue.html', pie_options1 = c1.dump_options(), pie_options2 = c2.dump_options()) else: error = 'Sorry! No data available Right Now.' return render_template('ComparisonRevenue.html',error = error) else: error = "Invalid Credentials" return render_template('error.html',error=error)
31,149
8,791
from fastapi import FastAPI try: from routes import analytics,templates,auth except: from .routes import analytics,templates,auth app = FastAPI() app.include_router(analytics.router) app.include_router(templates.router) app.include_router(auth.authr)
261
80
from __future__ import unicode_literals from __future__ import print_function from __future__ import division from __future__ import absolute_import from builtins import str from future import standard_library standard_library.install_aliases() import query_releases import unittest class TestQueryReleases(unittest.TestCase): def test_parse_url(self): """ Tests the parse_url function """ base_url = "https://test.url" token1, url1 = query_releases.parse_url(base_url) self.assertIsNone(token1) self.assertEqual(url1, base_url) test_url_2 = "https://token@test.url" # 2nd test URL is based on the base_url token2, url2 = query_releases.parse_url(test_url_2) self.assertEqual(token2, "token") self.assertEqual(url2, base_url) if __name__ == "__main__": unittest.main()
873
277
# simplePendulum.py # Model of a simple pendulum import math def simplePendulum(length = 1, angle = math.pi/4, angular_velocity = 0, DT = 0.0001, simLength = 12): numIterations = int(simLength/DT) + 1 g = 9.81 angle_change = angular_velocity angular_acceleration = -g * math.sin(angle)/length t = 0 timeLst = [0] angleLst = [angle] angular_velocityLst = [angular_velocity] angular_accelerationLst = [angular_acceleration] for i in range(1, numIterations): t = i * DT timeLst.append(t) angle = angle + (angle_change) * DT angleLst.append(angle) angular_velocity = angular_velocity + (angular_acceleration) * DT angular_velocityLst.append(angular_velocity) angle_change = angular_velocity angular_acceleration = -g * math.sin(angle)/length angular_accelerationLst.append(angular_acceleration) return timeLst, angleLst, angular_velocityLst, angular_accelerationLst
1,211
373
from setuptools import setup setup( name='dropboxbackup', version='0.1', py_modules=['dropboxbackup'], install_requires=[ 'click', 'dropbox', 'simple-crypt' ], entry_points=''' [console_scripts] dropboxbackup=dropboxbackup:cli ''', )
303
104
from os import path from shutil import copyfile import tvm from tvm import relay from tvm.driver import tvmc from tvm.driver.tvmc.model import TVMCModel from tvm.relay.transform import InferType, ToMixedPrecision """Copy pasted mostly from: https://github.com/AndrewZhaoLuo/TVM-Sandbox/blob/bb209e8845440ed9f40af1b2580618196c939745/fp16_pass/benchmark_fp16.py#L1 Creates centerface autoscheduler log files, which are included in this repo so you don't have to spend 24 hrs running the tuning script! Run on a 2020, 13-inch macbook pro (m1 mac) FP32: Processing centerface_autoscheduler_30000kt_fp32_llvm Execution time summary: mean (ms) median (ms) max (ms) min (ms) std (ms) 33.8869 33.6213 35.0154 33.1292 0.7192 Output Names: ['output_0', 'output_1', 'output_2', 'output_3'] FP16: Processing centerface_autoscheduler_10000kt_fp16_llvm Execution time summary: mean (ms) median (ms) max (ms) min (ms) std (ms) 22.3274 22.2959 23.4356 21.7442 0.4560 Output Names: ['output_0', 'output_1', 'output_2', 'output_3'] """ def load_model(name, **kwargs): return tvmc.load(path.join("models", name), **kwargs) def graph_optimize( tvmc_model, run_fp16_pass, run_other_opts, try_nhwc_layout=False, fast_math=True ): mod, params = tvmc_model.mod, tvmc_model.params mod = tvm.IRModule.from_expr(mod["main"]) # nhwc is typically better for autoscheduler -- more schedules available # also winograd is only available for nhwc if try_nhwc_layout: desired_layouts = { "nn.conv2d": ["NHWC", "default"], "nn.conv2d_transpose": ["NHWC", "default"], "nn.upsampling": ["NHWC", "default"], "image.resize2d": ["NHWC", "default"], "vision.roi_align": ["NHWC", "default"], } with tvm.transform.PassContext( opt_level=3, config={"relay.backend.use_auto_scheduler": True} ): mod = relay.transform.InferType()(mod) mod = relay.transform.ConvertLayout(desired_layouts)(mod) mod = tvm.relay.transform.EliminateCommonSubexpr()(mod) mod = tvm.relay.transform.FoldConstant()(mod) if run_other_opts: mod = tvm.relay.transform.FastMath()(mod) if fast_math else mod mod = tvm.relay.transform.EliminateCommonSubexpr()(mod) BindPass = tvm.relay.transform.function_pass( lambda fn, new_mod, ctx: tvm.relay.build_module.bind_params_by_name( fn, params ), opt_level=1, ) mod = BindPass(mod) mod = tvm.relay.transform.FoldConstant()(mod) mod = tvm.relay.transform.CombineParallelBatchMatmul()(mod) mod = tvm.relay.transform.FoldConstant()(mod) if run_fp16_pass: mod = InferType()(mod) mod = ToMixedPrecision()(mod) if run_other_opts and run_fp16_pass: # run one more pass to clean up new subgraph mod = tvm.relay.transform.EliminateCommonSubexpr()(mod) mod = tvm.relay.transform.FoldConstant()(mod) mod = tvm.relay.transform.CombineParallelBatchMatmul()(mod) mod = tvm.relay.transform.FoldConstant()(mod) mod = tvm.relay.transform.FastMath()(mod) if fast_math else mod return TVMCModel(mod, params) def benchmark_model( model_func, name, run_fp16_pass=True, run_other_opts=True, enable_autoscheduler=False, try_nhwc_layout=False, target="llvm", target_host="llvm", tuning_trials=10000, tuning_repeat_trials=5, measure_number=100, measure_repeats=10, ): print("*" * 30, name, "*" * 30) print("FP16 pass" if run_fp16_pass else "FP32 pass") """Get Module""" tvmc_model = model_func( run_pass=run_fp16_pass, run_opts=run_other_opts, try_nhwc_layout=try_nhwc_layout ) print("Final relay mod:") print(tvmc_model.mod) tuning_records = tvmc.tune( tvmc_model, target=target, enable_autoscheduler=enable_autoscheduler, trials=tuning_trials, repeat=tuning_repeat_trials, tuner="xgb_knob", target_host=target_host, ) copyfile(tuning_records, f"tuning_logs/{name}.log") # Create package artifacts package = tvmc.compile(tvmc_model, target=target, tuning_records=tuning_records) result = tvmc.run( package, device="cpu" if "llvm" in target else target, repeat=measure_number, number=measure_repeats, ) print(result) print() def get_centerface(run_pass=True, run_opts=True, try_nhwc_layout=False): tvmc_model = load_model("centerface.onnx") return graph_optimize( tvmc_model, run_pass, run_opts, try_nhwc_layout=try_nhwc_layout ) def benchmark_and_compile_so_and_whl( model_func, name, run_fp16_pass=True, run_other_opts=True, try_nhwc_layout=False, target="llvm", ): print(f"Processing {name}") tvmc_model = model_func( run_pass=run_fp16_pass, run_opts=run_other_opts, try_nhwc_layout=try_nhwc_layout ) tuning_records = f"tuning_logs/{name}.log" package = tvmc.compile(tvmc_model, target=target, tuning_records=tuning_records) result = tvmc.run( package, device="cpu" if "llvm" in target else target, repeat=10, number=100, ) print(result) print() copyfile(package.package_path, f"compiled_packages/{name}.tar") if __name__ == "__main__": benchmark_model( get_centerface, "centerface_autoscheduler_30000kt_fp16_llvm", run_fp16_pass=True, run_other_opts=True, enable_autoscheduler=True, try_nhwc_layout=True, tuning_trials=30000, target="llvm -mcpu=apple-latest -mtriple=arm64-apple-macos", target_host="llvm -mcpu=apple-latest -mtriple=arm64-apple-macos", ) benchmark_model( get_centerface, "centerface_autoscheduler_30000kt_fp32_llvm", run_fp16_pass=False, run_other_opts=True, enable_autoscheduler=True, try_nhwc_layout=True, tuning_trials=30000, target="llvm -mcpu=apple-latest -mtriple=arm64-apple-macos", target_host="llvm -mcpu=apple-latest -mtriple=arm64-apple-macos", ) benchmark_and_compile_so_and_whl( get_centerface, "centerface_autoscheduler_30000kt_fp16_llvm", run_fp16_pass=True, run_other_opts=True, try_nhwc_layout=True, target="llvm -mcpu=apple-latest -mtriple=arm64-apple-macos", ) benchmark_and_compile_so_and_whl( get_centerface, "centerface_autoscheduler_30000kt_fp32_llvm", run_fp16_pass=False, run_other_opts=True, try_nhwc_layout=True, target="llvm -mcpu=apple-latest -mtriple=arm64-apple-macos", )
6,896
2,628
# -*- coding: utf-8 -*- """ XJB Generate Images Module (/doodle) Created on Sun Sep 1 16:03:16 2019 @author: user """ import os import asyncio import uuid import tg_connection gen_path = "D:/AndroidProjects/ScarletKindom/flandre-generator/wgan/sample.png" inp_base = "D:/AndroidProjects/ScarletKindom/flandre-generator/wgan/" sketchr_query = set() def get_pf_path(): pf = str(uuid.uuid4()) return inp_base + "sketch-input-" + pf, inp_base + "sketch-output-" + pf + ".txt" async def command_doodle(session, chat_id): while not os.path.exists(gen_path): await asyncio.sleep(0.6) with open(gen_path, "rb") as fi: png_bytes = fi.read(-1) os.unlink(gen_path) await tg_connection.send_photo_by_contents(session, chat_id, png_bytes) async def command_sketchr(session, chat_id, from_user): sketchr_query.add(from_user) await tg_connection.send_msg(session, chat_id, "Send me a sketch!") async def command_sketchr_clbk(session, chat_id, frid, msg): if frid in sketchr_query: sketchr_query.remove(frid) else: return if "photo" in msg: fid = msg["photo"][0]["file_id"] b = await tg_connection.get_file_contents(session, fid) fi, fo = get_pf_path() with open(fi, "wb") as writer: writer.write(b) while os.path.exists(fi): await asyncio.sleep(0.6) with open(fo, "rt") as reader: txt = reader.read(-1) await tg_connection.send_msg(session, chat_id, txt) os.unlink(fo)
1,547
587
from rest_framework import viewsets, permissions from leads.serializers import LeadSerializer from leads.models import Lead class LeadViewSet(viewsets.ModelViewSet): """Manage CRUD operations for Leads in the database""" queryset = Lead.objects.all() permission_classes = [ permissions.AllowAny ] serializer_class = LeadSerializer
362
100
from flask import render_template from . import main @main.app_errorhandler(404) def fo_O_fo(error): """ Function to render the 404 error page """ return render_template('fo_O_fo.html'), 404
207
73
from django.contrib.auth.models import AbstractUser from django.db import models class User(AbstractUser): followers = models.ManyToManyField('self',symmetrical=False,related_name='following') class Post(models.Model): username = models.ForeignKey(User,on_delete=models.CASCADE,related_name='posts') content = models.TextField() datetime = models.DateTimeField() likes = models.ManyToManyField(User)
408
124
import PIL.Image as Img import numpy as np from tqdm.notebook import tqdm from PIL import ImageFilter import tables import time import gc """ all the insert/append function for collage generator _canvas_append takes the inserting operation, the rest are finding add_point logic """ class _insertion(): def _canvas_append(self, canvas: np.ndarray, add_point: np.ndarray, img: np.ndarray, mask: np.ndarray = None, mask_label: int = None, mode = "label", format = "pixel"): """ the actual working part, add a image to a canvas Args: canvas: np.ndarray, 3-channel canvas add_point: tuple of int, the topleft point of the image to be added img: np.ndarray, 3-channel, the vignette to be added mask: np.ndarray(if it's there), 1-channel/4-channels, the mask with the canvas mask_label: int/2d np.ndarray/4 channels np.ndarray, the value of this label onto the mask mode: str, "label" or "pattern", how the mask be overwritten, if "label", it will use the int mask_label if "pattern", it will copy the np.ndarray passed to mask_label format: str, "pixel" or "COCO", how the mask will be updates by new vignettes in "pixel", each individual mask will be saved on the same dimension if "COCO", each individual mask will be saved by a different color on a 3-channel mask Return: canvas: np.ndarray of 3 channels, the canvas with img added. mask: if format is "pixel" np.ndarray of 1 channel, the mask with img's label added. if format is "COCO", np.ndarray of 4-channels, the mask with img's label added. """ assert format in ["pixel", "COCO"] # if there's no mask (preview/background) if type(mask) != np.ndarray: # add img to canvas, if there's any overlap, skip it np.add(canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], img, where = (canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]] == 0), out = canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting = "unsafe") # return canvas return canvas #if there's a mask, from the logic of the functions below, #we are going to direcly add these value to a 0-filled canvas and mask else: if format == "pixel": # add image to canvas np.add(canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], img, out = canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting = "unsafe") # add label to mask if mode == "label": # if in label mode, we are adding this label int value to all nonzero space np.add(mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], mask_label*np.any(img, axis = 2), out = mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting = "unsafe") else: #else we are adding a pattern, copy the while pattern np.add(mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], mask_label, out = mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting = "unsafe") return canvas, mask # if we are building a COCO mode if format == "COCO": if mode == "label": # generate a new color for this object _new_color, self.existed_color = self._generate_new_color(self.existed_color) self.color_dict[str(tuple(_new_color.tolist()))] = mask_label # we have COCO format use as following, first layer will work as the pixel mask, # and the rest will following, the first layer will be removed when converted to COCO if mask.ndim == 2: # if the mask only have one layer, it must be the start mask # add 3 new layers as the RGB recording mask = np.stack((mask, np.zeros_like(mask),np.zeros_like(mask),np.zeros_like(mask)), axis = -1) # add image to canvas, add different label to different channel of the mask np.add(canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], img, out = canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting="unsafe") np.add(mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1],0], mask_label*np.any(img, axis = 2), out = mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1],0], casting="unsafe") np.add(mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1],1:4], np.any(img, axis = 2, keepdims = True)*_new_color, mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1],1:4], casting="unsafe") return canvas, mask else: if mask.ndim == 2: mask = np.stack((mask, np.zeros_like(mask),np.zeros_like(mask),np.zeros_like(mask)), axis = -1) np.add(canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], img, out = canvas[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting="unsafe") np.add(mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], mask_label, out = mask[add_point[0]:add_point[0]+img.shape[0], add_point[1]:add_point[1]+img.shape[1]], casting="unsafe") return canvas, mask def _init_insert(self, img: np.ndarray, canvas: np.ndarray, mask: np.ndarray, label: int, mode = "pattern", format = "pixel"): """ find a random legal position in canvas, append img to canvas and mask Args: img: np.ndarray of 3 channels, the vignette to be added canvas: np.ndarray of 3 channels, the canvas mask: 2d/4-channel np.ndarray, the mask label: the label to be added mode: str, "label" or "pattern", see mode in _canvas_append() format: str, see format in _canvas_append() Return: canvas: np.ndarray of 3 channels, the canvas with img added. mask: np.ndarray of 1 channel, the mask with img's label added. """ _outer_bound = (canvas.shape[0] - img.shape[0], canvas.shape[1] - img.shape[1]) # select an initial add_point _add_point = np.array((np.random.randint(low = self._scanning_constant, high = _outer_bound[0] - self._scanning_constant), np.random.randint(low = self._scanning_constant, high = _outer_bound[1] - self._scanning_constant))) # create a binary mask of the img _img_mask = np.any(img, axis = 2) # directly use the _add_point canvas, mask = self._canvas_append(canvas = canvas, add_point = _add_point, img = img, mask = mask, mask_label = label, mode = mode, format = format) return canvas, mask def _secondary_insert(self, img: np.ndarray, canvas: np.ndarray, mask: np.ndarray, label: int, patience: int, mode = "label", format = "pixel"): """ find a random non-overlapping position in canvas, append img to canvas and mask Args: img: np.ndarray of 3 channels, the vignette to be added canvas: np.ndarray of 3 channels, the canvas mask: 2d/4-channel np.ndarray, the mask label: the label to be added patience: int, the retry time for finding non-overlapping position mode: str, "label" or "pattern", see mode in _canvas_append() format: str, see format in _canvas_append() Return: canvas: np.ndarray of 3 channels, the canvas with img added, if the tries in {patience} times succssfully added the img onto canvas, otherwise the original canvas is returned mask: np.ndarray of 2d or 4 channels, the mask with img added, if the tries in {patience} times succssfully added the img onto canvas, otherwise the original mask if returned """ _outer_bound = (canvas.shape[0] - img.shape[0], canvas.shape[1] - img.shape[1]) # select an initial add_point _add_point = np.array((np.random.randint( low = self._scanning_constant, high = _outer_bound[0] - self._scanning_constant), np.random.randint( low = self._scanning_constant, high = _outer_bound[1] - self._scanning_constant) )) # create a binary mask of the img _img_mask = np.any(img, axis = 2) for retry in range(patience): # for each time make a small move _add_point = _add_point + np.random.randint( low = -1*self._scanning_constant, high = self._scanning_constant, size = 2) # make sure the new value is legal _add_point = np.clip(a = _add_point, a_min = (0,0), a_max = _outer_bound) # check if there's any overlap # in pixel format check the mask directly if format == "pixel": _check_zone = mask[_add_point[0]:_add_point[0]+_img_mask.shape[0], _add_point[1]:_add_point[1]+_img_mask.shape[1]] # in COCO format check the first layer of mask else: _check_zone = mask[_add_point[0]:_add_point[0]+_img_mask.shape[0], _add_point[1]:_add_point[1]+_img_mask.shape[1], 0] # if so if np.any(np.multiply(_check_zone,_img_mask)) == True: #retry for a new point continue # otherwise add the img to canvas and mask and stop retry else: canvas, mask = self._canvas_append(canvas = canvas, add_point = _add_point, img = img, mask = mask, mask_label = label, mode = mode, format = format) break gc.collect() return canvas, mask def _try_insert(self, img: np.ndarray, canvas: np.ndarray, mask: np.ndarray, label: int, patience: int, mode = "label", format = "pixel"): """ try to insert img into canvas and mask using a escape-overlapping algorithm if the initial point is overlapping, try to "escape" the overlapping and append at the first position successfuly escape if the initial point is not overlapping, try to find a overlapping point and append at the last non-overlapping point before this one Args: img: np.ndarray of 3 channels, the vignette to be added canvas: np.ndarray of 3 channels, the canvas mask: 2d/4-channel np.ndarray, the mask label: the label to be added patience: int, the retry time for finding non-overlapping position mode: str, "label" or "pattern", see mode in _canvas_append() format: str, see format in _canvas_append() Return: canvas: np.ndarray of 3 channels, the canvas with img added, if the tries in {patience} times succssfully added the img onto canvas, otherwise the original canvas is returned mask: np.ndarray of 2d or 4 channels, the mask with img added, if the tries in {patience} times succssfully added the img onto canvas, otherwise the original mask if returned """ _outer_bound = (canvas.shape[0] - img.shape[0], canvas.shape[1] - img.shape[1]) # select an initial add_point _add_point = np.array((np.random.randint(low = self._scanning_constant, high = _outer_bound[0] - self._scanning_constant), np.random.randint(low = self._scanning_constant, high = _outer_bound[1] - self._scanning_constant))) # create a binary mask of the img _img_mask = np.any(img, axis = 2) # check if there's any overlap if format == "pixel": _check_zone = mask[_add_point[0]:_add_point[0]+_img_mask.shape[0], _add_point[1]:_add_point[1]+_img_mask.shape[1]] # in COCO format check the first layer of mask else: _check_zone = mask[_add_point[0]:_add_point[0]+_img_mask.shape[0], _add_point[1]:_add_point[1]+_img_mask.shape[1], 0] # if we start with an overlap, we need to escape from overlap, otherwise we need to find a overlap _init_overlapped = np.any(np.multiply(_check_zone,_img_mask)) # if we are in a finding mode and need to record the last add point _last_add_point = _add_point # in the patience time for retry in range(patience): # for each time make a small move _add_point = _add_point + np.random.randint(low = -1*self._scanning_constant, high = self._scanning_constant, size = 2) # make sure the new value is legal _add_point = np.clip(a = _add_point, a_min = (0,0), a_max = _outer_bound) # check if there's any overlap if format == "pixel": _check_zone = mask[_add_point[0]:_add_point[0]+_img_mask.shape[0], _add_point[1]:_add_point[1]+_img_mask.shape[1]] # in COCO format check the first layer of mask else: _check_zone = mask[_add_point[0]:_add_point[0]+_img_mask.shape[0], _add_point[1]:_add_point[1]+_img_mask.shape[1], 0] # check if there's overlap _overlap = np.any(np.multiply(_check_zone,_img_mask)) # if we had a overlap in "escaping" if (_overlap == True) and (_init_overlapped == True): #retry for a new point continue # if we met the first non-overlap while escaping elif (_overlap == False) and (_init_overlapped == True): #stop the finding canvas, mask = self._canvas_append(canvas = canvas, add_point = _add_point, img = img, mask = mask, mask_label = label, mode = mode, format = format) break # if we are finding but not found elif (_overlap == False) and (_init_overlapped == False): #record last add_point and retry for a new point _last_add_point = _add_point continue # or we are finding a overlap and found it, we need to use the last else: canvas, mask = self._canvas_append(canvas = canvas, add_point = _last_add_point, img = img, mask = mask, mask_label = label, mode = mode, format = format) break gc.collect() return canvas, mask
18,476
4,902
# coding: utf-8 """ OpenShift API (with Kubernetes) OpenShift provides builds, application lifecycle, image content management, and administrative policy on top of Kubernetes. The API allows consistent management of those objects. All API operations are authenticated via an Authorization bearer token that is provided for service accounts as a generated secret (in JWT form) or via the native OAuth endpoint located at /oauth/authorize. Core infrastructure components may use client certificates that require no authentication. All API operations return a 'resourceVersion' string that represents the version of the object in the underlying storage. The standard LIST operation performs a snapshot read of the underlying objects, returning a resourceVersion representing a consistent version of the listed objects. The WATCH operation allows all updates to a set of objects after the provided resourceVersion to be observed by a client. By listing and beginning a watch from the returned resourceVersion, clients may observe a consistent view of the state of one or more objects. Note that WATCH always returns the update after the provided resourceVersion. Watch may be extended a limited time in the past - using etcd 2 the watch window is 1000 events (which on a large cluster may only be a few tens of seconds) so clients must explicitly handle the \"watch to old error\" by re-listing. Objects are divided into two rough categories - those that have a lifecycle and must reflect the state of the cluster, and those that have no state. Objects with lifecycle typically have three main sections: * 'metadata' common to all objects * a 'spec' that represents the desired state * a 'status' that represents how much of the desired state is reflected on the cluster at the current time Objects that have no state have 'metadata' but may lack a 'spec' or 'status' section. Objects are divided into those that are namespace scoped (only exist inside of a namespace) and those that are cluster scoped (exist outside of a namespace). A namespace scoped resource will be deleted when the namespace is deleted and cannot be created if the namespace has not yet been created or is in the process of deletion. Cluster scoped resources are typically only accessible to admins - resources like nodes, persistent volumes, and cluster policy. All objects have a schema that is a combination of the 'kind' and 'apiVersion' fields. This schema is additive only for any given version - no backwards incompatible changes are allowed without incrementing the apiVersion. The server will return and accept a number of standard responses that share a common schema - for instance, the common error type is 'metav1.Status' (described below) and will be returned on any error from the API server. The API is available in multiple serialization formats - the default is JSON (Accept: application/json and Content-Type: application/json) but clients may also use YAML (application/yaml) or the native Protobuf schema (application/vnd.kubernetes.protobuf). Note that the format of the WATCH API call is slightly different - for JSON it returns newline delimited objects while for Protobuf it returns length-delimited frames (4 bytes in network-order) that contain a 'versioned.Watch' Protobuf object. See the OpenShift documentation at https://docs.openshift.org for more information. OpenAPI spec version: latest Generated by: https://github.com/swagger-api/swagger-codegen.git """ from __future__ import absolute_import import sys import os import re # python 2 and python 3 compatibility library from six import iteritems from ..api_client import ApiClient class BuildOpenshiftIoV1Api(object): """ NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. Ref: https://github.com/swagger-api/swagger-codegen """ def __init__(self, api_client=None): if api_client is None: api_client = ApiClient() self.api_client = api_client def connect_post_namespaced_build_config_instantiatebinary(self, name, namespace, **kwargs): """ connect POST requests to instantiatebinary of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.connect_post_namespaced_build_config_instantiatebinary(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BinaryBuildRequestOptions (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str as_file: asFile determines if the binary should be created as a file within the source rather than extracted as an archive :param str revision_author_email: revision.authorEmail of the source control user :param str revision_author_name: revision.authorName of the source control user :param str revision_commit: revision.commit is the value identifying a specific commit :param str revision_committer_email: revision.committerEmail of the source control user :param str revision_committer_name: revision.committerName of the source control user :param str revision_message: revision.message is the description of a specific commit :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.connect_post_namespaced_build_config_instantiatebinary_with_http_info(name, namespace, **kwargs) else: (data) = self.connect_post_namespaced_build_config_instantiatebinary_with_http_info(name, namespace, **kwargs) return data def connect_post_namespaced_build_config_instantiatebinary_with_http_info(self, name, namespace, **kwargs): """ connect POST requests to instantiatebinary of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.connect_post_namespaced_build_config_instantiatebinary_with_http_info(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BinaryBuildRequestOptions (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str as_file: asFile determines if the binary should be created as a file within the source rather than extracted as an archive :param str revision_author_email: revision.authorEmail of the source control user :param str revision_author_name: revision.authorName of the source control user :param str revision_commit: revision.commit is the value identifying a specific commit :param str revision_committer_email: revision.committerEmail of the source control user :param str revision_committer_name: revision.committerName of the source control user :param str revision_message: revision.message is the description of a specific commit :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'as_file', 'revision_author_email', 'revision_author_name', 'revision_commit', 'revision_committer_email', 'revision_committer_name', 'revision_message'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method connect_post_namespaced_build_config_instantiatebinary" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `connect_post_namespaced_build_config_instantiatebinary`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `connect_post_namespaced_build_config_instantiatebinary`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'as_file' in params: query_params.append(('asFile', params['as_file'])) if 'revision_author_email' in params: query_params.append(('revision.authorEmail', params['revision_author_email'])) if 'revision_author_name' in params: query_params.append(('revision.authorName', params['revision_author_name'])) if 'revision_commit' in params: query_params.append(('revision.commit', params['revision_commit'])) if 'revision_committer_email' in params: query_params.append(('revision.committerEmail', params['revision_committer_email'])) if 'revision_committer_name' in params: query_params.append(('revision.committerName', params['revision_committer_name'])) if 'revision_message' in params: query_params.append(('revision.message', params['revision_message'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['*/*']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}/instantiatebinary', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def connect_post_namespaced_build_config_webhooks(self, name, namespace, **kwargs): """ connect POST requests to webhooks of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.connect_post_namespaced_build_config_webhooks(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str path: Path is the URL path to use for the current proxy request to pod. :return: str If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.connect_post_namespaced_build_config_webhooks_with_http_info(name, namespace, **kwargs) else: (data) = self.connect_post_namespaced_build_config_webhooks_with_http_info(name, namespace, **kwargs) return data def connect_post_namespaced_build_config_webhooks_with_http_info(self, name, namespace, **kwargs): """ connect POST requests to webhooks of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.connect_post_namespaced_build_config_webhooks_with_http_info(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str path: Path is the URL path to use for the current proxy request to pod. :return: str If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'path'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method connect_post_namespaced_build_config_webhooks" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `connect_post_namespaced_build_config_webhooks`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `connect_post_namespaced_build_config_webhooks`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'path' in params: query_params.append(('path', params['path'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['*/*']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}/webhooks', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='str', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def connect_post_namespaced_build_config_webhooks_with_path(self, name, namespace, path, **kwargs): """ connect POST requests to webhooks of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.connect_post_namespaced_build_config_webhooks_with_path(name, namespace, path, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str path: path to the resource (required) :param str path2: Path is the URL path to use for the current proxy request to pod. :return: str If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.connect_post_namespaced_build_config_webhooks_with_path_with_http_info(name, namespace, path, **kwargs) else: (data) = self.connect_post_namespaced_build_config_webhooks_with_path_with_http_info(name, namespace, path, **kwargs) return data def connect_post_namespaced_build_config_webhooks_with_path_with_http_info(self, name, namespace, path, **kwargs): """ connect POST requests to webhooks of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.connect_post_namespaced_build_config_webhooks_with_path_with_http_info(name, namespace, path, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str path: path to the resource (required) :param str path2: Path is the URL path to use for the current proxy request to pod. :return: str If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'path', 'path2'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method connect_post_namespaced_build_config_webhooks_with_path" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `connect_post_namespaced_build_config_webhooks_with_path`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `connect_post_namespaced_build_config_webhooks_with_path`") # verify the required parameter 'path' is set if ('path' not in params) or (params['path'] is None): raise ValueError("Missing the required parameter `path` when calling `connect_post_namespaced_build_config_webhooks_with_path`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] if 'path' in params: path_params['path'] = params['path'] query_params = [] if 'path2' in params: query_params.append(('path', params['path2'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['*/*']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}/webhooks/{path}', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='str', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def create_build_config_for_all_namespaces(self, body, **kwargs): """ create a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_build_config_for_all_namespaces(body, async_req=True) >>> result = thread.get() :param async_req bool :param V1BuildConfig body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.create_build_config_for_all_namespaces_with_http_info(body, **kwargs) else: (data) = self.create_build_config_for_all_namespaces_with_http_info(body, **kwargs) return data def create_build_config_for_all_namespaces_with_http_info(self, body, **kwargs): """ create a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_build_config_for_all_namespaces_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool :param V1BuildConfig body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ all_params = ['body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method create_build_config_for_all_namespaces" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `create_build_config_for_all_namespaces`") collection_formats = {} path_params = {} query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/buildconfigs', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfig', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def create_build_for_all_namespaces(self, body, **kwargs): """ create a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_build_for_all_namespaces(body, async_req=True) >>> result = thread.get() :param async_req bool :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.create_build_for_all_namespaces_with_http_info(body, **kwargs) else: (data) = self.create_build_for_all_namespaces_with_http_info(body, **kwargs) return data def create_build_for_all_namespaces_with_http_info(self, body, **kwargs): """ create a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_build_for_all_namespaces_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method create_build_for_all_namespaces" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `create_build_for_all_namespaces`") collection_formats = {} path_params = {} query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/builds', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def create_namespaced_build(self, namespace, body, **kwargs): """ create a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build(namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.create_namespaced_build_with_http_info(namespace, body, **kwargs) else: (data) = self.create_namespaced_build_with_http_info(namespace, body, **kwargs) return data def create_namespaced_build_with_http_info(self, namespace, body, **kwargs): """ create a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_with_http_info(namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method create_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `create_namespaced_build`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `create_namespaced_build`") collection_formats = {} path_params = {} if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def create_namespaced_build_clone(self, name, namespace, body, **kwargs): """ create clone of a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_clone(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildRequest (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildRequest body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildRequest If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.create_namespaced_build_clone_with_http_info(name, namespace, body, **kwargs) else: (data) = self.create_namespaced_build_clone_with_http_info(name, namespace, body, **kwargs) return data def create_namespaced_build_clone_with_http_info(self, name, namespace, body, **kwargs): """ create clone of a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_clone_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildRequest (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildRequest body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildRequest If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method create_namespaced_build_clone" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `create_namespaced_build_clone`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `create_namespaced_build_clone`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `create_namespaced_build_clone`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}/clone', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildRequest', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def create_namespaced_build_config(self, namespace, body, **kwargs): """ create a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_config(namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildConfig body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.create_namespaced_build_config_with_http_info(namespace, body, **kwargs) else: (data) = self.create_namespaced_build_config_with_http_info(namespace, body, **kwargs) return data def create_namespaced_build_config_with_http_info(self, namespace, body, **kwargs): """ create a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_config_with_http_info(namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildConfig body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ all_params = ['namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method create_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `create_namespaced_build_config`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `create_namespaced_build_config`") collection_formats = {} path_params = {} if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfig', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def create_namespaced_build_config_instantiate(self, name, namespace, body, **kwargs): """ create instantiate of a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_config_instantiate(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildRequest (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildRequest body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.create_namespaced_build_config_instantiate_with_http_info(name, namespace, body, **kwargs) else: (data) = self.create_namespaced_build_config_instantiate_with_http_info(name, namespace, body, **kwargs) return data def create_namespaced_build_config_instantiate_with_http_info(self, name, namespace, body, **kwargs): """ create instantiate of a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_namespaced_build_config_instantiate_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildRequest (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildRequest body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method create_namespaced_build_config_instantiate" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `create_namespaced_build_config_instantiate`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `create_namespaced_build_config_instantiate`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `create_namespaced_build_config_instantiate`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}/instantiate', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def delete_collection_namespaced_build(self, namespace, **kwargs): """ delete collection of Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_collection_namespaced_build(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1Status If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.delete_collection_namespaced_build_with_http_info(namespace, **kwargs) else: (data) = self.delete_collection_namespaced_build_with_http_info(namespace, **kwargs) return data def delete_collection_namespaced_build_with_http_info(self, namespace, **kwargs): """ delete collection of Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_collection_namespaced_build_with_http_info(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1Status If the method is called asynchronously, returns the request thread. """ all_params = ['namespace', 'pretty', '_continue', 'field_selector', 'include_uninitialized', 'label_selector', 'limit', 'resource_version', 'timeout_seconds', 'watch'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method delete_collection_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `delete_collection_namespaced_build`") collection_formats = {} path_params = {} if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if '_continue' in params: query_params.append(('continue', params['_continue'])) if 'field_selector' in params: query_params.append(('fieldSelector', params['field_selector'])) if 'include_uninitialized' in params: query_params.append(('includeUninitialized', params['include_uninitialized'])) if 'label_selector' in params: query_params.append(('labelSelector', params['label_selector'])) if 'limit' in params: query_params.append(('limit', params['limit'])) if 'resource_version' in params: query_params.append(('resourceVersion', params['resource_version'])) if 'timeout_seconds' in params: query_params.append(('timeoutSeconds', params['timeout_seconds'])) if 'watch' in params: query_params.append(('watch', params['watch'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Status', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def delete_collection_namespaced_build_config(self, namespace, **kwargs): """ delete collection of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_collection_namespaced_build_config(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1Status If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.delete_collection_namespaced_build_config_with_http_info(namespace, **kwargs) else: (data) = self.delete_collection_namespaced_build_config_with_http_info(namespace, **kwargs) return data def delete_collection_namespaced_build_config_with_http_info(self, namespace, **kwargs): """ delete collection of BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_collection_namespaced_build_config_with_http_info(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1Status If the method is called asynchronously, returns the request thread. """ all_params = ['namespace', 'pretty', '_continue', 'field_selector', 'include_uninitialized', 'label_selector', 'limit', 'resource_version', 'timeout_seconds', 'watch'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method delete_collection_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `delete_collection_namespaced_build_config`") collection_formats = {} path_params = {} if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if '_continue' in params: query_params.append(('continue', params['_continue'])) if 'field_selector' in params: query_params.append(('fieldSelector', params['field_selector'])) if 'include_uninitialized' in params: query_params.append(('includeUninitialized', params['include_uninitialized'])) if 'label_selector' in params: query_params.append(('labelSelector', params['label_selector'])) if 'limit' in params: query_params.append(('limit', params['limit'])) if 'resource_version' in params: query_params.append(('resourceVersion', params['resource_version'])) if 'timeout_seconds' in params: query_params.append(('timeoutSeconds', params['timeout_seconds'])) if 'watch' in params: query_params.append(('watch', params['watch'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Status', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def delete_namespaced_build(self, name, namespace, body, **kwargs): """ delete a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_namespaced_build(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1DeleteOptions body: (required) :param str pretty: If 'true', then the output is pretty printed. :param int grace_period_seconds: The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately. :param bool orphan_dependents: Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the \"orphan\" finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both. :param str propagation_policy: Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground. :return: V1Status If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.delete_namespaced_build_with_http_info(name, namespace, body, **kwargs) else: (data) = self.delete_namespaced_build_with_http_info(name, namespace, body, **kwargs) return data def delete_namespaced_build_with_http_info(self, name, namespace, body, **kwargs): """ delete a Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_namespaced_build_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1DeleteOptions body: (required) :param str pretty: If 'true', then the output is pretty printed. :param int grace_period_seconds: The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately. :param bool orphan_dependents: Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the \"orphan\" finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both. :param str propagation_policy: Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground. :return: V1Status If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty', 'grace_period_seconds', 'orphan_dependents', 'propagation_policy'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method delete_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `delete_namespaced_build`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `delete_namespaced_build`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `delete_namespaced_build`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if 'grace_period_seconds' in params: query_params.append(('gracePeriodSeconds', params['grace_period_seconds'])) if 'orphan_dependents' in params: query_params.append(('orphanDependents', params['orphan_dependents'])) if 'propagation_policy' in params: query_params.append(('propagationPolicy', params['propagation_policy'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Status', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def delete_namespaced_build_config(self, name, namespace, body, **kwargs): """ delete a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_namespaced_build_config(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1DeleteOptions body: (required) :param str pretty: If 'true', then the output is pretty printed. :param int grace_period_seconds: The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately. :param bool orphan_dependents: Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the \"orphan\" finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both. :param str propagation_policy: Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground. :return: V1Status If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.delete_namespaced_build_config_with_http_info(name, namespace, body, **kwargs) else: (data) = self.delete_namespaced_build_config_with_http_info(name, namespace, body, **kwargs) return data def delete_namespaced_build_config_with_http_info(self, name, namespace, body, **kwargs): """ delete a BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_namespaced_build_config_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1DeleteOptions body: (required) :param str pretty: If 'true', then the output is pretty printed. :param int grace_period_seconds: The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately. :param bool orphan_dependents: Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the \"orphan\" finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both. :param str propagation_policy: Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground. :return: V1Status If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty', 'grace_period_seconds', 'orphan_dependents', 'propagation_policy'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method delete_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `delete_namespaced_build_config`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `delete_namespaced_build_config`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `delete_namespaced_build_config`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if 'grace_period_seconds' in params: query_params.append(('gracePeriodSeconds', params['grace_period_seconds'])) if 'orphan_dependents' in params: query_params.append(('orphanDependents', params['orphan_dependents'])) if 'propagation_policy' in params: query_params.append(('propagationPolicy', params['propagation_policy'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Status', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_api_resources(self, **kwargs): """ get available resources This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_api_resources(async_req=True) >>> result = thread.get() :param async_req bool :return: V1APIResourceList If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_api_resources_with_http_info(**kwargs) else: (data) = self.get_api_resources_with_http_info(**kwargs) return data def get_api_resources_with_http_info(self, **kwargs): """ get available resources This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_api_resources_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool :return: V1APIResourceList If the method is called asynchronously, returns the request thread. """ all_params = [] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_api_resources" % key ) params[key] = val del params['kwargs'] collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1APIResourceList', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def list_build_config_for_all_namespaces(self, **kwargs): """ list or watch objects of kind BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_build_config_for_all_namespaces(async_req=True) >>> result = thread.get() :param async_req bool :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildConfigList If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.list_build_config_for_all_namespaces_with_http_info(**kwargs) else: (data) = self.list_build_config_for_all_namespaces_with_http_info(**kwargs) return data def list_build_config_for_all_namespaces_with_http_info(self, **kwargs): """ list or watch objects of kind BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_build_config_for_all_namespaces_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildConfigList If the method is called asynchronously, returns the request thread. """ all_params = ['pretty', '_continue', 'field_selector', 'include_uninitialized', 'label_selector', 'limit', 'resource_version', 'timeout_seconds', 'watch'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method list_build_config_for_all_namespaces" % key ) params[key] = val del params['kwargs'] collection_formats = {} path_params = {} query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if '_continue' in params: query_params.append(('continue', params['_continue'])) if 'field_selector' in params: query_params.append(('fieldSelector', params['field_selector'])) if 'include_uninitialized' in params: query_params.append(('includeUninitialized', params['include_uninitialized'])) if 'label_selector' in params: query_params.append(('labelSelector', params['label_selector'])) if 'limit' in params: query_params.append(('limit', params['limit'])) if 'resource_version' in params: query_params.append(('resourceVersion', params['resource_version'])) if 'timeout_seconds' in params: query_params.append(('timeoutSeconds', params['timeout_seconds'])) if 'watch' in params: query_params.append(('watch', params['watch'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf', 'application/json;stream=watch', 'application/vnd.kubernetes.protobuf;stream=watch']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/buildconfigs', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfigList', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def list_build_for_all_namespaces(self, **kwargs): """ list or watch objects of kind Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_build_for_all_namespaces(async_req=True) >>> result = thread.get() :param async_req bool :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildList If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.list_build_for_all_namespaces_with_http_info(**kwargs) else: (data) = self.list_build_for_all_namespaces_with_http_info(**kwargs) return data def list_build_for_all_namespaces_with_http_info(self, **kwargs): """ list or watch objects of kind Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_build_for_all_namespaces_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildList If the method is called asynchronously, returns the request thread. """ all_params = ['pretty', '_continue', 'field_selector', 'include_uninitialized', 'label_selector', 'limit', 'resource_version', 'timeout_seconds', 'watch'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method list_build_for_all_namespaces" % key ) params[key] = val del params['kwargs'] collection_formats = {} path_params = {} query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if '_continue' in params: query_params.append(('continue', params['_continue'])) if 'field_selector' in params: query_params.append(('fieldSelector', params['field_selector'])) if 'include_uninitialized' in params: query_params.append(('includeUninitialized', params['include_uninitialized'])) if 'label_selector' in params: query_params.append(('labelSelector', params['label_selector'])) if 'limit' in params: query_params.append(('limit', params['limit'])) if 'resource_version' in params: query_params.append(('resourceVersion', params['resource_version'])) if 'timeout_seconds' in params: query_params.append(('timeoutSeconds', params['timeout_seconds'])) if 'watch' in params: query_params.append(('watch', params['watch'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf', 'application/json;stream=watch', 'application/vnd.kubernetes.protobuf;stream=watch']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/builds', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildList', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def list_namespaced_build(self, namespace, **kwargs): """ list or watch objects of kind Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_namespaced_build(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildList If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.list_namespaced_build_with_http_info(namespace, **kwargs) else: (data) = self.list_namespaced_build_with_http_info(namespace, **kwargs) return data def list_namespaced_build_with_http_info(self, namespace, **kwargs): """ list or watch objects of kind Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_namespaced_build_with_http_info(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildList If the method is called asynchronously, returns the request thread. """ all_params = ['namespace', 'pretty', '_continue', 'field_selector', 'include_uninitialized', 'label_selector', 'limit', 'resource_version', 'timeout_seconds', 'watch'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method list_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `list_namespaced_build`") collection_formats = {} path_params = {} if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if '_continue' in params: query_params.append(('continue', params['_continue'])) if 'field_selector' in params: query_params.append(('fieldSelector', params['field_selector'])) if 'include_uninitialized' in params: query_params.append(('includeUninitialized', params['include_uninitialized'])) if 'label_selector' in params: query_params.append(('labelSelector', params['label_selector'])) if 'limit' in params: query_params.append(('limit', params['limit'])) if 'resource_version' in params: query_params.append(('resourceVersion', params['resource_version'])) if 'timeout_seconds' in params: query_params.append(('timeoutSeconds', params['timeout_seconds'])) if 'watch' in params: query_params.append(('watch', params['watch'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf', 'application/json;stream=watch', 'application/vnd.kubernetes.protobuf;stream=watch']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildList', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def list_namespaced_build_config(self, namespace, **kwargs): """ list or watch objects of kind BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_namespaced_build_config(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildConfigList If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.list_namespaced_build_config_with_http_info(namespace, **kwargs) else: (data) = self.list_namespaced_build_config_with_http_info(namespace, **kwargs) return data def list_namespaced_build_config_with_http_info(self, namespace, **kwargs): """ list or watch objects of kind BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_namespaced_build_config_with_http_info(namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param str _continue: The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server the server will respond with a 410 ResourceExpired error indicating the client must restart their list without the continue field. This field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications. :param str field_selector: A selector to restrict the list of returned objects by their fields. Defaults to everything. :param bool include_uninitialized: If true, partially initialized resources are included in the response. :param str label_selector: A selector to restrict the list of returned objects by their labels. Defaults to everything. :param int limit: limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true. The server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned. :param str resource_version: When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv. :param int timeout_seconds: Timeout for the list/watch call. :param bool watch: Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion. :return: V1BuildConfigList If the method is called asynchronously, returns the request thread. """ all_params = ['namespace', 'pretty', '_continue', 'field_selector', 'include_uninitialized', 'label_selector', 'limit', 'resource_version', 'timeout_seconds', 'watch'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method list_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `list_namespaced_build_config`") collection_formats = {} path_params = {} if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if '_continue' in params: query_params.append(('continue', params['_continue'])) if 'field_selector' in params: query_params.append(('fieldSelector', params['field_selector'])) if 'include_uninitialized' in params: query_params.append(('includeUninitialized', params['include_uninitialized'])) if 'label_selector' in params: query_params.append(('labelSelector', params['label_selector'])) if 'limit' in params: query_params.append(('limit', params['limit'])) if 'resource_version' in params: query_params.append(('resourceVersion', params['resource_version'])) if 'timeout_seconds' in params: query_params.append(('timeoutSeconds', params['timeout_seconds'])) if 'watch' in params: query_params.append(('watch', params['watch'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf', 'application/json;stream=watch', 'application/vnd.kubernetes.protobuf;stream=watch']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfigList', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def patch_namespaced_build(self, name, namespace, body, **kwargs): """ partially update the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_namespaced_build(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param object body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.patch_namespaced_build_with_http_info(name, namespace, body, **kwargs) else: (data) = self.patch_namespaced_build_with_http_info(name, namespace, body, **kwargs) return data def patch_namespaced_build_with_http_info(self, name, namespace, body, **kwargs): """ partially update the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_namespaced_build_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param object body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method patch_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `patch_namespaced_build`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `patch_namespaced_build`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `patch_namespaced_build`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['application/json-patch+json', 'application/merge-patch+json', 'application/strategic-merge-patch+json']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}', 'PATCH', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def patch_namespaced_build_config(self, name, namespace, body, **kwargs): """ partially update the specified BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_namespaced_build_config(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param object body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.patch_namespaced_build_config_with_http_info(name, namespace, body, **kwargs) else: (data) = self.patch_namespaced_build_config_with_http_info(name, namespace, body, **kwargs) return data def patch_namespaced_build_config_with_http_info(self, name, namespace, body, **kwargs): """ partially update the specified BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_namespaced_build_config_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param object body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method patch_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `patch_namespaced_build_config`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `patch_namespaced_build_config`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `patch_namespaced_build_config`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['application/json-patch+json', 'application/merge-patch+json', 'application/strategic-merge-patch+json']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}', 'PATCH', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfig', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def read_namespaced_build(self, name, namespace, **kwargs): """ read the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.read_namespaced_build(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param bool exact: Should the export be exact. Exact export maintains cluster-specific fields like 'Namespace'. :param bool export: Should this value be exported. Export strips fields that a user can not specify. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.read_namespaced_build_with_http_info(name, namespace, **kwargs) else: (data) = self.read_namespaced_build_with_http_info(name, namespace, **kwargs) return data def read_namespaced_build_with_http_info(self, name, namespace, **kwargs): """ read the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.read_namespaced_build_with_http_info(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param bool exact: Should the export be exact. Exact export maintains cluster-specific fields like 'Namespace'. :param bool export: Should this value be exported. Export strips fields that a user can not specify. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'pretty', 'exact', 'export'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method read_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `read_namespaced_build`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `read_namespaced_build`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if 'exact' in params: query_params.append(('exact', params['exact'])) if 'export' in params: query_params.append(('export', params['export'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def read_namespaced_build_config(self, name, namespace, **kwargs): """ read the specified BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.read_namespaced_build_config(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param bool exact: Should the export be exact. Exact export maintains cluster-specific fields like 'Namespace'. :param bool export: Should this value be exported. Export strips fields that a user can not specify. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.read_namespaced_build_config_with_http_info(name, namespace, **kwargs) else: (data) = self.read_namespaced_build_config_with_http_info(name, namespace, **kwargs) return data def read_namespaced_build_config_with_http_info(self, name, namespace, **kwargs): """ read the specified BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.read_namespaced_build_config_with_http_info(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str pretty: If 'true', then the output is pretty printed. :param bool exact: Should the export be exact. Exact export maintains cluster-specific fields like 'Namespace'. :param bool export: Should this value be exported. Export strips fields that a user can not specify. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'pretty', 'exact', 'export'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method read_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `read_namespaced_build_config`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `read_namespaced_build_config`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if 'exact' in params: query_params.append(('exact', params['exact'])) if 'export' in params: query_params.append(('export', params['export'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfig', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def read_namespaced_build_log(self, name, namespace, **kwargs): """ read log of the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.read_namespaced_build_log(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildLog (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str container: cointainer for which to stream logs. Defaults to only container if there is one container in the pod. :param bool follow: follow if true indicates that the build log should be streamed until the build terminates. :param int limit_bytes: limitBytes, If set, is the number of bytes to read from the server before terminating the log output. This may not display a complete final line of logging, and may return slightly more or slightly less than the specified limit. :param bool nowait: noWait if true causes the call to return immediately even if the build is not available yet. Otherwise the server will wait until the build has started. :param str pretty: If 'true', then the output is pretty printed. :param bool previous: previous returns previous build logs. Defaults to false. :param int since_seconds: sinceSeconds is a relative time in seconds before the current time from which to show logs. If this value precedes the time a pod was started, only logs since the pod start will be returned. If this value is in the future, no logs will be returned. Only one of sinceSeconds or sinceTime may be specified. :param int tail_lines: tailLines, If set, is the number of lines from the end of the logs to show. If not specified, logs are shown from the creation of the container or sinceSeconds or sinceTime :param bool timestamps: timestamps, If true, add an RFC3339 or RFC3339Nano timestamp at the beginning of every line of log output. Defaults to false. :param int version: version of the build for which to view logs. :return: V1BuildLog If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.read_namespaced_build_log_with_http_info(name, namespace, **kwargs) else: (data) = self.read_namespaced_build_log_with_http_info(name, namespace, **kwargs) return data def read_namespaced_build_log_with_http_info(self, name, namespace, **kwargs): """ read log of the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.read_namespaced_build_log_with_http_info(name, namespace, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildLog (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param str container: cointainer for which to stream logs. Defaults to only container if there is one container in the pod. :param bool follow: follow if true indicates that the build log should be streamed until the build terminates. :param int limit_bytes: limitBytes, If set, is the number of bytes to read from the server before terminating the log output. This may not display a complete final line of logging, and may return slightly more or slightly less than the specified limit. :param bool nowait: noWait if true causes the call to return immediately even if the build is not available yet. Otherwise the server will wait until the build has started. :param str pretty: If 'true', then the output is pretty printed. :param bool previous: previous returns previous build logs. Defaults to false. :param int since_seconds: sinceSeconds is a relative time in seconds before the current time from which to show logs. If this value precedes the time a pod was started, only logs since the pod start will be returned. If this value is in the future, no logs will be returned. Only one of sinceSeconds or sinceTime may be specified. :param int tail_lines: tailLines, If set, is the number of lines from the end of the logs to show. If not specified, logs are shown from the creation of the container or sinceSeconds or sinceTime :param bool timestamps: timestamps, If true, add an RFC3339 or RFC3339Nano timestamp at the beginning of every line of log output. Defaults to false. :param int version: version of the build for which to view logs. :return: V1BuildLog If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'container', 'follow', 'limit_bytes', 'nowait', 'pretty', 'previous', 'since_seconds', 'tail_lines', 'timestamps', 'version'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method read_namespaced_build_log" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `read_namespaced_build_log`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `read_namespaced_build_log`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'container' in params: query_params.append(('container', params['container'])) if 'follow' in params: query_params.append(('follow', params['follow'])) if 'limit_bytes' in params: query_params.append(('limitBytes', params['limit_bytes'])) if 'nowait' in params: query_params.append(('nowait', params['nowait'])) if 'pretty' in params: query_params.append(('pretty', params['pretty'])) if 'previous' in params: query_params.append(('previous', params['previous'])) if 'since_seconds' in params: query_params.append(('sinceSeconds', params['since_seconds'])) if 'tail_lines' in params: query_params.append(('tailLines', params['tail_lines'])) if 'timestamps' in params: query_params.append(('timestamps', params['timestamps'])) if 'version' in params: query_params.append(('version', params['version'])) header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}/log', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildLog', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def replace_namespaced_build(self, name, namespace, body, **kwargs): """ replace the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.replace_namespaced_build(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.replace_namespaced_build_with_http_info(name, namespace, body, **kwargs) else: (data) = self.replace_namespaced_build_with_http_info(name, namespace, body, **kwargs) return data def replace_namespaced_build_with_http_info(self, name, namespace, body, **kwargs): """ replace the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.replace_namespaced_build_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method replace_namespaced_build" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `replace_namespaced_build`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `replace_namespaced_build`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `replace_namespaced_build`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}', 'PUT', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def replace_namespaced_build_config(self, name, namespace, body, **kwargs): """ replace the specified BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.replace_namespaced_build_config(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildConfig body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.replace_namespaced_build_config_with_http_info(name, namespace, body, **kwargs) else: (data) = self.replace_namespaced_build_config_with_http_info(name, namespace, body, **kwargs) return data def replace_namespaced_build_config_with_http_info(self, name, namespace, body, **kwargs): """ replace the specified BuildConfig This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.replace_namespaced_build_config_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the BuildConfig (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1BuildConfig body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1BuildConfig If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method replace_namespaced_build_config" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `replace_namespaced_build_config`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `replace_namespaced_build_config`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `replace_namespaced_build_config`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/buildconfigs/{name}', 'PUT', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1BuildConfig', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def replace_namespaced_build_details(self, name, namespace, body, **kwargs): """ replace details of the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.replace_namespaced_build_details(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.replace_namespaced_build_details_with_http_info(name, namespace, body, **kwargs) else: (data) = self.replace_namespaced_build_details_with_http_info(name, namespace, body, **kwargs) return data def replace_namespaced_build_details_with_http_info(self, name, namespace, body, **kwargs): """ replace details of the specified Build This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.replace_namespaced_build_details_with_http_info(name, namespace, body, async_req=True) >>> result = thread.get() :param async_req bool :param str name: name of the Build (required) :param str namespace: object name and auth scope, such as for teams and projects (required) :param V1Build body: (required) :param str pretty: If 'true', then the output is pretty printed. :return: V1Build If the method is called asynchronously, returns the request thread. """ all_params = ['name', 'namespace', 'body', 'pretty'] all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method replace_namespaced_build_details" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'name' is set if ('name' not in params) or (params['name'] is None): raise ValueError("Missing the required parameter `name` when calling `replace_namespaced_build_details`") # verify the required parameter 'namespace' is set if ('namespace' not in params) or (params['namespace'] is None): raise ValueError("Missing the required parameter `namespace` when calling `replace_namespaced_build_details`") # verify the required parameter 'body' is set if ('body' not in params) or (params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `replace_namespaced_build_details`") collection_formats = {} path_params = {} if 'name' in params: path_params['name'] = params['name'] if 'namespace' in params: path_params['namespace'] = params['namespace'] query_params = [] if 'pretty' in params: query_params.append(('pretty', params['pretty'])) header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.\ select_header_accept(['application/json', 'application/yaml', 'application/vnd.kubernetes.protobuf']) # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.\ select_header_content_type(['*/*']) # Authentication setting auth_settings = ['BearerToken', 'Oauth2AccessToken', 'Oauth2Implicit'] return self.api_client.call_api('/apis/build.openshift.io/v1/namespaces/{namespace}/builds/{name}/details', 'PUT', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='V1Build', auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats)
190,066
45,999
import socket import sys IP_ADDR = "192.168.1.19" TCP_PORT = 10000 if __name__ == "__main__": # Create TCP socket sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) # Associate the socket with the server address server_address = (IP_ADDR, TCP_PORT) print("Start TCP server at address {} on port {} ".format(server_address[0], server_address[1])) sock.bind(server_address) # Mode TCP server sock.listen(1) while True: connection, client_address = sock.accept() while True : print("Connection from {} ".format(client_address)) data = connection.recv(16) print("Data : %s" % data) else: connection.close()
722
238
import pandas as pd countryInformation = pd.read_csv('resource/countryInformation.csv') #looping row #for index,row in countryInformation.iterrows(): #print(index, row['country_name']) print(countryInformation.loc[countryInformation['country_name'] == 'india'])
269
77
""" List Comprehensions Examples """ my_list = [] # my_list.append() # my_list.extend() """ When to use ListComps """ phones = [ { 'number': '111-111-1111', 'label': 'phone', 'extension': '1234', }, { 'number': '222-222-2222', 'label': 'mobile', 'extension': None, } ] my_phone_list = [] for phone in phones: my_phone_list.append(phone['number']) # List Comprehension [phone['number'] for phone in phones] """ Advanced Usage """ # Buld an explicit nested list table = [ [1, 2, 3], [1, 2, 3], [1, 2, 3], ] fields = ['x', 'y', 'z'] rows = [1, 2, 3] table = [] for r in rows: row = [] for field in fields: row.append(field) table.append(row) [field for field in fields] [row for row in rows] table = [[field for field in fields] for row in rows] """ Dictionary Comprehensions """ [{str(item): item} for item in [1, 2, 3, ]] dict1 = {'a': 1, 'b': 2, 'c': 3, 'd': 4, 'e': 5} double_dict1 = {k: v * 2 for (k, v) in dict1.items()} dict_map = { 'apple' : 1, 'cherry': 2, 'earwax': 3, } {v:k for (k, v) in dict_map.items()} items = dict_map.items() """ Logical Comparisons """ values = [1,2,3] [i for i in values if i < 3] [k for k, v in dict_map.items() if v < 3] """ Performance, Spongecase Example """ original_string = 'hello world' spongecase_letters = [] for index, letter in enumerate(original_string): if index % 2 == 1: spongecase_letters.append(letter.upper()) else: spongecase_letters.append(letter) spongecase_string = ''.join(spongecase_letters) # hElLo wOrLd def spongecase(index, letter): if index % 2 == 1: return letter.upper() else: return letter original_string = 'hello world' spongecase_letters = [] for index, letter in enumerate(original_string): transformed_letter = spongecase(index, letter) spongecase_letters.append(transformed_letter) spongecase_string = ''.join(spongecase_letters) # hElLo wOrLd
2,011
833
import tensorflow as tf from tensorflow.keras import backend as K from tensorflow.keras import models as KM from tensorflow.keras import layers as KL class DenoisingNetwork(object): def __new__(cls, mode: str) \ -> KM.Model: assert mode in ['base', 'skip', 'bn'] inputs = KL.Input(shape=[None, None, 3], name="input_image") x = inputs x = KL.Conv2D(64, (3, 3), padding="SAME", kernel_initializer='random_uniform', bias_initializer='zeros', name="layer1")(x) if mode == 'bn': x = KL.BatchNormalization()(x) x = KL.ReLU()(x) x = KL.Conv2D(64, (3, 3), padding="SAME", kernel_initializer='random_uniform', bias_initializer='zeros', name="layer2")(x) if mode == 'bn': x = KL.BatchNormalization()(x) x = KL.ReLU()(x) x = KL.Conv2D(64, (3, 3), padding="SAME", kernel_initializer='random_uniform', bias_initializer='zeros', name="layer3")(x) if mode == 'bn': x = KL.BatchNormalization()(x) x = KL.ReLU()(x) x = KL.Conv2D(64, (3, 3), padding="SAME", kernel_initializer='random_uniform', bias_initializer='zeros', name="layer4")(x) if mode == 'bn': x = KL.BatchNormalization()(x) x = KL.ReLU()(x) x = KL.Conv2D(3, (3, 3), padding="SAME", kernel_initializer='random_uniform', bias_initializer='zeros', name="layer5")(x) if mode == 'bn': x = KL.BatchNormalization()(x) x = KL.ReLU()(x) if mode == 'skip' or mode == 'bn': x = KL.average([x, inputs]) return KM.Model(inputs=inputs, outputs=x, name='denoising') @staticmethod def loss(y_true: tf.Tensor, y_pred: tf.Tensor) \ -> tf.Tensor: return K.mean(K.square(y_pred - y_true)) @classmethod def metric(cls, y_true: tf.Tensor, y_pred: tf.Tensor) \ -> tf.Tensor: return tf.image.psnr(y_true, y_pred, max_val=1.) @classmethod def compile(cls, model, optimizer, loss, metric)\ -> None: model.compile(optimizer=optimizer, loss=loss, metrics=[metric])
2,549
829
import time class Timer(object): def __init__(self, Log, *args, **kwargs): self.Log = Log return super().__init__(*args, **kwargs) @property def time(self): return time.strftime('%Y-%m-%d-%H-%M-%S', time.localtime()) def mktime(self, timex): return time.mktime(time.strptime(timex, '%Y-%m-%d-%H-%M-%S')) def timer(self, text, func, *args, **kwargs): start_time = self.time self.Log(start_time, _T=f'{text} Start:') result = func(*args, **kwargs) stop_time = self.time self.Log(stop_time, _T=f'{text} Stop:') cost_time = self.mktime(stop_time) - self.mktime(start_time) self.Log(cost_time, _T=f'{text} cost time (second):') time_dict = {f'{text}_start_time'.upper(): start_time, f'{text}_stop_time'.upper(): stop_time, f'{text}_cost_time'.upper(): cost_time} return time_dict, result
884
336
""" Select only a part of the instances .. todo: group instance selectors """ import random import logging from collections import defaultdict from pySPACE.missions.nodes.base_node import BaseNode from pySPACE.tools.memoize_generator import MemoizeGenerator class InstanceSelectionNode(BaseNode): """Retain only a certain percentage of the instances The node InstanceSelectionNode forwards only *train_percentage_selected* percent of the training instances passed to him to the successor node and only *test_percentage_selected* percent of the test instances. The forwarded instances are selected randomly but so that the class ratio is kept. If *reduce_class* is used, only the chosen class is reduced, without keeping the class ratio. So the total mount of reduced data does not match the percentage values. **Parameters** :train_percentage_selected: The percentage of training instances which is forwarded to successor node. (*optional, default: 100*) :test_percentage_selected: The percentage of test instances which is forwarded to successor node. (*optional, default: 100*) :reduce_class: If you want only to reduce one class, choose this parameter otherwise, both classes are reduced in a balanced fashion. (*optional, default: False*) :num_train_instances: Instead of specifying *train_percentage_selected*, this option allows to specify the absolute number of training instances of class *class_label* that should be in the training set. All instances that occur until *num_train_instances* are found are used for training. (*optional, default: None*) :class_label: If *num_train_instances*-option is used, this string determines the class of which training examples are count. (*optional, default: 'Target'*) :random: If *False*, the order of the data is retained. I.e. the first X percent or number of train instances are used for training. If *True*, the training data is sampled randomly without taking into consideration the data's order. (*optional, default: True*) **Exemplary call** .. code-block:: yaml - node : InstanceSelection parameters : train_percentage_selected : 80 test_percentage_selected : 100 reduce_class : Standard :Author: Jan Hendrik Metzen (jhm@informatik.uni-bremen.de) :Created: 2010/03/31 """ def __init__(self, train_percentage_selected=100, test_percentage_selected=100, reduce_class=False, num_train_instances=None, class_label='Target', random=True, **kwargs): super(InstanceSelectionNode, self).__init__(**kwargs) self.set_permanent_attributes( train_percentage_selected=train_percentage_selected, test_percentage_selected=test_percentage_selected, reduce_class=reduce_class, num_train_instances=num_train_instances, class_label=class_label, random=random) def get_num_data(self, iterator): """ Return a list of instances that contain *num_train_instances* many instances of class *class_label* and all other instances that occur up to this point """ counter = 0 retained_instances = [] while counter < self.num_train_instances: try: instance, label = iterator.next() except: #TODO: give some warning to user break else: if label == self.class_label: counter += 1 retained_instances.append((instance,label)) return retained_instances def request_data_for_training(self, use_test_data): """ Returns data for training of subsequent nodes .. todo:: to document .. note:: This method works differently in InstanceSelectionNode than in other nodes: Only *percentage_selected* of the available data are returned. """ assert(self.input_node is not None) if self.train_percentage_selected > 100: self._log("Train percentage of %f reduced to 100." % self.train_percentage_selected, level=logging.ERROR) self.train_percentage_selected = 100 self._log("Data for training is requested.", level=logging.DEBUG) if self.train_percentage_selected == 100 and \ self.num_train_instances is None: return super(InstanceSelectionNode, self).request_data_for_training( use_test_data) # If we haven't computed the data for training yet if self.data_for_training is None: self._log("Producing data for training.", level=logging.DEBUG) # Train this node self.train_sweep(use_test_data) if not self.num_train_instances is None and self.random == False: retained_instances = self.get_num_data( self.input_node.request_data_for_training(use_test_data)) else: # Store all data if self.num_train_instances is None: all_instances = defaultdict(list) for instance, label in self.input_node.request_data_for_training( use_test_data): all_instances[label].append(instance) else: all_instances = list( self.input_node.request_data_for_traning(use_test_data)) if self.random: r = random.Random(self.run_number) if not self.num_train_instances is None and self.random: r.shuffle(all_instances) retained_instances = self.get_num_data( all_instances.__iter__()) else: retained_instances = [] self._log("Keeping only %s percent of training data" % self.train_percentage_selected, level=logging.DEBUG) # Retain only *percentage_selected* percent of the data for label, instances in all_instances.iteritems(): # enable random choice of samples r.shuffle(instances) if not self.reduce_class or \ self.train_percentage_selected == 100: end_index = int(round(len(instances) * self.train_percentage_selected / 100)) elif not (self.reduce_class == label): end_index = len(instances) else: # self.reduce_class==label--> reduction needed end_index = int(round(len(instances) * self.train_percentage_selected / 100)) retained_instances.extend(zip(instances[0:end_index], [label]*end_index)) if self.random: # mix up samples between the different labels r.shuffle(retained_instances) # Compute a generator the yields the train data and # encapsulate it in an object that memoizes its outputs and # provides a "fresh" method that returns a new generator that will # yield the same sequence train_data_generator = ((self.execute(data), label) for (data, label) in retained_instances) self.data_for_training = MemoizeGenerator(train_data_generator, caching=self.caching) self._log("Data for training finished", level=logging.DEBUG) # Return a fresh copy of the generator return self.data_for_training.fresh() def request_data_for_testing(self): """ Returns data for testing of subsequent nodes .. todo:: to document """ assert(self.input_node is not None) if self.test_percentage_selected > 100: self._log("Test percentage of %f reduced to 100." % self.test_percentage_selected, level=logging.ERROR) self.test_percentage_selected = 100 self._log("Data for testing is requested.", level=logging.DEBUG) if self.test_percentage_selected == 100: return super(InstanceSelectionNode, self).request_data_for_testing() # If we haven't computed the data for testing yet if self.data_for_testing is None: # Assert that this node has already been trained assert(not self.is_trainable() or self.get_remaining_train_phase() == 0) # Divide available instances according to label all_instances = defaultdict(list) for instance, label in self.input_node.request_data_for_testing(): all_instances[label].append(instance) self._log("Keeping only %s percent of test data" % self.test_percentage_selected, level=logging.DEBUG) r = random.Random(self.run_number) # Retain only *percentage_selected* percent of the data retained_instances = [] for label, instances in all_instances.iteritems(): # enable random choice of samples r.shuffle(instances) if not self.reduce_class or \ self.test_percentage_selected == 100: end_index = int(round(len(instances) * self.test_percentage_selected / 100)) elif not (self.reduce_class == label): end_index = len(instances) else: # self.reduce_class==label--> reduction needed end_index = int(round(len(instances) * self.test_percentage_selected / 100)) retained_instances.extend(zip(instances[0:end_index], [label]*end_index)) # mix up samples between the different labels r.shuffle(retained_instances) # Compute a generator the yields the test data and # encapsulate it in an object that memoizes its outputs and # provides a "fresh" method that returns a new generator that'll # yield the same sequence self._log("Producing data for testing.", level=logging.DEBUG) test_data_generator = ((self.execute(data), label) for (data, label) in retained_instances) self.data_for_testing = MemoizeGenerator(test_data_generator, caching=self.caching) self._log("Data for testing finished", level=logging.DEBUG) # Return a fresh copy of the generator return self.data_for_testing.fresh() def _execute(self, time_series): return time_series # We don't do anything with the kept instances class ReduceOverrepresentedClassNode(BaseNode): """ Reject instances to balance categories for classification The node forwards only a reduced number of the training and test instances of the bigger class to get a balanced ratio of the classes. The forwarded instances are selected randomly. All data of the underrepresented class is forwarded. **Parameters** **Exemplary call** .. code-block:: yaml - node : Reduce_Overrepresented_Class :Author: Hendrik Woehrle (hendrik.woehrle@dfki.de) :Created: 2010/09/22 """ def __init__(self, **kwargs): super(ReduceOverrepresentedClassNode, self).__init__(**kwargs) def request_data_for_training(self, use_test_data): """ Returns data for training of subsequent nodes .. todo:: to document """ assert(self.input_node is not None) self._log("Data for testing is requested.", level=logging.DEBUG) if self.data_for_training is None: self._log("Producing data for training.", level=logging.DEBUG) # Train this node self.train_sweep(use_test_data) # Divide available instances according to label all_instances = defaultdict(list) for instance, label in self.input_node.request_data_for_training( use_test_data): all_instances[label].append(instance) retained_instances = self.balance_instances(all_instances) # Compute a generator the yields the test data and # encapsulate it in an object that memoizes its outputs and # provides a "fresh" method that returns a new generator that will # yield the same sequence self._log("Producing data for testing.", level=logging.DEBUG) train_data_generator = ((self.execute(data), label) for (data, label) in retained_instances) self.data_for_training = MemoizeGenerator(train_data_generator, caching=self.caching) self._log("Data for training finished", level=logging.DEBUG) # Return a fresh copy of the generator return self.data_for_training.fresh() def request_data_for_testing(self): """ Returns data for testing of subsequent nodes .. todo:: to document """ assert(self.input_node is not None) self._log("Data for testing is requested.", level=logging.DEBUG) # If we haven't computed the data for testing yet if self.data_for_testing is None: # Assert that this node has already been trained assert(not self.is_trainable() or self.get_remaining_train_phase() == 0) # Divide available instances according to label all_instances = defaultdict(list) for instance, label in self.input_node.request_data_for_testing(): all_instances[label].append(instance) retained_instances = self.balance_instances(all_instances) # Compute a generator the yields the test data and # encapsulate it in an object that memoizes its outputs and # provides a "fresh" method that returns a new generator that will # yield the same sequence self._log("Producing data for testing.", level=logging.DEBUG) test_data_generator = ((self.execute(data), label) for (data, label) in retained_instances) self.data_for_testing = MemoizeGenerator(test_data_generator, caching=self.caching) self._log("Data for testing finished", level=logging.DEBUG) # Return a fresh copy of the generator return self.data_for_testing.fresh() def _execute(self, time_series): return time_series # We don't do anything with the kept instances def balance_instances(self, all_instances): """Method that performs the rejections of the data in the oversized class""" retained_instances = [] # it is supposed to have a binary classifier, e.g. to have exactly 2 classes #if not len(all_instances.keys())==2: # raise ValueError("Too many classes: only binary classification supported") # count the number of instances per class min_num_instances_per_class = float("+inf") for label, instances in all_instances.iteritems(): min_num_instances_per_class = min(min_num_instances_per_class, len(instances)) r = random.Random(self.run_number) # retain only the number of instances that corresponds # to the size of smaller class for label, instances in all_instances.iteritems(): r.shuffle(instances) retained_instances.extend( zip(instances[0:min_num_instances_per_class], [label]*min_num_instances_per_class)) r.shuffle(retained_instances) return retained_instances _NODE_MAPPING = {"RandomInstanceSelection": InstanceSelectionNode, "Reduce_Overrepresented_Class": ReduceOverrepresentedClassNode}
17,464
4,321
# -*- coding: utf-8 -*- """ """ from __future__ import absolute_import from contextlib import contextmanager import imp import posixpath from zipfile import ZipFile from click.testing import CliRunner import pkginfo import pytest from six import PY3 def test_pyfile_compiled(packages, tmpdir): packages.require_eggs('dist1') unzip = False if PY3: # Python >= 3.2 doesn't seem to run .pyc files from PEP 3147 # (__pycache__) repository directories. unzip = True venv = packages.get_venv('dist1', unzip=unzip) assert venv.run("__import__('dist1').test_is_compiled()") == 0 @pytest.fixture def dist1_metadata(packages): egg = packages.get_egg('dist1') return pkginfo.BDist(str(egg)) def test_summary(dist1_metadata): assert dist1_metadata.summary == "A dummy distribution" def test_description(dist1_metadata): assert dist1_metadata.description.rstrip() \ == u"Long description.\n\nGruß." def test_script_wrapper(packages): packages.require_eggs('dist1') venv = packages.get_venv('dist1') assert venv.call(['dist1_wrapper']) == 42 def test_old_style_script(packages): packages.require_eggs('dist1') venv = packages.get_venv('dist1') assert venv.call(['dist1_script']) == 42 def test_namespace_package(packages): packages.require_eggs('dist1', 'dist2') venv = packages.get_venv('dist2') prog = ( 'import sys\n' 'from dist2.plugins.builtin import the_answer\n' 'sys.exit(the_answer)\n' ) assert venv.run(prog) == 42 def test_namespace_stubs_in_egg(packages): dist2_egg = packages.get_egg('dist2') dist2_stubs = with_byte_compiled(['dist2/__init__.py', 'dist2/plugins/__init__.py']) with fileobj(ZipFile(str(dist2_egg))) as zf: files_in_egg = dist2_stubs.intersection(zf.namelist()) # Make sure we generated the stubs (or not, depending on python # version) stubs_in_egg = files_in_egg.intersection(dist2_stubs) assert stubs_in_egg == dist2_stubs # Make sure we didn't copy the .pth file that the wheel installer # creates for the namespaces assert not any(fn.lower().endswith('.pth') for fn in files_in_egg) def test_extension(packages): packages.require_eggs('extension_dist') venv = packages.get_venv('extension_dist') assert venv.run("__import__('extension_dist').test_extension()") == 0 def test_eager_resources(packages): packages.require_eggs('extension_dist') venv = packages.get_venv('extension_dist') assert venv.run("__import__('extension_dist').test_eager_resources()") == 0 def test_extras(packages): packages.require_eggs('dist1', 'extension_dist') venv = packages.get_venv('dist1[extras]') assert venv.run("__import__('dist1').test_extras()") == 0 def test_no_extras(packages): packages.require_eggs('dist1', 'extension_dist') venv = packages.get_venv('dist1') assert venv.run("__import__('dist1').test_no_extras()") == 0 def test_main(packages, tmpdir): from humpty import main wheel = packages.get_wheel('dist1') runner = CliRunner() result = runner.invoke(main, ['-d', str(tmpdir), str(wheel)]) assert result.exit_code == 0 eggs = list(tmpdir.listdir(fil="*.egg")) assert len(eggs) == 1 egg = eggs[0] assert egg.isfile() assert egg.fnmatch("dist1-*") @contextmanager def fileobj(fp): try: yield fp finally: fp.close() def with_byte_compiled(paths): """ Augment PATHS with paths of byte-compiled files. """ get_tag = getattr(imp, 'get_tag', None) compiled = set() for path in paths: head, tail = posixpath.split(path) root, ext = posixpath.splitext(tail) if ext == '.py': if get_tag: root = '%s.%s' % (root, get_tag()) head = posixpath.join(head, '__pycache__') compiled.add(posixpath.join(head, root + '.pyc')) return compiled.union(paths)
4,050
1,411
import logging from abc import ABC, abstractmethod logger = logging.getLogger(__name__) class ActionsMasterListBase(ABC): """ Base class meant to hold the entire Set of IAM resource actions. It is up to a concrete class to implement a source document parser (parse_actions_source) """ def __init__(self, source_master): """ :param source_master: :type source_master: str """ self._actions_set = self.parse_actions_source(source_master) self._actions_set_case_insensitive_lookup = {resource_action.lower(): resource_action for resource_action in self._actions_set} super().__init__() @abstractmethod def parse_actions_source(self, source_master): """ :param source_master: :type source_master: str :return: :rtype: set """ pass @abstractmethod def all_actions_for_resource(self, resource_name): """ This must return a sorted list of all actions for the given resource :param resource_name: :type resource_name: str :return: :rtype: list """ def all_actions_set(self, lower=False): return set(item.lower() for item in self._actions_set) if lower else self._actions_set def lookup_action(self, action): """ Case insensitive lookup for all known actions. Returned in PascalCase :param action: :type action: str :return: :rtype: str """ return self._actions_set_case_insensitive_lookup.get(action.lower())
1,649
446
import pybullet as p from gym import spaces import pybullet_planning as pbp import numpy as np from diy_gym.addons.addon import Addon class JointTrace(Addon): """ JointTrace Trace the follows a joints movements """ def __init__(self, parent, config): super().__init__(parent, config) self.uid = parent.uid joint_info = [p.getJointInfo(self.uid, i) for i in range(p.getNumJoints(self.uid))] if 'joint' in config: joints = [config.get('joint')] elif 'joints' in config: joints = config.get('joints') else: joints = [info[1].decode('UTF-8') for info in joint_info] self.joint_ids = [info[0] for info in joint_info if info[1].decode('UTF-8') in joints and info[3] > -1] self.last = None def reset(self): p.removeAllUserDebugItems() self.last = None def update(self, action): # A colored trace for each joint joint_pos = np.array([pbp.get_link_pose(self.uid, i)[0] for i in self.joint_ids]) if self.last is not None: m = len(joint_pos) for i in range(n): p.addUserDebugLine( lineFromXYZ=joint_pos[i], lineToXYZ=self.last[i], lineColorRGB=[(n-i)/(n+1), 0.9, i/(n+1)], lineWidth=1, lifeTime=360) self.last = joint_pos
1,401
464
def solution(S): rs = "" for i in S: if i != " ": rs += i else: rs += "%20" return rs S = "Mr John Smith" print(solution(S))
181
67
from __future__ import print_function import ngraph.transformers as ngt from ngraph.flex.names import flex_gpu_transformer_name import argparse class FlexNgraphArgparser(): """ Flex specific command line args """ @staticmethod def setup_flex_args(argParser): """ Add flex specific arguments to other default args used by ngraph """ # use fixed point for flex backend argParser.add_argument('--fixed_point', action="store_true", help=argparse.SUPPRESS) # turn on flex verbosity for debug argParser.add_argument('--flex_verbose', action="store_true", help=argparse.SUPPRESS) # collect flex data and save it to h5py File argParser.add_argument('--collect_flex_data', action="store_true", default=argparse.SUPPRESS) @staticmethod def make_and_set_transformer_factory(args): flex_args = ('fixed_point', 'flex_verbose', 'collect_flex_data') # default value for all flex args if not given, confusing with store_true in add_argument default = False if args.backend == flex_gpu_transformer_name: flex_args_dict = dict((a, getattr(args, a, default)) for a in flex_args) factory = ngt.make_transformer_factory(args.backend, **flex_args_dict) else: factory = ngt.make_transformer_factory(args.backend) ngt.set_transformer_factory(factory)
1,612
433
""" Simple reduced order solver. More of a no-op, in that it doesn't actually perform a flux solution """ import numpy from hydep.internal.features import FeatureCollection from hydep.internal import TransportResult from .lib import ReducedOrderSolver class SimpleROSolver(ReducedOrderSolver): """The simplest reduced order flux solution where nothing happens""" needs = FeatureCollection() def __init__(self): self._flux = None def processBOS(self, txResult, _timestep, _power): """Store flux from a high fidelity transport solution""" self._flux = txResult.flux def substepSolve(self, *args, **kwargs): """Return the beginning-of-step flux with no modifications Returns ------- hydep.internal.TransportResult Transport result with the flux provided in :meth:`processBOS` """ return TransportResult(self._flux, [numpy.nan, numpy.nan], runTime=numpy.nan)
969
270
import FWCore.ParameterSet.Config as cms # # produce ttGenEvent with all necessary ingredients # from TopQuarkAnalysis.TopEventProducers.producers.TopInitSubset_cfi import * from TopQuarkAnalysis.TopEventProducers.producers.TopDecaySubset_cfi import * from TopQuarkAnalysis.TopEventProducers.producers.TtGenEvtProducer_cfi import * makeGenEvtTask = cms.Task( initSubset, decaySubset, genEvt ) makeGenEvt = cms.Sequence(makeGenEvtTask)
449
160
import os from os.path import join as pjoin import time import numpy as np import torch from torch import nn import torch.nn.functional as F from torch.optim.lr_scheduler import CosineAnnealingLR try: from .radam import RAdam except (ImportError, ModuleNotFoundError) as err: from radam import RAdam try: from torch.nn import Flatten except ImportError: class Flatten(nn.Module): __constants__ = ['start_dim', 'end_dim'] def __init__(self, start_dim=1, end_dim=-1): super(Flatten, self).__init__() self.start_dim = start_dim self.end_dim = end_dim def forward(self, input): return input.flatten(self.start_dim, self.end_dim) class HPF(nn.Conv1d): def __init__(self, in_channels, out_channels, kernel_size, *args, **kwargs): super(HPF, self).__init__(in_channels, out_channels, kernel_size, *args, bias=False, **kwargs) self.hpf_kernel = np.array([[[ 1, -1, 0, 0, 0]], [[ 1, -2, 1, 0, 0]], [[ 1, -3, 3, -1, 0]], [[ 1, -4, 6, -4, 1]]]) self.weight.data = torch.tensor(self.hpf_kernel, dtype=self.weight.dtype) def initialize_parameters(self): device = next(iter(self.parameters())).device self.weight.data = torch.tensor(self.hpf_kernel, dtype=self.weight.dtype, device=device) # The following settings does not allow training HPF. #self.bias.data.fill_(0) #self.hpf.bias.requires_grad = False class TLU(nn.Module): def __init__(self, thr=3.0): """truncated linear unit (TLU) """ super(TLU, self).__init__() self.thr = thr def forward(self, x): return x.clamp(-self.thr, self.thr) #torch.min(torch.max(x, -self.thr), self.thr) class Group1(nn.Module): def __init__(self): super(Group1, self).__init__() self.module = nn.Sequential(nn.Conv1d(4, 8, 1), TLU(3.0), nn.Conv1d(8, 8, 5, padding=2), nn.Conv1d(8, 16, 1)) def forward(self, x): return self.module(x) class Group2(nn.Module): def __init__(self): super(Group2, self).__init__() self.module = nn.Sequential(nn.Conv1d(16, 16, 5, padding=2), nn.ReLU(), nn.Conv1d(16, 32, 1), nn.ReLU(), nn.AvgPool1d(3, stride=2, padding=1)) def forward(self, x): return self.module(x) class Group3(nn.Module): def __init__(self): super(Group3, self).__init__() self.module = nn.Sequential(nn.Conv1d(32, 32, 5, padding=2), nn.ReLU(), nn.Conv1d(32, 64, 1), nn.ReLU(), nn.AvgPool1d(3, stride=2, padding=1)) def forward(self, x): return self.module(x) class Group4(nn.Module): def __init__(self): super(Group4, self).__init__() self.module = nn.Sequential(nn.Conv1d(64, 64, 5, padding=2), nn.ReLU(), nn.Conv1d(64, 128, 1), nn.ReLU(), nn.AvgPool1d(3, stride=2, padding=1)) def forward(self, x): return self.module(x) class Group5(nn.Module): def __init__(self): super(Group5, self).__init__() self.module = nn.Sequential(nn.Conv1d(128, 128, 5, padding=2), nn.ReLU(), nn.Conv1d(128, 256, 1), nn.ReLU(), nn.AvgPool1d(3, stride=2, padding=1)) def forward(self, x): return self.module(x) class Group6(nn.Module): def __init__(self): super(Group6, self).__init__() self.module = nn.Sequential(nn.Conv1d(256, 256, 5, padding=2), nn.ReLU(), nn.Conv1d(256, 512, 1), nn.ReLU(), nn.AdaptiveAvgPool1d(1)) def forward(self, x): return self.module(x) class Classifier(nn.Module): def __init__(self, n_classes=2): super(Classifier, self).__init__() self.module = nn.Sequential(Flatten(1), nn.Linear(512, n_classes)) def forward(self, x): return self.module(x) class LinNet(nn.Module): @staticmethod def get_optimizer(model, lr): return RAdam(model.parameters(), lr=lr, weight_decay=1e-5) @staticmethod def get_lr_scheduler(optimizer): return CosineAnnealingLR(optimizer, T_max=20, eta_min=1e-7) def __str__(self): return self._name def __init__(self, n_classes=2): super(LinNet, self).__init__() self._name = "linnet" # HPF self.hpf = HPF(1, 4, 5, padding=2) self.group1 = Group1() self.group2 = Group2() self.group3 = Group3() self.group4 = Group4() self.group5 = Group5() self.group6 = Group6() self.classifier = Classifier(n_classes) self.initialize_parameters() def forward(self, x): y = self.hpf(x) g1 = self.group1(y) g2 = self.group2(g1) g3 = self.group3(g2) g4 = self.group4(g3) g5 = self.group5(g4) g6 = self.group6(g5) logits = self.classifier(g6) return logits def initialize_parameters(self): """ In the original paper, Lin et al. Conv1d: Xavier uniform initializer with zero biases """ """ [Original] for m in self.modules(): if isinstance(m, HPF): self.hpf.initialize_parameters() elif isinstance(m, nn.Conv1d): nn.init.kaiming_normal_(m.weight, mode='fan_in', nonlinearity='relu') nn.init.constant_(m.bias.data, val=1e-3) elif isinstance(m, nn.Linear): # Zero mean Gaussian with std 0.01 nn.init.normal_(m.weight, 0.0, 0.01) if m.bias is not None: nn.init.constant_(m.bias, 1e-3) """ # Following settings is the same with that of BSN. for m in self.modules(): if isinstance(m, HPF): self.hpf.initialize_parameters() if isinstance(m, nn.Conv1d): nn.init.xavier_uniform_(m.weight) if m.bias is not None: nn.init.zeros_(m.bias.data) elif isinstance(m, nn.Linear): nn.init.xavier_uniform_(m.weight) if m.bias is not None: nn.init.zeros_(m.bias.data) def initialize_curriculum_learning(self): for m in self.modules(): if isinstance(m, HPF): self.hpf.initialize_parameters() elif isinstance(m, nn.Linear): # Zero mean Gaussian with std 0.01 nn.init.normal_(m.weight, 0.0, 0.01) if m.bias is not None: nn.init.constant_(m.bias, 1e-3) if __name__ == "__main__": model = LinNet() n_ch = 1 for i in range(1, 2): x = torch.randn(1, n_ch, i*16000) t_beg = time.time() out = model(x) t_end = time.time() print("LinNet model output:", out) print("Execution time:", t_end - t_beg) # end of for
8,442
2,773
import os import sys import unittest from nose.importer import Importer class TestImporter(unittest.TestCase): def setUp(self): self.dir = os.path.normpath(os.path.join(os.path.dirname(__file__), 'support')) self.imp = Importer() self._mods = sys.modules.copy() self._path = sys.path[:] sys.modules.pop('mod', None) sys.modules.pop('pak', None) sys.modules.pop('pak.mod', None) sys.modules.pop('pak.sub', None) def tearDown(self): to_del = [ m for m in sys.modules.keys() if m not in self._mods ] if to_del: for mod in to_del: del sys.modules[mod] sys.modules.update(self._mods) sys.path = self._path[:] def test_import_from_dir(self): imp = self.imp d1 = os.path.join(self.dir, 'dir1') d2 = os.path.join(self.dir, 'dir2') # simple name m1 = imp.importFromDir(d1, 'mod') m2 = imp.importFromDir(d2, 'mod') self.assertNotEqual(m1, m2) self.assertNotEqual(m1.__file__, m2.__file__) # dotted name p1 = imp.importFromDir(d1, 'pak.mod') p2 = imp.importFromDir(d2, 'pak.mod') self.assertNotEqual(p1, p2) self.assertNotEqual(p1.__file__, p2.__file__) def test_import_from_path(self): imp = self.imp jn = os.path.join d1 = jn(self.dir, 'dir1') d2 = jn(self.dir, 'dir2') # simple name m1 = imp.importFromPath(jn(d1, 'mod.py'), 'mod') m2 = imp.importFromPath(jn(d2, 'mod.py'), 'mod') self.assertNotEqual(m1, m2) self.assertNotEqual(m1.__file__, m2.__file__) # dotted name p1 = imp.importFromPath(jn(d1, 'pak', 'mod.py'), 'pak.mod') p2 = imp.importFromPath(jn(d2, 'pak', 'mod.py'), 'pak.mod') self.assertNotEqual(p1, p2) self.assertNotEqual(p1.__file__, p2.__file__) # simple name -- package sp1 = imp.importFromPath(jn(d1, 'pak'), 'pak') sp2 = imp.importFromPath(jn(d2, 'pak'), 'pak') self.assertNotEqual(sp1, sp2) assert sp1.__path__ assert sp2.__path__ self.assertNotEqual(sp1.__path__, sp2.__path__) # dotted name -- package dp1 = imp.importFromPath(jn(d1, 'pak', 'sub'), 'pak.sub') dp2 = imp.importFromPath(jn(d2, 'pak', 'sub'), 'pak.sub') self.assertNotEqual(dp1, dp2) assert dp1.__path__ assert dp2.__path__ self.assertNotEqual(dp1.__path__, dp2.__path__) def test_import_sets_intermediate_modules(self): imp = self.imp path = os.path.join(self.dir, 'package2', 'test_pak', 'test_sub', 'test_mod.py') mod = imp.importFromPath(path, 'test_pak.test_sub.test_mod') print mod, dir(mod) assert 'test_pak' in sys.modules, 'test_pak was not imported?' test_pak = sys.modules['test_pak'] assert hasattr(test_pak, 'test_sub'), "test_pak.test_sub was not set" def test_cached_no_reload(self): imp = self.imp d1 = os.path.join(self.dir, 'dir1') m1 = imp.importFromDir(d1, 'mod') m2 = imp.importFromDir(d1, 'mod') assert m1 is m2, "%s is not %s" % (m1, m2) def test_cached_no_reload_dotted(self): imp = self.imp d1 = os.path.join(self.dir, 'dir1') p1 = imp.importFromDir(d1, 'pak.mod') p2 = imp.importFromDir(d1, 'pak.mod') assert p1 is p2, "%s is not %s" % (p1, p2) def test_import_sets_sys_modules(self): imp = self.imp d1 = os.path.join(self.dir, 'dir1') p1 = imp.importFromDir(d1, 'pak.mod') assert sys.modules['pak.mod'] is p1, "pak.mod not in sys.modules" assert sys.modules['pak'], "pak not in sys.modules" assert sys.modules['pak'].mod is p1, \ "sys.modules['pak'].mod is not the module we loaded" def test_failed_import_raises_import_error(self): imp = self.imp def bad_import(): imp.importFromPath(self.dir, 'no.such.module') self.assertRaises(ImportError, bad_import) def test_sys_modules_same_path_no_reload(self): imp = self.imp d1 = os.path.join(self.dir, 'dir1') d2 = os.path.join(self.dir, 'dir2') sys.path.insert(0, d1) mod_sys_imported = __import__('mod') mod_nose_imported = imp.importFromDir(d1, 'mod') assert mod_nose_imported is mod_sys_imported, \ "nose reimported a module in sys.modules from the same path" mod_nose_imported2 = imp.importFromDir(d2, 'mod') assert mod_nose_imported2 != mod_sys_imported, \ "nose failed to reimport same name, different dir" def test_import_pkg_from_path_fpw(self): imp = self.imp imp.config.firstPackageWins = True jn = os.path.join d1 = jn(self.dir, 'dir1') d2 = jn(self.dir, 'dir2') # dotted name p1 = imp.importFromPath(jn(d1, 'pak', 'mod.py'), 'pak.mod') p2 = imp.importFromPath(jn(d2, 'pak', 'mod.py'), 'pak.mod') self.assertEqual(p1, p2) self.assertEqual(p1.__file__, p2.__file__) # simple name -- package sp1 = imp.importFromPath(jn(d1, 'pak'), 'pak') sp2 = imp.importFromPath(jn(d2, 'pak'), 'pak') self.assertEqual(sp1, sp2) assert sp1.__path__ assert sp2.__path__ self.assertEqual(sp1.__path__, sp2.__path__) # dotted name -- package dp1 = imp.importFromPath(jn(d1, 'pak', 'sub'), 'pak.sub') dp2 = imp.importFromPath(jn(d2, 'pak', 'sub'), 'pak.sub') self.assertEqual(dp1, dp2) assert dp1.__path__ assert dp2.__path__ self.assertEqual(dp1.__path__, dp2.__path__) if __name__ == '__main__': import logging logging.basicConfig(level=logging.DEBUG) unittest.main()
6,043
2,214
from sklearn.base import clone import pandas as pd from abc import ABCMeta from time import time from datetime import datetime import numpy as np from sklearn.model_selection import ParameterGrid from sklearn.base import BaseEstimator, MetaEstimatorMixin from mvmm.utils import get_seeds from mvmm.multi_view.utils import linspace_zero_to, \ expspace_zero_to, polyspace_zero_to from mvmm.multi_view.block_diag.graph.linalg import geigh_Lsym_bp_smallest from mvmm.multi_view.block_diag.utils import asc_sort from mvmm.clustering_measures import unsupervised_cluster_scores, \ several_unsupervised_cluster_scores, MEASURE_MIN_GOOD class SpectralPenSearchByBlockMVMM(MetaEstimatorMixin, BaseEstimator, metaclass=ABCMeta): """ Does a grid search over the continuous hyper-parameter for the spentral penalized MVMM. Stores the best MVMM for each block. Parameters ---------- base_mvmm_0: Unconstrained MVMM. base_wbd_mvmm: mvmm.multi_view.BlockDiagMVMM.BlockDiagMVMM() The base class for the spectral penalized MVMM eval_weights: The weights to put on the generalized eigenvalues. adapt_expon: max_n_blocks: Maximum number of blocks to get i.e. the number of eigenvalues to penalized. user_eval_weights: (Optional) User provied eignvalue weights. pen_max: str, float Largest penalty value to try. If 'default' will make an automatic, educated guess. n_pen_seq: int Number of penalty values to try. user_pen_vals: None, list (Optional) User provided penalty values to try default_c: float Multiplicative factor for infering pen_max with the default method. pen_seq_spacing: str How to space the penalty values along the penalty sequence. n_init: int Number of random initalizations. random_state: None, int Random seed. select_metric: str How to pick the best model for each fixed number of blocks. metrics2compute: list of st Model selection measures to compute for tracking purposes. verbosity: int Level of printout """ def __init__(self, base_mvmm_0, base_wbd_mvmm, eval_weights='adapt', adapt_expon=1, max_n_blocks='default', user_eval_weights=None, pen_max='default', n_pen_seq=100, user_pen_vals=None, # adapt_pen=False, pen_incr=.5, max_n_pen_incr=200, default_c=100, pen_seq_spacing='lin', n_init=1, random_state=None, select_metric='bic', metrics2compute=['aic', 'bic'], verbosity=0): self.base_mvmm_0 = base_mvmm_0 self.base_wbd_mvmm = base_wbd_mvmm self.eval_weights = eval_weights self.adapt_expon = adapt_expon self.max_n_blocks = max_n_blocks self.user_eval_weights = user_eval_weights self.pen_max = pen_max self.n_pen_seq = n_pen_seq self.user_pen_vals = user_pen_vals self.default_c = default_c self.pen_seq_spacing = pen_seq_spacing assert pen_seq_spacing in ['lin', 'quad', 'exp'] # self.adapt_pen = adapt_pen # self.pen_incr = pen_incr # self.max_n_pen_incr = max_n_pen_incr # if self.adapt_pen: # assert self.user_pen_vals is None self.random_state = random_state self.n_init = n_init self.select_metric = select_metric self.metrics2compute = metrics2compute self.verbosity = verbosity def get_pen_seq_from_max(self, pen_max): if self.pen_seq_spacing == 'lin': return linspace_zero_to(stop=pen_max, num=self.n_pen_seq) elif self.pen_seq_spacing == 'quad': return polyspace_zero_to(stop=pen_max, num=self.n_pen_seq, deg=2) elif self.pen_seq_spacing == 'exp': return expspace_zero_to(stop=pen_max, num=self.n_pen_seq, base=10) @property def n_pen_vals_(self): if self.user_pen_vals is not None: return len(self.user_pen_vals) + 1 else: return self.n_pen_seq + 1 @property def param_grid_(self): """ List of all parameter settings """ if hasattr(self, 'est_n_blocks_'): param_grid = {'n_blocks': self.est_n_blocks_} return list(ParameterGrid(param_grid)) else: return None def get_default_pen_max(self, model, X): # steup temp model temp_model = clone(model) temp_model.view_models_ = \ [temp_model.base_view_models[v] for v in range(temp_model.n_views)] temp_model.initialize_parameters(X) eval_pen_default = temp_model.\ get_eval_pen_guess(X=X, c=self.default_c, use_bipt_sp=True, K='default') if self.verbosity >= 1: print('default pen val', eval_pen_default) return eval_pen_default def fit(self, X): # assert all(self.pen_vals_[1:] > 0) # assert len(np.unique(self.pen_vals)) == len(self.pen_vals) init_seeds = get_seeds(n_seeds=self.n_init, random_state=self.random_state) fit_data = pd.DataFrame() n_blocks_best_models = {} n_blocks_best_idx = {} init_adapt_weights = [] for init in range(self.n_init): if self.verbosity >= 1: current_time = datetime.now().strftime("%H:%M:%S") print('Initialization {}/{} at {}'. format(init + 1, self.n_init, current_time)) # max number of evals to penalize if self.max_n_blocks == 'default': K = min(self.base_mvmm_0.n_view_components) else: K = int(self.max_n_blocks) for pen_idx in range(self.n_pen_vals_): if self.verbosity >= 1: current_time = datetime.now().strftime("%H:%M:%S") print('Penalty {}/{} at {}'. format(pen_idx + 1, self.n_pen_vals_, current_time)) data = {'pen_idx': pen_idx, 'init': init} start_time = time() if pen_idx == 0: pen_val = None # fit model fit_model = clone(self.base_mvmm_0) fit_model.set_params(random_state=init_seeds[init], n_init=1) fit_model.fit(X) # get current parameter values for warm starting current_view_params = fit_model._get_parameters()['views'] current_bd_weights = fit_model.weights_mat_ current_bd_weights = current_bd_weights * \ self.base_wbd_mvmm.epsilon_tilde / \ current_bd_weights.sum() # track data data['n_blocks'] = 1 data['n_steps'] = fit_model.opt_data_['n_steps'] # compute adaptive weights if self.eval_weights == 'adapt': evals = geigh_Lsym_bp_smallest(X=self.bd_weights_, rank=K, zero_tol=1e-10, method='tsym') # deal with 0 evals by artificially setting # them to the smallest non-zero eval zero_evals = evals < 1e-6 if np.mean(zero_evals) == 1: # edge case: if all evals are 0 just use uiniform evals = np.ones(len(evals)) else: evals[zero_evals] = min(evals[~zero_evals]) # clip for numerical stability eval_weights = (1 / evals) ** self.adapt_expon init_adapt_weights.append(eval_weights) else: # setup and fit model fit_model = clone(self.base_wbd_mvmm) params = {'init_params_method': 'user', 'init_params_value': current_view_params, 'init_weights_method': 'user', 'init_weights_value': current_bd_weights # 'eval_pen_base': pen_val, } params.update({'n_pen_tries': 1, 'n_init': 1, # 'fine_tune_n_steps': None }) fit_model.set_params(**params) # set eval weights if self.user_eval_weights: eval_weights = self.user_eval_weights elif self.eval_weights == 'adapt': eval_weights = init_adapt_weights[init] elif self.eval_weights == 'uniform': eval_weights = np.ones(K) elif self.eval_weights == 'lin': eval_weights = 1 / np.arange(1, K + 1) elif self.eval_weights == 'quad': eval_weights = (1 / np.arange(1, K + 1)) ** 2 elif self.eval_weights == 'exp': eval_weights = .5 ** np.arange(1, K + 1) else: raise ValueError("invalid input for eval_weights: {}" .format(self.eval_weights)) def process(x): x = np.clip(x, a_min=0, a_max=1e5) return asc_sort(x * len(x) / np.sum(x)) # eval_weights = np.clip(eval_weights, a_min=0, a_max=1e5) # superficial normalization step keeps # penalty value reasonable # eval_weights *= K / np.sum(eval_weights) # eval_weights = desc_sort(eval_weights) eval_weights = process(eval_weights) fit_model.set_params(eval_weights=eval_weights) # set penalty sequence for this initialization if pen_idx == 1: if self.user_pen_vals is not None: pen_seq = np.sort(self.user_pen_vals) elif self.pen_max == 'default': # compute default max penalty default_pen_max = \ self.get_default_pen_max(model=fit_model, X=X) pen_seq = self.\ get_pen_seq_from_max(pen_max=default_pen_max) elif self.pen_max != 'default': pen_seq = self.\ get_pen_seq_from_max(pen_max=self.pen_max) pen_seq = np.concatenate([[None], pen_seq]) # set penalty value pen_val = pen_seq[pen_idx] fit_model.set_params(eval_pen_base=pen_val) fit_model.fit(X) # get current parameter values for warm starting current_view_params = fit_model._get_parameters()['views'] current_bd_weights = fit_model.bd_weights_ # track data data['pen_val'] = pen_val data['n_blocks'] = fit_model.opt_data_['n_blocks_est'] data['n_steps'] = \ fit_model.opt_data_['adpt_opt_data']['n_steps'] # store tracking data data['fit_time'] = time() - start_time tracking_data = fit_model.compute_tracking_data(X) data['loss_val'] = tracking_data['loss_val'] data['obs_nll'] = tracking_data['obs_nll'] # TODO: possibly precompute distances model_sel_scores = \ unsupervised_cluster_scores(X=X, estimator=fit_model, measures=self.metrics2compute) for measure in model_sel_scores.keys(): data[measure] = model_sel_scores[measure] # data['bic'] = fit_model.bic(X) # data['aic'] = fit_model.aic(X) fit_data = fit_data.append(data, ignore_index=True) # save this model if it is the best current_n_blocks = data['n_blocks'] # current n_blocks # get th block_scores = fit_data.query("n_blocks == @current_n_blocks") if MEASURE_MIN_GOOD[self.select_metric]: best_idx = block_scores[self.select_metric].idxmin() else: best_idx = block_scores[self.select_metric].idxmax() # best_idx = fit_data.\ # query("n_blocks == @n_blocks")[self.select_metric].\ # idxmin() if fit_data.loc[best_idx, 'init'] == init: n_blocks_best_models[current_n_blocks] = fit_model n_blocks_best_idx[current_n_blocks] = best_idx self.est_n_blocks_ = np.sort(list(n_blocks_best_models.keys())) self.estimators_ = [n_blocks_best_models[n_blocks] for n_blocks in self.est_n_blocks_] int_cols = ['init', 'pen_idx', 'n_blocks', 'n_steps'] fit_data[int_cols] = fit_data[int_cols].astype(int) self.init_fit_data_ = fit_data self.fit_init_best_idxs = [n_blocks_best_idx[n_blocks] for n_blocks in self.est_n_blocks_] if self.eval_weights == 'adapt': self.init_adapt_weights_ = init_adapt_weights self.model_sel_scores_ = \ several_unsupervised_cluster_scores(X=X, estimators=self.estimators_, measures=self.metrics2compute) return self def check_fit(self): return hasattr(self, 'estimators_') @property def best_idx_(self): """ Index of selected model. """ if self.check_fit(): if MEASURE_MIN_GOOD[self.select_metric]: return self.model_sel_scores_[self.select_metric].idxmin() else: return self.model_sel_scores_[self.select_metric].idxmax() else: return None @property def best_estimator_(self): """ Selected estimator. """ if self.check_fit(): return self.estimators_[self.best_idx_] else: return None def predict(self, X): """ Predict the labels for the data samples in X using trained model. """ return self.best_estimator_.predict(X) def predict_proba(self, X): """ Predict posterior probability of each component given the data. """ return self.best_estimator_.predict_proba(X) def sample(self, n_samples=1): """ Generate random samples from the fitted Gaussian distribution. """ return self.best_estimator_.sample(n_samples=n_samples) def score(self, X, y=None): """ Compute the per-sample average log-likelihood of the given data X. """ return self.best_estimator_.score(X) def score_samples(self, X): """ Compute the weighted log probabilities for each sample. """ return self.best_estimator_.score_samples(X)
16,278
4,774
"""Delegate provider traversal tests.""" from dependency_injector import providers def test_traversal_provider(): another_provider = providers.Provider() provider = providers.Delegate(another_provider) all_providers = list(provider.traverse()) assert len(all_providers) == 1 assert another_provider in all_providers def test_traversal_provider_and_overriding(): provider1 = providers.Provider() provider2 = providers.Provider() provider3 = providers.Provider() provider3.override(provider2) provider = providers.Delegate(provider1) provider.override(provider3) all_providers = list(provider.traverse()) assert len(all_providers) == 3 assert provider1 in all_providers assert provider2 in all_providers assert provider3 in all_providers
812
237
# -*- coding: utf-8 -*- """Top-level package for RHG Compute Tools.""" __author__ = """Michael Delgado""" __email__ = 'mdelgado@rhg.com' __version__ = '0.2.1'
161
72
# Generated by Django 3.2.3 on 2021-12-19 17:24 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('grocers_panel', '0004_shop_grocer'), ] operations = [ migrations.AlterField( model_name='shop', name='food', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='grocers_panel.food'), ), ]
481
174
class Colabore(): def getContext(self): return self.__contextColabore(self) def __contextColabore(self): context = { "grupo": "geral", "grupo_link": "saiba_mais", "titulo": "Observatório UFJ Covid-19 - Colabore" } return context
271
106
import datetime from sqlalchemy.orm import sessionmaker from database import db from database.order_history import OrderHistory from stock_analysis.logic import order_history from stock_analysis.logic.order_history import Order from stock_analysis.logic.order_history import OrderHistoryLogic from stock_analysis.logic.order_history import TickerDate Session = sessionmaker(bind=db.engine) class TestOrderHistoryLogic(object): @db.in_sandbox def test_add_buy_orders(self): logic = OrderHistoryLogic() user_id = 1 order = Order( user_id=user_id, order_type=order_history.BUY_ORDER_TYPE, date=datetime.date(2015, 8, 9), ticker='AAPL', num_shares=20, price=150.001, ) logic.add_orders([order]) session = Session() resp = session.query(OrderHistory).all() assert len(resp) == 1 order_in_db = resp[0] assert order_in_db.user_id == user_id assert order_in_db.date == order.date.isoformat() assert order_in_db.ticker == order.ticker assert order_in_db.num_shares == order.num_shares assert order_in_db.price == order.price session.close() @db.in_sandbox def test_get_tickers_and_min_dates_for_user(self): logic = OrderHistoryLogic() user_id = 1 order1 = Order( user_id=user_id, order_type=order_history.BUY_ORDER_TYPE, date=datetime.date(2015, 8, 9), ticker='AAPL', num_shares=20, price=150.002, ) order2 = Order( user_id=user_id, order_type=order_history.BUY_ORDER_TYPE, date=datetime.date(2017, 1, 1), ticker='AAPL', num_shares=20, price=152.333, ) logic.add_orders([order1, order2]) ticker_dates = logic.get_tickers_and_min_dates_for_user(user_id) assert ticker_dates == [TickerDate(order1.ticker, order1.date)] @db.in_sandbox def test_get_all_order_tickers_min_date(self): logic = OrderHistoryLogic() user_id_1 = 1 user_id_2 = 2 order1 = Order( user_id=user_id_1, order_type=order_history.BUY_ORDER_TYPE, date=datetime.date(2015, 8, 9), ticker='AAPL', num_shares=20, price=150.001, ) order2 = Order( user_id=user_id_2, order_type=order_history.BUY_ORDER_TYPE, date=datetime.date(2017, 1, 1), ticker='AAPL', num_shares=20, price=152.333, ) logic.add_orders([order1, order2]) ticker_dates = logic.get_all_order_tickers_min_date() assert ticker_dates == [TickerDate(order1.ticker, order1.date)] @db.in_sandbox def test_get_orders_for_user(self): logic = OrderHistoryLogic() order1 = Order( user_id=1, order_type=order_history.BUY_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 12), num_shares=2, price=150.0, ) order2 = Order( user_id=1, order_type=order_history.BUY_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 19), num_shares=3, price=170.0, ) order3 = Order( user_id=2, order_type=order_history.BUY_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 19), num_shares=3, price=170.0, ) logic.add_orders([order1, order2, order3]) assert set(logic.get_orders_for_user(1)) == set([order1, order2]) assert logic.get_ticker_to_orders(1) == {'AAPL': [order1, order2]} @db.in_sandbox def test_get_portfolio_shares_owned_on_date(self): logic = OrderHistoryLogic() order1 = Order( user_id=1, order_type=order_history.BUY_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 12), num_shares=2, price=150.0, ) order2 = Order( user_id=1, order_type=order_history.BUY_ORDER_TYPE, ticker='ATVI', date=datetime.date(2017, 6, 19), num_shares=3, price=170.0, ) order3 = Order( user_id=1, order_type=order_history.SELL_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 26), num_shares=1, price=180.0, ) logic.add_orders([order1, order2, order3]) results = logic.get_portfolio_shares_owned_on_date(order1.user_id, datetime.date(2017, 6, 27)) assert results == {'AAPL': 1, 'ATVI': 3} @db.in_sandbox def test_get_ticker_total_purchased_sold(self): logic = OrderHistoryLogic() order1 = Order( user_id=1, order_type=order_history.BUY_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 12), num_shares=2, price=150.0, ) order2 = Order( user_id=1, order_type=order_history.BUY_ORDER_TYPE, ticker='ATVI', date=datetime.date(2017, 6, 19), num_shares=3, price=170.0, ) order3 = Order( user_id=1, order_type=order_history.SELL_ORDER_TYPE, ticker='AAPL', date=datetime.date(2017, 6, 26), num_shares=1, price=180.0, ) order4 = Order( user_id=1, order_type=order_history.SELL_ORDER_TYPE, ticker='ATVI', date=datetime.date(2017, 6, 29), num_shares=3, price=120.0, ) logic.add_orders([order1, order2, order3, order4]) purchased, sold = logic.get_ticker_total_purchased_sold(order1.user_id) assert purchased == { 'AAPL': order1.num_shares * order1.price, 'ATVI': order2.num_shares * order2.price } assert sold == { 'AAPL': order3.num_shares * order3.price, 'ATVI': order4.num_shares * order4.price }
6,339
2,209
import os import tensorflow as tf from densenet_3d_model import DenseNet3D def model_fn(features, labels, mode, params): # Define the model model = DenseNet3D( video_clips=features['video_clips'], labels=labels, **params) # Get the prediction result if mode == tf.estimator.ModeKeys.PREDICT: model.is_training = False return _predict_result(model) return tf.estimator.EstimatorSpec( mode=mode, loss=model.losses, train_op=model.train_op, eval_metric_ops={'eval_accuracy': model.accuracy}) def _predict_result(model): predictions = {'prediction': model.prediction, 'probability': model.probability, 'logits': model.logits} return tf.estimator.EstimatorSpec( mode=tf.estimator.ModeKeys.PREDICT, predictions=predictions) def serving_input_fn(params): inputs = { 'video_clips': tf.placeholder( tf.float32, shape=[ None, params['num_frames_per_clip'], params['height'], params['width'], params['channel'] ]) } return tf.estimator.export.build_raw_serving_input_receiver_fn(inputs)() def train_input_fn(training_dir, params): directory = os.path.join(training_dir, 'train.tfrecord') return _build_tfrecord_dataset(directory, params['train_total_video_clip'], **params) def eval_input_fn(evaluating_dir, params): directory = os.path.join(evaluating_dir, 'eval.tfrecord') return _build_tfrecord_dataset(directory, params['eval_total_video_clip'], **params) def _build_tfrecord_dataset(directory, total_clip_num, batch_size, **params): ''' Buffer the training dataset to TFRecordDataset with the following video shape [num_frames_per_clip, height, width, channel] ex: [16, 100, 120, 3] ''' print('Building dataset, number of clips: ' + str(total_clip_num)) dataset = tf.data.TFRecordDataset(directory) dataset = dataset.shuffle(buffer_size=total_clip_num) dataset = dataset.map( map_func= lambda serialized_example: _parser(serialized_example, **params)) dataset = dataset.repeat() iterator = dataset.batch(batch_size=batch_size).make_one_shot_iterator() clips, labels = iterator.get_next() return {'video_clips': clips}, labels def _parser(serialized_example, num_frames_per_clip, **params): features = tf.parse_single_example( serialized_example, features={ 'clip/width': tf.FixedLenFeature([], tf.int64), 'clip/height': tf.FixedLenFeature([], tf.int64), 'clip/channel': tf.FixedLenFeature([], tf.int64), 'clip/raw': tf.FixedLenFeature([num_frames_per_clip], tf.string), 'clip/label': tf.FixedLenFeature([], tf.int64) }) def mapping_func(image): return _decode_image(image, **params) clip = tf.map_fn(mapping_func, features['clip/raw'], dtype=tf.float32) return clip, features['clip/label'] def _decode_image(image, channel, width, height, **params): image = tf.image.decode_jpeg(image, channels=channel) # This set_shape step is necesary for the last trainsition_layer_to_classes layer in the model image.set_shape([height, width, channel]) image = tf.cast(image, tf.float32) return image
3,380
1,058