text
string
size
int64
token_count
int64
n = int(input('Qual tabuada deseja ver: ')) c=1 print(11*'=') while c <= 10: print('{} x {:2} = {}'.format(n,c,c*n)) c += 1 print(11*'=')
146
77
import argparse import json import os import matplotlib.pyplot as plt import numpy as np import tensorflow as tf import tqdm from config import Config from dataset import LJSpeech from model import DiffWave class Trainer: """WaveGrad trainer. """ def __init__(self, model, lj, config): """Initializer. Args: model: DiffWave, diffwave model. lj: LJSpeech, LJ-speec dataset which provides already batched and normalized speech dataset. config: Config, unified configurations. """ self.model = model self.lj = lj self.config = config self.split = config.train.split // config.data.batch self.trainset = self.lj.dataset().take(self.split) \ .shuffle(config.train.bufsiz) \ .prefetch(tf.data.experimental.AUTOTUNE) self.testset = self.lj.dataset().skip(self.split) \ .prefetch(tf.data.experimental.AUTOTUNE) self.optim = tf.keras.optimizers.Adam( config.train.lr(), config.train.beta1, config.train.beta2, config.train.eps) self.eval_intval = config.train.eval_intval // config.data.batch self.ckpt_intval = config.train.ckpt_intval // config.data.batch self.train_log = tf.summary.create_file_writer( os.path.join(config.train.log, config.train.name, 'train')) self.test_log = tf.summary.create_file_writer( os.path.join(config.train.log, config.train.name, 'test')) self.ckpt_path = os.path.join( config.train.ckpt, config.train.name, config.train.name) self.alpha_bar = np.cumprod(1 - config.model.beta()) self.cmap = tf.constant(plt.get_cmap('viridis').colors, dtype=tf.float32) def compute_loss(self, signal, logmel): """Compute loss for noise estimation. Args: signal: tf.Tensor, [B, T], raw audio signal segment. logmel: tf.Tensor, [B, T // hop, mel], mel-spectrogram. Returns: loss: tf.Tensor, [], L1-loss between noise and estimation. """ # [B] bsize = tf.shape(signal)[0] # [B] timesteps = tf.random.uniform( [bsize], 1, self.config.model.iter + 1, dtype=tf.int32) # [B] noise_level = tf.gather(self.alpha_bar, timesteps - 1) # [B, T], [B, T] noised, noise = self.model.diffusion(signal, noise_level) # [B, T] eps = self.model.pred_noise(noised, timesteps, logmel) # [] loss = tf.reduce_mean(tf.abs(eps - noise)) return loss def train(self, step=0, ir_unit=5): """Train wavegrad. Args: step: int, starting step. ir_unit: int, log ir units. """ for _ in tqdm.trange(step // self.split, self.config.train.epoch): with tqdm.tqdm(total=self.split, leave=False) as pbar: for signal, logmel in self.trainset: with tf.GradientTape() as tape: tape.watch(self.model.trainable_variables) loss = self.compute_loss(signal, logmel) grad = tape.gradient(loss, self.model.trainable_variables) self.optim.apply_gradients( zip(grad, self.model.trainable_variables)) norm = tf.reduce_mean([tf.norm(g) for g in grad]) del grad step += 1 pbar.update() pbar.set_postfix( {'loss': loss.numpy().item(), 'step': step, 'grad': norm.numpy().item()}) with self.train_log.as_default(): tf.summary.scalar('loss', loss, step) tf.summary.scalar('grad norm', norm, step) if step % self.eval_intval == 0: pred, _ = self.model(logmel) tf.summary.audio( 'train', pred[..., None], self.config.data.sr, step) tf.summary.image( 'train mel', self.mel_img(pred), step) del pred if step % self.ckpt_intval == 0: self.model.write( '{}_{}.ckpt'.format(self.ckpt_path, step), self.optim) loss = [ self.compute_loss(signal, logmel).numpy().item() for signal, logmel in self.testset ] loss = sum(loss) / len(loss) with self.test_log.as_default(): tf.summary.scalar('loss', loss, step) gt, pred, ir = self.eval() tf.summary.audio( 'gt', gt[None, :, None], self.config.data.sr, step) tf.summary.audio( 'eval', pred[None, :, None], self.config.data.sr, step) tf.summary.image( 'gt mel', self.mel_img(gt[None]), step) tf.summary.image( 'eval mel', self.mel_img(pred[None]), step) for i in range(0, len(ir), ir_unit): tf.summary.audio( 'ir_{}'.format(i), np.clip(ir[i][None, :, None], -1., 1.), self.config.data.sr, step) del gt, pred, ir def mel_img(self, signal): """Generate mel-spectrogram images. Args: signal: tf.Tensor, [B, T], speech signal. Returns: tf.Tensor, [B, mel, T // hop, 3], mel-spectrogram in viridis color map. """ # [B, T // hop, mel] _, mel = self.lj.mel_fn(signal) # [B, mel, T // hop] mel = tf.transpose(mel, [0, 2, 1]) # minmax norm in range(0, 1) mel = (mel - tf.reduce_min(mel)) / (tf.reduce_max(mel) - tf.reduce_min(mel)) # in range(0, 255) mel = tf.cast(mel * 255, tf.int32) # [B, mel, T // hop, 3] mel = tf.gather(self.cmap, mel) # make origin lower mel = tf.image.flip_up_down(mel) return mel def eval(self): """Generate evaluation purpose audio. Returns: speech: np.ndarray, [T], ground truth. pred: np.ndarray, [T], predicted. ir: List[np.ndarray], config.model.iter x [T], intermediate represnetations. """ # [T] speech = next(iter(lj.rawset)) # [1, T // hop, mel] _, logmel = lj.mel_fn(speech[None]) # [1, T], iter x [1, T] pred, ir = self.model(logmel) # [T] pred = tf.squeeze(pred, axis=0).numpy() # config.model.iter x [T] ir = [np.squeeze(i, axis=0) for i in ir] return speech.numpy(), pred, ir if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument('--config', default=None) parser.add_argument('--load-step', default=0, type=int) parser.add_argument('--ir-unit', default=10, type=int) parser.add_argument('--data-dir', default=None) parser.add_argument('--download', default=False, action='store_true') parser.add_argument('--from-raw', default=False, action='store_true') args = parser.parse_args() config = Config() if args.config is not None: print('[*] load config: ' + args.config) with open(args.config) as f: config = Config.load(json.load(f)) log_path = os.path.join(config.train.log, config.train.name) if not os.path.exists(log_path): os.makedirs(log_path) ckpt_path = os.path.join(config.train.ckpt, config.train.name) if not os.path.exists(ckpt_path): os.makedirs(ckpt_path) lj = LJSpeech(config.data, args.data_dir, args.download, not args.from_raw) diffwave = DiffWave(config.model) trainer = Trainer(diffwave, lj, config) if args.load_step > 0: super_path = os.path.join(config.train.ckpt, config.train.name) ckpt_path = '{}_{}.ckpt'.format(config.train.name, args.load_step) ckpt_path = next( name for name in os.listdir(super_path) if name.startswith(ckpt_path) and name.endswith('.index')) ckpt_path = os.path.join(super_path, ckpt_path[:-6]) print('[*] load checkpoint: ' + ckpt_path) trainer.model.restore(ckpt_path, trainer.optim) with open(os.path.join(config.train.ckpt, config.train.name + '.json'), 'w') as f: json.dump(config.dump(), f) trainer.train(args.load_step, args.ir_unit)
8,782
2,788
# boudoir - chuchotement pantophobique # https://www.youtube.com/watch?v=KL2zW6Q5hWs # https://gist.github.com/jf-parent/c8ea7e54e30593af01512f4e21b54670 Scale.default = Scale.major Root.default = 0 Clock.bpm = 120 b1.reset() >> glass( [0], dur = 16, ).after(16, 'stop') def play1(): print('play1') p1.reset() >> bass( [0, [0, 1, 2]], dur = 8, oct = 5, lpf = 600, room = 1, mix = 1, ).after(320, 'stop') def play2(): print('play2') p2.reset() >> soft( [0], formant = linvar([0, 1], 8), sus = 2, vib = linvar([0, 2], 4), oct = var([3, 4, 5], 4), ).every( 16, 'stutter' ).after(290, 'stop') def play3(): print('play3') p3.reset() >> sawbass( P[0, 2, 4, [5, 6, 7]], dur = [1, [1, 2], 2, [2, 4]], oct = [4, 5, 6, 6], vib = [0, 0, 0, [0, 1]], formant = var([0, 1], 8), ).after(48, 'stop') Clock.set_time(0) Clock.future(0, play1) Clock.future(30, play2) Clock.future(60, play3) Clock.future(120, play3) Clock.future(240, play3)
1,122
555
import random import urllib.parse import sqlite3 import asyncio import aiohttp import discord from discord.ext import commands import loadconfig class fun(commands.Cog): def __init__(self, bot): self.bot = bot async def cog_command_error(self, ctx, error): print('Error in {0.command.qualified_name}: {1}'.format(ctx, error)) def userOnline(self, memberList): online = [] for i in memberList: if i.status == discord.Status.online and i.bot == False: online.append(i) return online @commands.command(aliases=['javascript', 'nodejs', 'js']) async def java(self, ctx): '''Weil Java != Javscript''' await ctx.send(':interrobang: Meintest du jQuery, Javascript oder Node.js? https://abload.de/img/2016-05-102130191kzpu.png') @commands.command(aliases=['c++', 'c#', 'objective-c']) async def csharp(self, ctx): '''Wie soll man da überhaupt durchblicken???''' await ctx.send(':interrobang: Meintest du C, C++, C# oder Objective-C? https://i.imgur.com/Nd4aAXO.png') @commands.command() async def praise(self, ctx): '''Praise the Sun''' await ctx.send('https://i.imgur.com/K8ySn3e.gif') @commands.command() async def css(self, ctx): '''Counter Strike: Source''' await ctx.send('http://i.imgur.com/TgPKFTz.gif') @commands.command() async def countdown(self, ctx): '''It's the final countdown''' countdown = ['five', 'four', 'three', 'two', 'one'] for num in countdown: await ctx.send('**:{0}:**'.format(num)) await asyncio.sleep(1) await ctx.send('**:ok:** DING DING DING') @commands.command(aliases=['cat', 'randomcat']) async def neko(self, ctx): '''Zufällige Katzen Bilder nyan~''' #http://discordpy.readthedocs.io/en/latest/faq.html#what-does-blocking-mean async with aiohttp.ClientSession() as cs: async with cs.get('http://aws.random.cat/meow') as r: res = await r.json() emojis = [':cat2: ', ':cat: ', ':heart_eyes_cat: '] await ctx.send(random.choice(emojis) + res['file']) @commands.command(aliases=['rand']) async def random(self, ctx, *arg): '''Gibt eine zufällige Zahl oder Member aus Benutzung: ----------- :random Gibt eine zufällige Zahl zwischen 1 und 100 aus :random coin Wirft eine Münze (Kopf oder Zahl) :random 6 Gibt eine zufällige Zahl zwischen 1 und 6 aus :random 10 20 Gibt eine zufällige Zahl zwischen 10 und 20 aus :random user Gibt einen zufällige Benutzer der gerade online ist aus :random choice Dani Eddy Shinobu Wählt aus der vorgegebenen Liste einen Namen aus ''' if ctx.invoked_subcommand is None: if not arg: start = 1 end = 100 elif arg[0] == 'flip' or arg[0] == 'coin': coin = ['Kopf', 'Zahl'] await ctx.send(f':arrows_counterclockwise: {random.choice(coin)}') return elif arg[0] == 'choice': choices = list(arg) choices.pop(0) await ctx.send(f':congratulations: The winner is {random.choice(choices)}') return elif arg[0] == 'user': online = self.userOnline(ctx.guild.members) randomuser = random.choice(online) if ctx.channel.permissions_for(ctx.author).mention_everyone: user = randomuser.mention else: user = randomuser.display_name await ctx.send(f':congratulations: The winner is {user}') return elif len(arg) == 1: start = 1 end = int(arg[0]) elif len(arg) == 2: start = int(arg[0]) end = int(arg[1]) await ctx.send(f'**:arrows_counterclockwise:** Zufällige Zahl ({start} - {end}): {random.randint(start, end)}') @commands.command() async def steinigt(self, ctx, member:str): '''Monty Python''' await ctx.send(f'R.I.P. {member}\nhttps://media.giphy.com/media/l41lGAcThnMc29u2Q/giphy.gif') @commands.command(aliases=['hypu', 'train']) async def hype(self, ctx): '''HYPE TRAIN CHOO CHOO''' hypu = ['https://cdn.discordapp.com/attachments/102817255661772800/219514281136357376/tumblr_nr6ndeEpus1u21ng6o1_540.gif', 'https://cdn.discordapp.com/attachments/102817255661772800/219518372839161859/tumblr_n1h2afSbCu1ttmhgqo1_500.gif', 'https://gfycat.com/HairyFloweryBarebirdbat', 'https://i.imgur.com/PFAQSLA.gif', 'https://abload.de/img/ezgif-32008219442iq0i.gif', 'https://i.imgur.com/vOVwq5o.jpg', 'https://i.imgur.com/Ki12X4j.jpg', 'https://media.giphy.com/media/b1o4elYH8Tqjm/giphy.gif'] msg = f':train2: CHOO CHOO {random.choice(hypu)}' await ctx.send(msg) @commands.command() async def xkcd(self, ctx, *searchterm: str): '''Zeigt den letzten oder zufälligen XKCD Comic Beispiel: ----------- :xkcd :xkcd random ''' apiUrl = 'https://xkcd.com{}info.0.json' async with aiohttp.ClientSession() as cs: async with cs.get(apiUrl.format('/')) as r: js = await r.json() if ''.join(searchterm) == 'random': randomComic = random.randint(0, js['num']) async with cs.get(apiUrl.format('/' + str(randomComic) + '/')) as r: if r.status == 200: js = await r.json() comicUrl = 'https://xkcd.com/{}/'.format(js['num']) date = '{}.{}.{}'.format(js['day'], js['month'], js['year']) msg = '**{}**\n{}\nAlt Text:```{}```XKCD Link: <{}> ({})'.format(js['safe_title'], js['img'], js['alt'], comicUrl, date) await ctx.send(msg) @commands.command(aliases=['witz', 'joke']) async def pun(self, ctx): '''Weil jeder schlechte Witze mag''' #ToDo: Add some way to fetch https://github.com/derphilipp/Flachwitze puns = ['Was sagt das eine Streichholz zum anderen Streichholz?\n Komm, lass uns durchbrennen', 'Wieviele Deutsche braucht man um eine Glühbirne zu wechseln?\n Einen, wir sind humorlos und effizient.', 'Wo wohnt die Katze?\n Im Miezhaus.', 'Wie begrüßen sich zwei plastische Chirurgen?\n "Was machst du denn heute für ein Gesicht?"', 'Warum essen Veganer kein Huhn?\n Könnte Ei enthalten', '85% der Frauen finden ihren Arsch zu dick, 10% zu dünn, 5% finden ihn so ok, wie er ist und sind froh, dass sie ihn geheiratet haben...', 'Meine Freundin meint, ich wär neugierig...\n...zumindest\' steht das in ihrem Tagebuch.', '"Schatz, Ich muss mein T-Shirt waschen! Welches Waschmaschinen Programm soll ich nehmen?" - "Was steht denn auf dem T-Shirt drauf?"\n "Slayer!"', 'Gestern erzählte ich meinem Freund, dass ich schon immer dieses Ding aus Harry Potter reiten wollte.\n"einen Besen?" "nein, Hermine."', 'Warum gehen Ameisen nicht in die Kirche?\nSie sind in Sekten.', 'Was steht auf dem Grabstein eines Mathematikers?\n"Damit hat er nicht gerechnet."', 'Wenn ein Yogalehrer seine Beine senkrecht nach oben streckt und dabei furzt, welche Yoga Figur stellt er da?\n Eine Duftkerze', 'Warum ging der Luftballon kaputt?\n Aus Platzgründen.', 'Ich wollte Spiderman anrufen, aber er hatte kein Netz und beim Bäcker war alles belegt.', 'Was vermisst eine Schraube am meisten? Einen Vater', 'Geht ein Panda über die Straße. Bam....Bus!', 'Unterhalten sich zwei Gletscher. Sagt der eine: "Was meinst du, was wird die Zukunft bringen?" Sagt der Andere: "Naja, wir werden Seen."', 'Wenn sich ein Professor ein Brot macht ist das dann wissenschaftlich belegt?', 'Knabbern zwei Männern an einer Eisenbahnschiene. Sagt der eine: "Ganz schön hart, oder?"\nSagt der andere: "Aber guck mal, da drübern ist ne Weiche"', 'Warum sind bei IKEA Pfeile auf dem Boden?\nWeil es ein Einrichtungshaus ist', 'Was macht die Security in der Nudelfabrik?\nDie Pasta auf.', 'Wie nennt man einen kleinwüchsigen Securitymenschen?\nSicherheitshalber', 'Habe bei Weight Watchers angerufen. Hat keiner abgenommen.\nDanach beim DJ. Hat aber aufgelegt.' 'Meine Schwester hat eine Tochter bekommen.\nDa wurde mein Wunsch nach einem Neffen zur Nichte gemacht.', 'Praktizieren sie Inzest?\n"Mitnichten"', 'Wann sinkt ein U-Boot?\nAm Tag der offenen Tür.', 'Auf St. Pauli wurde letztens ein Sarg gefunden. Er konnte aber nicht geöffnet werden, war ein Zuhälter drin!', 'Treffen sich zwei Anwälte. Fragt der eine "Na, wie geht\'s?" Antwortet der andere "Schlecht. Ich kann nicht klagen"', 'Treffen sich zwei Jäger. Beide tot.', 'Treffen sich zwei Päpste.', 'Treffen sich zwei Psychologen, sagt der eine: "Dir geht\'s gut, wie geht\'s mir?"', 'Treffen sich zwei Linksextreme in einer Bar, kommen drei Splittergruppen raus.', 'Was macht man mit nem Hund ohne Beine?\nUm die Häuser ziehen.', 'Wo findest du nen Hund ohne Beine?\nDa wo du ihn liegen lassen hast.', 'Was macht eine Bombe im Bordell?\nPuff', 'Und was macht eine Bombe im Treppenhaus?\nHochgehen', 'Wo war Lucy nach der Explosion?\nÜberall', 'Egal, wie dicht du bist. Göthe war dichter!', 'Egal, wie gut du fährst. Züge fahren Güter!', 'Egal, wie sauer du bist, Dinos sind Saurier!', 'Egal, wie leer du bist, es gibt Menschen die sind Lehrer.', 'Wissenschaftler haben herausgefunden\nund sind dann wieder reingegangen.', 'Was ist klein, braun, rund und sitzt hinter Gittern? Eine Knastanie.', 'Was liegt am Strand und kann nicht richtig reden? - Eine Nuschel!', 'Was ist grün und klopft an die Tür? - Klopfsalat', 'Was ist rot und steht am Straßenrand? Eine Hagenutte', 'Und was ist blau und steht am Wegesrand? Eine Frostituierte', 'Was ist rosa und schwimmt durchs Meer? Eine Meerjungsau.', 'Was ist braun und schwimmt auch im Meer? Ein U-Brot.', 'Was raucht und springt durch den Wald? Ein Kaminchen.', 'Was machen Bits am liebsten? Busfahren.', 'Warum ist der Programmierer in der Dusche gestorben? Auf der Flasche stand “einschäumen, ausspülen, wiederholen"', 'Wo gehen Datenspeicher hin, wenn sie sich prügeln wollen? In den Byte Club.\n Und Regel Nummer Eins: Ihr verliert kein dword über den Byte Club!', 'Wer wohnt im Dschungel und schummelt? Mogli', 'Geht ein Mann zum Arzt weil er sich schlecht fühlt. Sagt der Arzt: "Sie müssen mit dem Masturbieren aufhören!"\nSagt der Mann: "Wieso das denn?!".\nSagt der Arzt: "Ja, sonst kann ich Sie nicht untersuchen."', 'Wie heißt ein Spanier ohne Auto?\nCarlos', 'Wie nennt man ein Cowboy ohne Pferd?\nSattelschlepper', 'Kommt ein Cowboy aus dem Frisiersalon heraus\nPony weg', 'Wie nennt man einen Schäfer, der seine Schafe schlägt?\nMähdrescher', 'Was trinkt die Chefin?\nLeitungswasser', 'Vampir in der Verkehrskontrolle.\n"Haben Sie was getrunken?"\n"Ja, zwei Radler."', 'Wie nennt man jemanden, der DIN A4 Blätter scannt?\nScandinavier', 'Wie nennt man einen Europäer aus Lettland?\nEuropalette', 'Hab nem Hipster ins Bein geschossen\nJetzt hopster', 'Wie viel wiegt ein Influencer?\nEin Instagramm', 'Was ist gelb und kann nicht schwimmen?\nEin Bagger\nUnd warum nicht?\nHat nur einen Arm', 'Was hat ein Mann ohne Beine?\nErdnüsse', 'Welcher Vogel hat Darth Vader auf denn Kopf geschissen?\nDer Star wars', 'Wie heißt ein Veganer Russe?\nMooskauer', 'Was ist der Unterschied zwischen Grießbrei und einem Epileptiker?\nDer Grießbrei liegt in Zucker und Zimt, der Epileptiker liegt im Zimmer und zuckt.', 'Was macht ein Clown im Büro?\nFaxen', 'Was ist grūn und nuschelt im Gurkensalat?\nDill Schweiger', 'Was ist die Vergangenheitsform von Tomate? Passierte Tomate', 'Gehören abgetriebene Babys eigentlich auch zu den entfernen Verwandten?', 'Kommt ein Dachdecker in ne Bar\nDa sagt der Barkeeper: "Der geht aufs Haus!"', 'Was spricht man in der Sauna? Schwitzerdeutsch.', 'Was ist grün und wird auf Knopfdruck rot?\nEin Frosch im Mixer', 'Was ist weiß und fliegt über die Wiese?\nBiene Majo', 'Warum trinken Veganer kein Leitungswasser?\nWeil es aus dem Hahn kommt'] emojis = [':laughing:', ':smile:', ':joy:', ':sob:', ':rofl:'] msg = f'{random.choice(emojis)} {random.choice(puns)}' await ctx.send(msg) def setup(bot): bot.add_cog(fun(bot))
14,101
4,655
# !/usr/bin/env python # -*- coding: utf-8 -*- import base64 import json import copy import socket import subprocess import six class SSR: class Service: def __init__(self, conf): self.conf = conf def update(self, array): self.conf.update(array) def port_open(self, first=True): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) for i in range(2): try: s.connect(('127.0.0.1', int(self.conf['port']))) s.shutdown(2) return True except socket.error: if first: self.restart() return self.port_open(first=False) return False return False def restart(self): if 'restart' in self.conf: return subprocess.call(self.conf['restart'], shell=True) return def get_data(self): if not self.port_open(): return None self.conf.pop('restart') return base64.urlsafe_b64encode(json.dumps(self.conf).encode()).decode() def __init__(self, conf, host, group, remarks, restart): self.data_list = [] self.conf = conf with open(self.conf, 'r') as f: try: self.config = json.load(f) except Exception: raise ValueError('SSR config is not json: %s' % conf) self.host = host self.group = group self.remarks = remarks self.restart = restart def get_services(self): if ('server_port' not in self.config or 'password' not in self.config) and 'port_password' not in self.config: raise KeyError('SSR config is incorrect') base_service = SSR.Service( { 'host': self.host, 'protocol': self.config.get('protocol', 'origin'), 'protoparam': self.config.get('protocol_param', ''), 'method': self.config.get('method', 'none'), 'obfs': self.config.get('obfs', 'plain'), 'obfsparam': self.config.get('obfs_param', ''), 'remarks': self.remarks, 'group': self.group, 'restart': self.restart, 'password': '', 'port': 0 } ) if 'port_password' in self.config: for port, data in six.iteritems(self.config['port_password']): service = copy.deepcopy(base_service) if type(data) is str: array = {'password': data} elif type(data) is dict: array = data else: continue array['remarks'] = self.remarks + ('_%s' % str(port)) array['port'] = port service.update(array) url = service.get_data() if url is None: continue self.data_list.append(url) else: base_service.update({'port': self.config['server_port'], 'password': self.config['password']}) url = base_service.get_data() if url is not None: self.data_list.append(url) return self.data_list
3,362
912
"""Build beta detail models for the API""" import enum from typing import Dict, Optional import deserialize from asconnect.models.common import BaseAttributes, Links, Relationship, Resource class ExternalBetaState(enum.Enum): """External beta state.""" PROCESSING = "PROCESSING" PROCESSING_EXCEPTION = "PROCESSING_EXCEPTION" MISSING_EXPORT_COMPLIANCE = "MISSING_EXPORT_COMPLIANCE" READY_FOR_BETA_TESTING = "READY_FOR_BETA_TESTING" IN_BETA_TESTING = "IN_BETA_TESTING" EXPIRED = "EXPIRED" READY_FOR_BETA_SUBMISSION = "READY_FOR_BETA_SUBMISSION" IN_EXPORT_COMPLIANCE_REVIEW = "IN_EXPORT_COMPLIANCE_REVIEW" WAITING_FOR_BETA_REVIEW = "WAITING_FOR_BETA_REVIEW" IN_BETA_REVIEW = "IN_BETA_REVIEW" BETA_REJECTED = "BETA_REJECTED" BETA_APPROVED = "BETA_APPROVED" class InternalBetaState(enum.Enum): """Internal beta state.""" PROCESSING = "PROCESSING" PROCESSING_EXCEPTION = "PROCESSING_EXCEPTION" MISSING_EXPORT_COMPLIANCE = "MISSING_EXPORT_COMPLIANCE" READY_FOR_BETA_TESTING = "READY_FOR_BETA_TESTING" IN_BETA_TESTING = "IN_BETA_TESTING" EXPIRED = "EXPIRED" IN_EXPORT_COMPLIANCE_REVIEW = "IN_EXPORT_COMPLIANCE_REVIEW" @deserialize.key("identifier", "id") class BuildBetaDetail(Resource): """Represents a build localization.""" @deserialize.key("auto_notify_enabled", "autoNotifyEnabled") @deserialize.key("external_build_state", "externalBuildState") @deserialize.key("internal_build_state", "internalBuildState") class Attributes(BaseAttributes): """Represents beta build localization attributes.""" auto_notify_enabled: bool external_build_state: ExternalBetaState internal_build_state: InternalBetaState identifier: str attributes: Attributes relationships: Optional[Dict[str, Relationship]] links: Links
1,866
709
import math from plotter import Plotter from plots import LinePlot import board import digitalio import busio import adafruit_sdcard import storage from adafruit_bitmapsaver import save_pixels def plot(): sines = list(math.sin(math.radians(x)) for x in range(0, 361, 4)) lineplot = LinePlot([sines],'MicroPlot line') plotter = Plotter() lineplot.plot(plotter) def save(): spi = busio.SPI(board.SCK, MOSI=board.MOSI, MISO=board.MISO) cs = digitalio.DigitalInOut(board.SD_CS) sdcard = adafruit_sdcard.SDCard(spi, cs) vfs = storage.VfsFat(sdcard) storage.mount(vfs, "/sd") save_pixels("/sd/screenshot2.bmp") plot() #save() print('done') #import jax.numpy as np # #def periodic_spikes(firing_periods, duration: int): # return 0 == (1 + np.arange(duration))[:, None] % firing_periods # # #periodic_spikes(5, 22)
922
373
from os import listdir import subprocess for f in listdir("tests/vulkan"): if f.endswith(".spv"): continue print(f"-- compiling test {f}") p = subprocess.run(["glslangValidator", f"tests/vulkan/{f}", "-H", "-o", f"tests/vulkan/{f}.spv"], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) if p.returncode != 0: print(bytes.decode(p.stdout))
385
143
# Taken straight from Patter https://github.com/ryanleary/patter # TODO: review, and copyright and fix/add comments import torch from torch.utils.data import Dataset from .manifest import Manifest def seq_collate_fn(batch): def find_max_len(seq, index): max_len = -1 for item in seq: if item[index].size(0) > max_len: max_len = item[index].size(0) return max_len batch_size = len(batch) audio_signal, audio_lengths = None, None if batch[0][0] is not None: max_audio_len = find_max_len(batch, 0) audio_signal = torch.zeros(batch_size, max_audio_len, dtype=torch.float) audio_lengths = [] for i, s in enumerate(batch): audio_signal[i].narrow(0, 0, s[0].size(0)).copy_(s[0]) audio_lengths.append(s[1]) audio_lengths = torch.tensor(audio_lengths, dtype=torch.long) max_transcript_len = find_max_len(batch, 2) transcript = torch.zeros(batch_size, max_transcript_len, dtype=torch.long) transcript_lengths = [] for i, s in enumerate(batch): transcript[i].narrow(0, 0, s[2].size(0)).copy_(s[2]) transcript_lengths.append(s[3]) transcript_lengths = torch.tensor(transcript_lengths, dtype=torch.long) return audio_signal, audio_lengths, transcript, transcript_lengths def audio_seq_collate_fn(batch): """ collate a batch (iterable of (sample tensor, label tensor) tuples) into properly shaped data tensors :param batch: :return: inputs (batch_size, num_features, seq_length), targets, input_lengths, target_sizes """ # sort batch by descending sequence length (for packed sequences later) batch.sort(key=lambda x: -x[0].size(0)) minibatch_size = len(batch) # init tensors we need to return inputs = torch.zeros(minibatch_size, batch[0][0].size(0)) input_lengths = torch.zeros(minibatch_size, dtype=torch.long) target_sizes = torch.zeros(minibatch_size, dtype=torch.long) targets = [] metadata = [] # iterate over minibatch to fill in tensors appropriately for i, sample in enumerate(batch): input_lengths[i] = sample[0].size(0) inputs[i].narrow(0, 0, sample[0].size(0)).copy_(sample[0]) target_sizes[i] = len(sample[1]) targets.extend(sample[1]) metadata.append(sample[2]) targets = torch.tensor(targets, dtype=torch.long) return inputs, targets, input_lengths, target_sizes, metadata class AudioDataset(Dataset): def __init__(self, manifest_filepath, labels, featurizer, max_duration=None, min_duration=None, max_utts=0, normalize=True, trim=False, eos_id=None, logger=False, load_audio=True): """ Dataset that loads tensors via a json file containing paths to audio files, transcripts, and durations (in seconds). Each new line is a different sample. Example below: {"audio_filepath": "/path/to/audio.wav", "text_filepath": "/path/to/audio.txt", "duration": 23.147} ... {"audio_filepath": "/path/to/audio.wav", "text": "the transcription", offset": 301.75, "duration": 0.82, "utt": "utterance_id", "ctm_utt": "en_4156", "side": "A"} Args: manifest_filepath: Path to manifest json as described above. Can be coma-separated paths. labels: String containing all the possible characters to map to featurizer: Initialized featurizer class that converts paths of audio to feature tensors max_duration: If audio exceeds this length, do not include in dataset min_duration: If audio is less than this length, do not include in dataset max_utts: Limit number of utterances normalize: whether to normalize transcript text (default): True eos_id: Id of end of sequence symbol to append if not None load_audio: Boolean flag indicate whether do or not load audio """ m_paths = manifest_filepath.split(',') self.manifest = Manifest(m_paths, labels, max_duration=max_duration, min_duration=min_duration, max_utts=max_utts, normalize=normalize) self.featurizer = featurizer self.trim = trim self.eos_id = eos_id self.load_audio = load_audio if logger: logger.info( "Dataset loaded with {0:.2f} hours. Filtered {1:.2f} " "hours.".format( self.manifest.duration / 3600, self.manifest.filtered_duration / 3600)) def __getitem__(self, index): sample = self.manifest[index] if self.load_audio: duration = sample['duration'] if 'duration' in sample else 0 offset = sample['offset'] if 'offset' in sample else 0 features = self.featurizer.process(sample['audio_filepath'], offset=offset, duration=duration, trim=self.trim) f, fl = features, torch.tensor(features.shape[0]).long() # f = f / (torch.max(torch.abs(f)) + 1e-5) else: f, fl = None, None t, tl = sample["transcript"], len(sample["transcript"]) if self.eos_id is not None: t = t + [self.eos_id] tl += 1 return \ f, fl, \ torch.tensor(t).long(), torch.tensor(tl).long() def __len__(self): return len(self.manifest)
5,769
1,755
""" ~~~ IMPORT EXPERIMENTAL DATA, PROCESS, AND NONDIMENSIONALIZE ~~~ This code reads in the rescaled Snodgrass data and compares parameters to known parameters found in the Henderson and Segur paper. 1. Get distances 2. Read in the gauge data for each event (get frequencies and Fourier magnitudes) 3. Adjust the y axis units 4. Get the k vector using integer division and clean up 5. Get the carrier wave location (requires some restricting) 6. Factor out carrier wave 7. Get the energies at each gauge 8. Get nondimensionalization constants """ import numpy as np import os import glob import matplotlib.pyplot as plt from numpy.fft import fft, ifft import NLS import random as rand from scipy import interpolate ### STEP 1: Get distance information distv = np.array([0.0,2400000.0,4200000.0,8700000.0]) # Distances between gauges in METERS ### STEP 2: Read in information at each gauge for each event subdirs = ['Aug1Data','Aug2Data','JulyData'] # Define something that will list directories that are not hidden def listdirNH(path): return glob.glob(os.path.join(path, '*')) # Read in the data j = 0 for sd in subdirs: files = listdirNH(sd+'/Rescaled') # Initialize some values n = 0 pi =0 fig1,ax1 = plt.subplots(4,1) plt.suptitle(sd) # Get files Deltavals = [] for f in files: datavals = np.transpose(np.loadtxt(f).view(float)) N = len(datavals[1]) x = datavals[0] # Frequencies sly = datavals[1] # Magnitudes #ly = np.sqrt(sly*x)*0.01 #MULTIPLY VERSION (get the amplitude in meters) mns = [] for w in range(N-1): mns.append(np.abs(x[w+1]-x[w])) #mns.append(np.mean(mns)) ### STEP 3: Adjust the y axis units ly = np.sqrt(sly*np.mean(mns))*0.01 # INTEGRATED VERSION ### STEP 4: Get the k vector using integer division and clean up L = 3600*3 # The period k = (x*0.001)//(2*np.pi/L) # Convert to mHz, then divide by 2pi/L to get the k vector # REMOVE DUPLICATE VALUES ndk = np.array(()) for fi in range(len(k)): num = k[fi] if num not in ndk: ndk = np.append(ndk,num) lll =[] for h in ndk: l1=np.where(k==h)[0] lll.append(l1) ndy = np.array(()) for ar in lll: val = np.mean(ly[ar]) ndy=np.append(ndy,val) ### STEP 5: Get the location of the carrier wave (defined by the first gauge) if n == 0: m = max(ndy) i = np.where(ndy == m) if len(i[0]) > 1: newi = i[0][len(i[0])//2] carriermode = np.array(newi) carrierloc = ndk[carriermode] else: newi = i[0][0] carriermode = np.array(newi) carrierloc = ndk[carriermode] # First, find the carrier mode in ANY file, not just the first one loc = np.where(np.logical_and(ndk>carrierloc*0.99, ndk<carrierloc*1.001)) #loc = np.where(np.logical_and(ndk>carrierloc-1, ndk<carrierloc+1)) # Be a little more restrictive if len(loc[0])>1: loc = loc[0][0] else: loc = loc[0][0] ### STEP 6: Redefine the k vector so that the carrier mode is at 0 (factor it out) knew = ndk-ndk[loc] xnew = x-x[loc] ### STEP 7: Get the "Energy" integrals fnc = interpolate.interp1d(x, sly,kind ='cubic') longx = np.linspace(x[0],x[-1],1000) newy = fnc(longx) A0 = np.sqrt(2*NLS.trapezoid(newy,(x[-1]-x[0])))*0.01 figg,axx = plt.subplots() axx.plot(x,sly,'.',markersize=7) axx.plot(longx,newy) plt.show() M000 = NLS.trapezoid(newy[np.where(np.logical_and(longx>41.2,longx<75.6))],(74.6-41.2)) Deltavals.append(M000) ### STEP 8: Get nondimensionalization constants g = 9.81 #(m/s^2) if n==0: w0 = (2*np.pi)**2/L*ndk[loc] # Get the value from the integer k0 = w0**2/g # The carrier wavenumber m = max(ndy) epsilon = 2*m*k0 # The nondimensionalization constant epsilon heps = A0*k0 print(f,'Special Values') print('2A0',A0) print('Maximum value',m) print('Carrier frequency',w0) print('Wavenumber',k0) print('MY epsilon',epsilon) print('HENDERSON EPSILON', heps) print('period',L) n = n+1 M0 = Deltavals[0] MX = Deltavals/M0 energyy = np.log(MX) # Get the fit and define a new y vector A = np.vstack([distv, np.ones(len(distv))]).T m, b = np.linalg.lstsq(A, energyy,rcond=-1)[0] # m is delta hdeltab = -m hdeltas = hdeltab/(2*heps**2*k0) xplot = np.linspace(distv[0],distv[-1],100) newy = m*xplot+b print('HENDERSON BIG Delta ',hdeltab, 'b ', b) print('HENDERSON LITTLE delta', hdeltas) print()
5,277
1,976
from .rip import *
19
7
# -*- coding: utf-8 -*- # Time : 2022/1/17 15:20 # Author : QIN2DIM # Github : https://github.com/QIN2DIM # Description: import os.path import time from hashlib import sha256 from typing import List, Optional, Union, Dict import cloudscraper import yaml from lxml import etree # skipcq: BAN-B410 - Ignore credible sources from selenium.common.exceptions import WebDriverException, InvalidCookieDomainException from selenium.webdriver.common.action_chains import ActionChains from selenium.webdriver.common.by import By from selenium.webdriver.common.keys import Keys from selenium.webdriver.support import expected_conditions as EC from selenium.webdriver.support.wait import WebDriverWait from services.settings import DIR_EXPLORER, EPIC_EMAIL from services.settings import logger from services.utils import ToolBox, ChallengerContext, StandardContext from .exceptions import DiscoveryTimeoutException, ProtocolOutdatedWarning class EpicAwesomeExplorer: """游戏商店探索者 获取免费游戏数据以及促销信息""" # 平台对象参数 URL_STORE_HOME = "https://store.epicgames.com/zh-CN/" URL_FREE_GAMES = "https://store.epicgames.com/zh-CN/free-games" URL_STORE_PREFIX = "https://store.epicgames.com/zh-CN/browse?" URL_STORE_FREE_GAME = ( f"{URL_STORE_PREFIX}sortBy=releaseDate&sortDir=DESC&priceTier=tierFree&count=40" ) URL_STORE_FREE_DLC = f"{URL_STORE_PREFIX}sortBy=releaseDate&sortDir=DESC&priceTier=tierFree&category=GameAddOn&count=40&start=0" # noqa URL_PROMOTIONS = ( "https://store-site-backend-static.ak.epicgames.com/freeGamesPromotions?locale=zh-CN" ) URL_PRODUCT_PAGE = "https://store.epicgames.com/zh-CN/p/" def __init__(self, silence: bool = None): self.silence = True if silence is None else silence # 驱动参数 self.action_name = "AwesomeFreeGirl" # 运行缓存 self.runtime_workspace = None self.path_free_games = "ctx_store.yaml" self.game_objs = {} # {index0:{name:value url:value}, } self.category_details = { "game": {"url": self.URL_STORE_FREE_GAME, "flag": "免费游戏"}, "dlc": {"url": self.URL_STORE_FREE_DLC, "flag": "免费附加内容"}, } # 初始化工作空间 self._init_workspace() def _init_workspace(self) -> None: """初始化工作目录 缓存游戏商店数据""" self.runtime_workspace = "." if not os.path.exists(DIR_EXPLORER) else DIR_EXPLORER self.path_free_games = os.path.join(self.runtime_workspace, self.path_free_games) def _discovery_free_games( self, ctx: Union[ChallengerContext, StandardContext], ctx_cookies: List[dict], category: str = "game", ) -> None: """发现玩家所属地区可视的常驻免费游戏数据""" url = self.category_details[category]["url"] flag = self.category_details[category]["flag"] # 重载玩家令牌 if ctx_cookies: ctx.get(self.URL_STORE_FREE_GAME) for cookie_dict in ctx_cookies: try: ctx.add_cookie(cookie_dict) except InvalidCookieDomainException: pass _mode = "(深度搜索)" if ctx_cookies else "(广度搜索)" logger.debug( ToolBox.runtime_report( motive="DISCOVERY", action_name=self.action_name, message=f"📡 正在为玩家搜集{flag}{_mode}...", ) ) # 获取免费游戏链接 _start = time.time() _url_store_free = url while True: ctx.get(_url_store_free) time.sleep(1) WebDriverWait(ctx, 10, ignored_exceptions=WebDriverException).until( EC.presence_of_element_located( (By.XPATH, "//section[@data-testid='section-wrapper']") ) ) # 滑到底部 action = ActionChains(ctx) action.send_keys(Keys.END) action.perform() # 判断异常跳转 if "tierFree" not in ctx.current_url: break if time.time() - _start > 80: raise DiscoveryTimeoutException(f"获取{flag}链接超时") # 断言最后一页 WebDriverWait(ctx, 5, ignored_exceptions=WebDriverException).until( EC.element_to_be_clickable((By.XPATH, "//a[@data-component='PaginationItem']")) ) page_switcher = ctx.find_elements(By.XPATH, "//a[@data-component='PaginationItem']")[-1] # 提取价值信息 game_objs = ctx.find_elements(By.XPATH, "//a[@class='css-1jx3eyg']") for game_obj in game_objs: name = game_obj.get_attribute("aria-label") url = game_obj.get_attribute("href") self.game_objs.update( { self.game_objs.__len__(): { "name": name.split(",")[0].replace("\n", "").strip(), "url": url.strip(), "in_library": None, } } ) # 页面跳转判断 page_end = page_switcher.get_attribute("href") if page_end in ctx.current_url: break # 更新跳转链接 _url_store_free = page_end logger.success( ToolBox.runtime_report( motive="DISCOVERY", action_name=self.action_name, message=f"{flag}搜集完毕", qsize=len(self.game_objs), ) ) def stress_expressions(self, ctx: Union[ChallengerContext, StandardContext]) -> Dict[str, str]: """应力表达式的主要实现""" logger.debug( ToolBox.runtime_report( motive="DISCOVERY", action_name=self.action_name, message="📡 使用应力表达式搜索周免游戏..." ) ) # 访问链接 游戏名称 pending_games = {} for i in range(2): try: ctx.get(self.URL_STORE_HOME) time.sleep(3) # 定位周免游戏的绝对位置 WebDriverWait(ctx, 45, ignored_exceptions=WebDriverException).until( EC.presence_of_element_located((By.XPATH, "//a[contains(string(),'当前免费')]")) ) # 周免游戏基本信息 stress_operator = ctx.find_elements(By.XPATH, "//a[contains(string(),'当前免费')]") title_seq = ctx.find_elements( By.XPATH, "//a[contains(string(),'当前免费')]//span[@data-testid='offer-title-info-title']", ) # 重组周免游戏信息 for index, _ in enumerate(stress_operator): href = stress_operator[index].get_attribute("href") try: pending_games[href] = f"{title_seq[index].text}".strip() except AttributeError as err: if i == 0: raise AttributeError from err pending_games[href] = "null" break except (WebDriverException, AttributeError): continue return pending_games class GameLibManager(EpicAwesomeExplorer): """游戏对象管理 缓存商城数据以及判断游戏在库状态""" def __init__(self): super().__init__() self.action_name = "GameLibManager" self.email = EPIC_EMAIL self.auth_str = "explorer" def _z(self) -> str: return ( sha256(f"{self.email[-3::-2]}{self.auth_str}".encode("utf-8")).hexdigest() if self.email else "" ) def _check_protocol(self): """ 读取协议文件,检查协议头 :except ProtocolOutdatedWarning: 缓存文件异常,请返回空数据 :return: stream """ try: with open(self.path_free_games, "r", encoding="utf8") as file: ctx_protocol = yaml.safe_load(file) except FileNotFoundError as err: raise ProtocolOutdatedWarning from err else: if not ctx_protocol or not isinstance(ctx_protocol, dict): raise ProtocolOutdatedWarning return ctx_protocol @staticmethod def _update_status(game_objs, ctx_content=None, runtime: bool = False): """ 更新实体信息 - case1 文件不存在 空文件 过时 协议。 跳过读取,初始化协议对象,创建并写入内容。 - case2 缓存协议镜像,扩建实体,并逐条比对更新实体信息 加入新增内容,更新 ``in_library`` 状态 :param game_objs: 详见 `Game Object` :param ctx_content: 实际上就是上一次保存的 game_objs :param runtime: 默认False,是否为运行时更新实例状态。 一个显然的认知是,``IF runtime is False`` 为全量更新,应进行全库实例比对 `` IF runtime is True `` 为局部更新,此时 new_objs 数量一般会远少于 memory_objs 的数量, 应在比对后将 new_objs 的缺省实例补回。 :type game_objs: List[Dict[str, str|bool]] :type ctx_content: List[Dict[str, str|bool]] :return: """ if not ctx_content: return game_objs runtime = bool(runtime) new_objs_map = {game["url"]: game for game in game_objs} memory_objs_map = {c["url"]: c for c in ctx_content} new_content_objs = [] # 对新内容迭代会自动丢弃过期的周免实例 for flag in new_objs_map.keys(): # 添加上一轮未发现的新游戏 if not memory_objs_map.get(flag): new_content_objs.append(new_objs_map[flag]) # 逐条对比 更新实体在库状态 else: memory = memory_objs_map[flag] new_obj = new_objs_map[flag] if new_obj.get("in_library") is None: new_obj["in_library"] = memory.get("in_library") # 添加更新在库状态后的实例 new_content_objs.append(new_obj) # 将 new_objs 的缺省实例补回 if runtime is True: for flag in memory_objs_map.keys(): if not new_objs_map.get(flag): new_content_objs.append(memory_objs_map[flag]) return new_content_objs def save_game_objs(self, game_objs, category: str, runtime: Optional[bool] = None): """ 缓存免费商城数据 :param runtime: :param game_objs: :param category: :type game_objs: List[Dict[str, str|bool]] :return: """ if not game_objs: return try: ctx_protocol = self._check_protocol() except ProtocolOutdatedWarning: content = {self._z(): {category: game_objs}} with open(self.path_free_games, "w", encoding="utf8") as file: yaml.dump(content, file, allow_unicode=True) logger.success( ToolBox.runtime_report( motive="SAVE", action_name=self.action_name, message="Cache Epic store information.", ) ) else: # 切换账号 键值补全 if not ctx_protocol.get(self._z()): ctx_protocol[self._z()] = {} # 键值比对更新 merged_content = self._update_status( game_objs=game_objs, ctx_content=ctx_protocol[self._z()].get(category), runtime=runtime, ) ctx_protocol[self._z()].update({category: merged_content}) # 缓存更新 with open(self.path_free_games, "w", encoding="utf8") as file: yaml.dump(ctx_protocol, file, allow_unicode=True) logger.success( ToolBox.runtime_report( motive="MERGE", action_name=self.action_name, message="Update Epic store information.", ) ) def load_game_objs(self, category: str, only_url: bool = None): """ 加载缓存在本地的免费游戏对象 :param category: :param only_url: :return: """ try: ctx_protocol = self._check_protocol() except ProtocolOutdatedWarning: return [] else: ctx_content = ctx_protocol.get(self._z(), {}).get(category, []) if not ctx_content: return [] if only_url is True: return [obj["url"] for obj in ctx_content] return ctx_content @staticmethod def is_my_game(ctx_cookies, page_link: str, pre_assert_content: bytes = None) -> Optional[dict]: """ 判断游戏在库状态 :param pre_assert_content: 前置协同响应流,将耗时的网络请求操作前置, 封装成协程任务,而仅将此函数用于解析上游模块的静态返回值。 :param ctx_cookies: :param page_link: :type ctx_cookies: List[dict]|str :return: None 异常状态 True 跳过任务 False 继续任务 仅当返回值为 False 时可以继续任务,并可以进一步筛选掉 AjaxLoadingReject 目标。 """ def response_wrapper(new_params: dict): resp_ = {"assert": "", "status": None, "warning": ""} resp_.update(new_params) return resp_ # 模式匹配 --> 上下文呈递|重新握手 if pre_assert_content is None: headers = { "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) " "Chrome/100.0.4896.75 Safari/537.36 Edg/100.0.1185.36", "cookie": ctx_cookies if isinstance(ctx_cookies, str) else ToolBox.transfer_cookies(ctx_cookies), } scraper = cloudscraper.create_scraper() response = scraper.get(page_link, headers=headers) content = response.content else: content = pre_assert_content # 清洗页面的促销信息 tree = etree.HTML(content) assert_obj = tree.xpath( "//span[@data-component='PurchaseCTA']//span[@data-component='Message']" ) # 模式匹配 --> 向下游呈递资源对象的状态 # 1. 剔除已在库的、付费的、未推出的资源; # 2. 剔除被限制的免费资源; # 3. 呈递待领取的免费资源; # 🚧 异常状态 忽略尚未发布的游戏对象 if not assert_obj: return response_wrapper({"assert": "AssertObjectNotFound", "status": None}) assert_message = assert_obj[0].text # 🚧 跳过 `无法认领` 的日志信息 if assert_message in ["已在游戏库中", "已在库中", "立即购买", "购买", "即将推出"]: return response_wrapper({"assert": assert_message, "status": True}) # 🚧 惰性加载,前置节点不处理动态加载元素 if assert_message in ["正在载入"]: return response_wrapper({"assert": "AjaxLoadingReject", "status": False}) # 🍟 未领取的免费游戏 if assert_message in ["获取"]: # 無遮挡警告 繼續任務 warning_obj = tree.xpath("//h1[@class='css-1gty6cv']//span") if not warning_obj: return response_wrapper({"assert": assert_message, "status": False}) # 成人内容可获取 warning_message = warning_obj[0].text if warning_message in ["成人内容"]: return response_wrapper( {"assert": assert_message, "warning": warning_message, "status": False} ) # 地区限制無法獲取 return response_wrapper( {"assert": assert_message, "warning": warning_message, "status": None} ) class _Game: """基础游戏对象""" # 游戏名称 name: str = "《堡垒之夜》" # 商城访问链接 url: str = "https://store.epicgames.com/zh-CN/p/fortnite" # 在库情况 True在库 False不在 None不到啊(初始化状态) in_library: bool = None class _Dlc: """游戏附加内容对象""" # 附加内容名称 name: str = "《消逝的光芒》-《求生之路 2》Weapon Pack" # 商城访问链接 url: str = "https://store.epicgames.com/zh-CN/p/dying-light--left-4-dead-2-weapon-pack" # 在库情况 True在库 False不在 None不到啊(初始化状态) in_library: bool = None class _Report: """消息推送|资源封装对象""" # 资源名称 name: str = "" # 资源链接 url: str = "" # 执行结果 status: str = ""
15,542
5,893
import torch import torch.nn as nn from torch.nn import functional as F class LinearNorm(nn.Module): def __init__(self, in_features, out_features, activation=None, use_bias=True, kernel_initializer='glorot_uniform', bias_initializer='zeros'): super(LinearNorm, self).__init__() self.linear = nn.Linear( in_features=in_features, out_features=out_features, bias=use_bias) # init weight if kernel_initializer == 'glorot_uniform': nn.init.xavier_uniform_(self.linear.weight) elif kernel_initializer == 'zeros': nn.init.zeros_(self.linear.weight) # init bias if use_bias: if bias_initializer == 'zeros': nn.init.constant_(self.linear.bias, 0.0) else: raise NotImplementedError self.activation = activation if activation is not None else nn.Identity() def forward(self, x): x = self.activation(self.linear(x)) return x class ConvNorm(nn.Module): def __init__( self, in_channels, out_channels, kernel_size=1, stride=1, padding=None, dilation=1, activation=None, use_bias=True, kernel_initializer='glorot_uniform', bias_initializer='zeros'): super(ConvNorm, self).__init__() self.conv = nn.Conv1d(in_channels, out_channels, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation, bias=use_bias ) # init weight if kernel_initializer == 'glorot_uniform': nn.init.xavier_uniform_(self.conv.weight) elif kernel_initializer == 'zeros': nn.init.zeros_(self.conv.weight) # init bias if use_bias: if bias_initializer == 'zeros': nn.init.constant_(self.conv.bias, 0.0) def forward(self, signal): conv_signal = self.conv(signal) return conv_signal class PreNet(nn.Module): def __init__(self, in_features, units, drop_rate, activation): super(PreNet, self).__init__() self.dense1 = LinearNorm( in_features, units, activation=activation) self.dense2 = LinearNorm( units, units, activation=activation) self.dropout_layer = nn.Dropout(p=drop_rate) def forward(self, inputs): dense1_out = self.dense1(inputs) dense1_out = self.dropout_layer(dense1_out) dense2_out = self.dense2(dense1_out) dense2_out = self.dropout_layer(dense2_out) return dense2_out class ConvPreNet(nn.Module): def __init__(self, nconv, hidden, conv_kernel, drop_rate, activation=nn.ReLU(), bn_before_act=True): super(ConvPreNet, self).__init__() self.conv_stack = nn.ModuleList( [ Conv1D(in_channels=hidden, out_channels=hidden, kernel_size=conv_kernel, activation=activation, drop_rate=drop_rate, bn_before_act=bn_before_act) for i in range(nconv) ] ) self.projection = LinearNorm(hidden, hidden) def forward(self, inputs, mask=None): conv_outs = inputs for conv in self.conv_stack: conv_outs = conv(conv_outs, mask) projections = self.projection(conv_outs) return projections class FFN(nn.Module): def __init__(self, in_features, hidden1, hidden2): super(FFN, self).__init__() self.dense1 = LinearNorm(in_features, hidden1, activation=nn.ReLU()) self.dense2 = LinearNorm(hidden1, hidden2, activation=None) self.layer_norm = nn.LayerNorm(hidden2) def forward(self, inputs, mask=None): dense1_outs = self.dense1(inputs) dense2_outs = self.dense2(dense1_outs) outs = dense2_outs + inputs outs = self.layer_norm(outs) if mask is not None: outs = outs.masked_fill(mask.unsqueeze(-1), 0.0) return outs class Conv1D(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, activation, drop_rate, bn_before_act=False, strides=1): super(Conv1D, self).__init__() self.conv1d = ConvNorm(in_channels=in_channels, out_channels=out_channels, kernel_size=kernel_size, stride=strides, padding=int((kernel_size - 1) / 2), dilation=1, activation=None) self.activation = activation if activation is not None else nn.Identity() self.bn = nn.BatchNorm1d(out_channels) self.dropout = nn.Dropout(p=drop_rate) self.bn_before_act = bn_before_act def forward(self, inputs, mask=None): conv_outs = inputs.contiguous().transpose(1, 2) conv_outs = self.conv1d(conv_outs) if self.bn_before_act: conv_outs = self.bn(conv_outs) conv_outs = self.activation(conv_outs) else: conv_outs = self.activation(conv_outs) conv_outs = self.bn(conv_outs) dropouts = self.dropout(conv_outs) dropouts = dropouts.contiguous().transpose(1, 2) if mask is not None: dropouts = dropouts.masked_fill(mask.unsqueeze(-1), 0.0) return dropouts class PostNet(nn.Module): def __init__(self, n_conv, hidden, conv_filters, conv_kernel, drop_rate): super(PostNet, self).__init__() activations = [nn.Tanh()] * (n_conv - 1) + [nn.Identity()] self.conv_stack = nn.ModuleList( [ Conv1D(in_channels=hidden if i == 0 else conv_filters, out_channels=conv_filters, kernel_size=conv_kernel, activation=activations[i], drop_rate=drop_rate) for i in range(n_conv) ] ) def forward(self, inputs, mask=None): conv_out = inputs for conv in self.conv_stack: conv_out = conv(conv_out, mask) return conv_out class PositionalEncoding(nn.Module): def __init__(self): super(PositionalEncoding, self).__init__() @staticmethod def positional_encoding(len, dim, device, step=1.): """ :param len: int scalar :param dim: int scalar :param device: :param step: :return: position embedding """ pos_mat = torch.tile( (torch.arange(0, len, dtype=torch.float32, device=device) * step).unsqueeze(-1), [1, dim]) dim_mat = torch.tile( torch.arange(0, dim, dtype=torch.float32, device=device).unsqueeze(0), [len, 1]) dim_mat_int = dim_mat.type(torch.int32) pos_encoding = torch.where( # [time, dims] torch.eq(torch.fmod(dim_mat_int, 2), 0), torch.sin(pos_mat / torch.pow(10000., dim_mat / float(dim))), torch.cos(pos_mat / torch.pow(10000., (dim_mat - 1) / float(dim)))) return pos_encoding
7,312
2,408
# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Interface definitions for federated algorithms.""" import abc from typing import Generic, List, TypeVar from fedjax.legacy.core.model import Model from fedjax.legacy.core.typing import FederatedData T = TypeVar('T') class FederatedAlgorithm(Generic[T], metaclass=abc.ABCMeta): """Interface for federated algorithms. This interface structures federated algorithms at the per round level. Algorithm state contains any round specific parameters (e.g. model parameters) that will be passed from round to round. This state is initialized by `init_state` and passed as input into and returned as output from `run_round`. We suggest defining state using `typing.NamedTuple` as this provides immutability, type hinting, and works by default with JAX transformations. Anything else that is static and doesn't update at each round can be defined in the `__init__`. The expected usage of FederatedAlgorithm is as follows: ``` algorithm = FederatedAlgorithm() state = algorithm.init_state() for i in range(num_rounds): client_ids = sample_clients(i) state = algorithm.run_round(state, client_ids) ``` """ @abc.abstractproperty def federated_data(self) -> FederatedData: """Returns federated data used in federated training.""" @abc.abstractproperty def model(self) -> Model: """Returns model used in federated training.""" @abc.abstractmethod def init_state(self) -> T: """Returns initial state of algorithm.""" @abc.abstractmethod def run_round(self, state: T, client_ids: List[str]) -> T: """Runs one round of federated training."""
2,186
630
from enum import Enum from typing import ( Any, Dict, ) from galaxy import ( exceptions, model, ) from galaxy.managers import hdas from galaxy.managers.context import ProvidesUserContext from galaxy.managers.jobs import ( JobManager, JobSearch, view_show_job, ) from galaxy.schema.fields import EncodedDatabaseIdField from galaxy.schema.schema import JobIndexQueryPayload class JobIndexViewEnum(str, Enum): collection = "collection" admin_job_list = "admin_job_list" class JobIndexPayload(JobIndexQueryPayload): view: JobIndexViewEnum = JobIndexViewEnum.collection class JobsService: job_manager: JobManager job_search: JobSearch hda_manager: hdas.HDAManager def __init__( self, job_manager: JobManager, job_search: JobSearch, hda_manager: hdas.HDAManager, ): self.job_manager = job_manager self.job_search = job_search self.hda_manager = hda_manager def show( self, trans: ProvidesUserContext, id: EncodedDatabaseIdField, full: bool = False, ) -> Dict[str, Any]: id = trans.app.security.decode_id(id) job = self.job_manager.get_accessible_job(trans, id) return view_show_job(trans, job, bool(full)) def index( self, trans: ProvidesUserContext, payload: JobIndexPayload, ): security = trans.security is_admin = trans.user_is_admin if payload.view == JobIndexViewEnum.admin_job_list: payload.user_details = True user_details = payload.user_details if payload.view == JobIndexViewEnum.admin_job_list and not is_admin: raise exceptions.AdminRequiredException("Only admins can use the admin_job_list view") query = self.job_manager.index_query(trans, payload) out = [] view = payload.view for job in query.yield_per(model.YIELD_PER_ROWS): job_dict = job.to_dict(view, system_details=is_admin) j = security.encode_all_ids(job_dict, True) if view == JobIndexViewEnum.admin_job_list: j["decoded_job_id"] = job.id if user_details: j["user_email"] = job.get_user_email() out.append(j) return out
2,306
705
import pytest import numpy as np from deephub.models.registry.toy import DebugToyModel from deephub.models.feeders import MemorySamplesFeeder from deephub.trainer import Trainer @pytest.mark.slow def test_early_stopping(tmpdir): model_params = { 'type': 'toy:DebugToyModel', 'model_dir': str(tmpdir / 'test_early_stopping/'), 'learning_rate': 0.01, 'num_classes': 2, 'num_steps_per_epoch': 1, 'hidden_neurons': 512 } # Initialize the model model = DebugToyModel(**model_params) # Read training data train_feeder = MemorySamplesFeeder(np.asarray(np.arange(0, 10).reshape(10, 1), dtype=np.float32), np.array([int(0 if i < 5 else 1) for i in range(10)], dtype=np.int64, ndmin=1), batch_size=10, feed_as_dict=False) # Read validation data validation_feeder = MemorySamplesFeeder(np.asarray(np.arange(10, 20).reshape(10, 1), dtype=np.float32), np.array([int(0 if i < 5 else 1) for i in range(10)], dtype=np.int64, ndmin=1), batch_size=10, feed_as_dict=False) # Initialize the Trainer trainer = Trainer() train_params = { 'epochs': 12, 'save_summary_steps': 1, 'save_checkpoint_steps': 1, 'early_stopping_metric': 'loss', 'early_stopping_steps_without_decrease': 3, 'early_stopping_min_steps': 1, 'early_stopping_hook_run_every_steps': 1 } # Start training process trainer.train(model=model, train_feeder=train_feeder, eval_feeder=validation_feeder, **train_params) # Grab global step from model.estimator object global_step = model.estimator().get_variable_value('global_step') assert global_step < train_params['epochs']
1,984
668
from typing import TYPE_CHECKING import pytest from . import CapturedOutput from utsc.switchconfig import config from prompt_toolkit.application import create_app_session from prompt_toolkit.input import create_pipe_input if TYPE_CHECKING: from .. import MockedUtil from pytest_mock import MockerFixture @pytest.fixture() def mock_config(mocker: "MockerFixture", mock_util: "MockedUtil"): mocker.patch.object(config, "util", mock_util) return config @pytest.fixture(scope="function") def mock_pt_app(): pipe_input = create_pipe_input() pipe_output = CapturedOutput() try: with create_app_session(input=pipe_input, output=pipe_output) as app: yield app finally: pipe_input.close() # endregion # region failed interactive fixture experiment # def pytest_addoption(parser): # parser.addoption( # "--interactive", action="store_true", default=False, help="run interactive tests" # ) # @pytest.fixture() # def interactive(request, capfd: 'CaptureFixture'): # if request.config.getoption("--interactive") or os.getenv("VSCODE_DEBUGGER"): # # here we reach directly into capsys._capture, # # because the capsys.disabled context manager # # does not suspend capturing of stdin. # capmanager: 'CaptureManager' = capfd.request.config.pluginmanager.getplugin("capturemanager") # capmanager.suspend(in_=True) # assert capfd._capture # noqa # capfd._capture.suspend_capturing(in_=True) # noqa # yield # capmanager.resume() # capfd._capture.resume_capturing() # noqa # else: # pytest.skip("This test can only be run with the --interactive option") # def pytest_collection_modifyitems(config, items): # if config.getoption("--interactive"): # # --interactive given in cli: do not skip interactive tests # return # skip_interactive = pytest.mark.skip(reason="need --interactive option to run") # for item in items: # if "interactive" in item.keywords and not os.getenv("VSCODE_DEBUGGER"): # item.add_marker(skip_interactive) # endregion
2,157
664
# Copyright (c) 2014, HashFast Technologies LLC # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # 1. Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # 3. Neither the name of HashFast Technologies LLC nor the # names of its contributors may be used to endorse or promote products # derived from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL HASHFAST TECHNOLOGIES LLC BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. from abc import ABCMeta, abstractmethod from ..load import crc from ..util import with_metaclass, int_to_lebytes, lebytes_to_int # Operation codes from hf_protocol.h. opcodes = { # Serial protocol operation codes (Second header byte) 'OP_NULL': 0, 'OP_ROOT': 1, 'OP_RESET': 2, 'OP_PLL_CONFIG': 3, 'OP_ADDRESS': 4, 'OP_READDRESS': 5, 'OP_HIGHEST': 6, 'OP_BAUD': 7, 'OP_UNROOT': 8, 'OP_HASH': 9, 'OP_NONCE': 10, 'OP_ABORT': 11, 'OP_STATUS': 12, 'OP_GPIO': 13, 'OP_CONFIG': 14, 'OP_STATISTICS': 15, 'OP_GROUP': 16, 'OP_CLOCKGATE': 17, # Factory Codes 'OP_SERIAL': 50, # Serial number read/write 'OP_LIMITS': 51, # Operational limits read/write 'OP_HISTORY': 52, # Read operational history data 'OP_CHARACTERIZE': 53, # Characterize one or more die 'OP_CHAR_RESULT': 54, # Characterization result 'OP_SETTINGS': 55, # Read or write settings 'OP_FAN_SETTINGS': 56, 'OP_POWER': 57, 'OP_BAD_CORE': 58, # Set or clear bad core status # USB interface specific operation codes 'OP_USB_INIT': 128, # Initialize USB interface details 'OP_GET_TRACE': 129, # Send back the trace buffer if present 'OP_LOOPBACK_USB': 130, 'OP_LOOPBACK_UART': 131, 'OP_DFU': 132, # Jump into the boot loader 'OP_USB_SHUTDOWN': 133, # Initialize USB interface details 'OP_DIE_STATUS': 134, # Die status. There are 4 die per ASIC 'OP_GWQ_STATUS': 135, # Global Work Queue protocol status 'OP_WORK_RESTART': 136, # Stratum work restart regime 'OP_USB_STATS1': 137, # Statistics class 1 'OP_USB_GWQSTATS': 138, # GWQ protocol statistics 'OP_USB_NOTICE': 139, # Asynchronous notification event 'OP_PING': 140, # Echo 'OP_CORE_MAP': 141, # Return core map 'OP_VERSION': 142, # Version information 'OP_FAN': 143, # Set Fan Speed 'OP_NAME': 144, # System name write/read 'OP_USB_DEBUG': 255 } opnames = {} for opcode_name, opcode in opcodes.items(): assert opcode not in opnames opnames[opcode] = opcode_name known_opcodes = set(opcodes.keys()) known_opnames = set(opnames.keys()) def check_framebytes(framebytes): assert {x >= 0 and x < 256 for x in framebytes} == set([True]) assert len(framebytes) >= 8 assert framebytes[0] == 0xaa assert framebytes[7] == crc.crc8(framebytes[1:7]) if framebytes[6] == 0: assert len(framebytes) == 8 else: data_length = 4 * framebytes[6] # Eight byte frame header, data, plus 4 crc32 bytes. # Fix: Restore when using serial line directly # expected_framebytes_length = 8 + data_length + 4 expected_framebytes_length = 8 + data_length assert expected_framebytes_length == len(framebytes) data = framebytes[8:8+data_length] # Fix: Restore when using serial line directly # crc32 = framebytes[-4:] # if crc32 != crc.crc32_to_bytelist(crc.crc32(data)): # raise HF_Error("Bad CRC32 checksum.") class hf_frame_data(with_metaclass(ABCMeta, object)): def __init__(self, bytes=None): self.initialize() if bytes is None: self.generate_frame_data() else: self.parse_frame_data(bytes) @abstractmethod def initialize(self): pass @abstractmethod def parse_frame_data(self, bytes): pass @abstractmethod def generate_frame_data(self): pass class hf_frame_data_base(hf_frame_data): LENGTH = 0 def intialize(self): pass def parse_frame_data(self, bytes): assert len(bytes) >= self.LENGTH def generate_frame_data(self): self.frame_data = [0x00] return self.frame_data # Fix: Document terminology: frame is the whole thing and consists of up to # three parts: the header, the data, and the CRC32 checksum. # Fix: Wants to verify checksums and throw exception if they are not right. # And check for 0xaa. # Fix: Wants to make all the fields of the header accessible, but also provide raw bytes. # Fix: Should be able to initialize with stream of bytes or by filling in fields # and asking for the bytes. Throw exception if field values are out of bounds. # Fix: Maybe want something which checks for known opcode and whether fields are # plausible for that opcode -- problem is that if we are using this to report # what was seen on the wire, we need to make those assumptions, maybe. # Fix: The really pure way to do this is to create a subclass for every opcode type # and then have specific methods for that type. Probably more trouble than # its worth, but it would also let us have specific methods for parameters # that just occupy a couple bits. class HF_Frame(): def __init__(self, initial_state): self.initialize() if initial_state is None: pass elif isinstance(initial_state, list): self.off_the_wire(initial_state) elif isinstance(initial_state, dict): self.buildframe(initial_state) else: raise HF_Error("Argument type not supported: %s" % (inital_state)) def initialize(self): self.framebytes = [] self.operation_code = None self.chip_address = 0 self.core_address = 0 self.hdata = 0 self.data_length_field = 0 self.crc8 = 0 self.data = None # Fix: Restore when using serial line directly # self.crc32 = None self.data_length = 0; def off_the_wire(self, framebytes): check_framebytes(framebytes) self.framebytes = framebytes self.operation_code = framebytes[1] self.chip_address = framebytes[2] self.core_address = framebytes[3] self.hdata = lebytes_to_int(framebytes[4:6]) self.data_length_field = framebytes[6] self.data_length = 4 * self.data_length_field self.crc8 = framebytes[7] if self.data_length > 0: assert {x >= 0 and x < 256 for x in framebytes} == set([True]) self.data = framebytes[8:8+self.data_length] # Fix: Restore when using serial line directly # self.crc32 = framebytes[8+self.data_length:] def set_data(self, data): self.data = data self.data_length = len(data) self.data_length_field = int(self.data_length / 4) # Fix: Restore when using serial line directly # self.crc32 = crc.crc32(self.data) def construct_framebytes(self): crc8_input = [self.operation_code] crc8_input += [self.chip_address] crc8_input += [self.core_address] crc8_input += int_to_lebytes(self.hdata, 2) crc8_input += [self.data_length_field] self.crc8 = crc.crc8(crc8_input) frameheader = [0xaa, self.operation_code, self.chip_address, self.core_address] + \ int_to_lebytes(self.hdata, 2) + [self.data_length_field, self.crc8] if self.data_length > 0: # Fix: Restore when using serial line directly # return frameheader + self.data + crc.crc32_to_bytelist(self.crc32) self.framebytes = frameheader + self.data else: self.framebytes = frameheader return self.framebytes def buildframe(self, framedict): legal_fields = set(['operation_code', 'chip_address', 'core_address', 'hdata', 'data']) received_fields = set(framedict.keys()) assert received_fields.issubset(legal_fields) assert 'operation_code' in framedict assert framedict['operation_code'] in opnames self.operation_code = framedict['operation_code'] if 'chip_address' in framedict: if framedict['chip_address'] < 0 or framedict['chip_address'] > 255: raise HF_Error("chip_address is out of range: %d" % (framedict['chip_address'])) self.chip_address = framedict['chip_address'] if 'core_address' in framedict: if framedict['core_address'] < 0 or framedict['core_address'] > 255: raise HF_Error("core_address is out of range: %d" % (framedict['core_address'])) self.core_address = framedict['core_address'] if 'hdata' in framedict: if framedict['hdata'] < 0 or framedict['hdata'] > 65535: raise HF_Error("hdata is out of range: %d" % (framedict['hdata'])) self.hdata = framedict['hdata'] if 'data' in framedict: assert len(framedict['data']) == 0 or {x >= 0 and x < 256 for x in framedict['data']} == set([True]) assert len(framedict['data']) <= 1020 and len(framedict['data']) % 4 == 0 if len(framedict['data']) > 0: self.set_data(framedict['data']) return self.construct_framebytes() def __str__(self): string = "" #string += "framebytes: {}\n".format(self.framebytes) string += "operation_code: {:#x}\n".format(self.operation_code) string += "chip_address: {:#x}\n".format(self.chip_address) string += "core_address: {:#x}\n".format(self.core_address) string += "hdata: {:#x}\n".format(self.hdata) string += "data_length_field: {}\n".format(self.data_length) #string += "data: {}\n".format(self.data) return string
10,742
3,712
from django.db import models from django.conf import settings from courses.models import Course # Create your models here. class Update(models.Model): user = models.ForeignKey( settings.AUTH_USER_MODEL, on_delete=models.CASCADE, null=True ) course = models.ForeignKey(Course, on_delete=models.CASCADE) update_title = models.CharField(max_length=200) update_description = models.TextField() #course_icon = models.CharField(max_length=200) update_created_datetime = models.DateTimeField(auto_now_add=True)
561
176
import argparse import os import re import sys from operator import itemgetter from typing import Optional import sentry_sdk import youtube_dl from selenium.common.exceptions import SessionNotCreatedException from cmd_tool import ( get_execution_path, exit_enter, get_input_path_or_exit, get_chrome_driver_path_or_exit, get_resource_path, cd ) from imagetools import Size from qrcode import NaverQrCode, make_qr_image, make_redirect_html from thumbnail import composite_thumbnail, capture_video from youtube_uploader import YoutubeUploader, YoutubeUploaderException sentry_sdk.init("https://1ff694f9169a4fa383a867fe10ed9329@o342398.ingest.sentry.io/5243685") def make_dirs(): os.chdir(get_execution_path()) print(os.path.abspath(os.curdir)) print('폴더 생성기') print('존재하는 폴더는 건너뜀') path = input('생성할 경로: ') if not os.path.isdir(path): print('없는 경로입니다') exit_enter(1) s, e = map(int, (input('시작할 숫자: '), input('끝나는 숫자: '))) for i in range(s, e + 1): os.makedirs(os.path.join(path, f'{i:04}'), exist_ok=True) print('완료') exit_enter() def make_thumbnail(): input_path = os.path.join(get_execution_path(), 'input') for cur_dir, _, files in os.walk(input_path): dir_name = os.path.basename(cur_dir) def _is_mp4(filename): _, ext = os.path.splitext(filename) return ext == '.mp4' mp4_files = tuple(filter(_is_mp4, files)) if 1 < len(mp4_files): print(f'pass: "{dir_name}" 안에 한개 이상의 mp4 파일이 존재합니다') if not dir_name.isnumeric(): print(f'skip: "{dir_name}" 는 숫자로 구성된 폴더이름이 아닙니다') continue idx_text = f'{int(dir_name):04}' mp4_filename = mp4_files[0] jpg_filename = f'{idx_text}.jpg' jpg_filepath = os.path.join(cur_dir, jpg_filename) print(f'capture:\t{mp4_filename} to {jpg_filename}') capture_video(os.path.join(cur_dir, mp4_filename), jpg_filepath) target_filename = f'p{idx_text}.jpg' print(f'composite:\t{jpg_filename} to {target_filename} ...') composite_thumbnail(jpg_filepath, os.path.join(cur_dir, target_filename)) print('완료되었습니다!') exit_enter() def upload_videos(): path = get_execution_path() chrome_driver_path = get_chrome_driver_path_or_exit() input_path = get_input_path_or_exit() cookie_path = os.path.join(get_execution_path(), 'cookies.txt') if not os.path.isfile(cookie_path): print('최상위 폴더에 cookies.txt 를 작성해야 합니다') exit_enter(1) uploader = None video_dirs = {} to_upload = {} for cur_dir, _, files in os.walk(input_path): dir_name = os.path.basename(cur_dir) video_path = video_name = thumbnail_path = None if 1 < len(tuple(filter(lambda s: s.endswith('.mp4'), files))): print(f'"{cur_dir}" 에 여러개의 .mp4 파일이 존재합니다!') continue for filename in files: if filename == 'youtube_url.txt': video_path = thumbnail_path = None print(f'already uploaded: {dir_name}') break current_video_name, ext = os.path.splitext(filename) if ext == '.mp4': if not dir_name.isnumeric(): print(f'skip: "{dir_name}" 는 숫자로 구성된 폴더이름이 아닙니다') break video_name = f'{int(dir_name):04}' video_path = os.path.join(cur_dir, f'{video_name}.mp4') if current_video_name != video_name: print(f'rename "{filename}" to "{video_name}.mp4"') os.rename(os.path.join(cur_dir, filename), video_path) video_dirs[video_name] = cur_dir elif ext == '.jpg' and re.match(r'^\d+[.]jpg$', filename): thumbnail_path = os.path.join(cur_dir, filename) if not (video_path and thumbnail_path): continue to_upload[int(dir_name)] = (video_name, video_path, thumbnail_path) for dir_number, (video_name, video_path, thumbnail_path) in sorted(to_upload.items(), key=lambda e: e[0]): # if not uploader: uploader = YoutubeUploader() try: my_channel_id = uploader.init(chrome_driver_path, cookie_path) except SessionNotCreatedException as e: print(e) print('컴퓨터에 설치된 chrome 과 chromedriver 의 버전이 일치하지 않습니다.') print('https://chromedriver.chromium.org/downloads 에서 다시 chromedriver 를 다운로드 해주세요.') break with open(os.path.join(path, '.mychannelid'), 'w') as f: f.write(my_channel_id) print(f'uploading {video_name}') try: if uploader.upload_video(video_path, thumbnail_path): print(f'success: {video_name}') else: print(f'failure: {video_name}') except YoutubeUploaderException as e: print(e) print(f'failure: {video_name}') try: uploader.browser.close() except: pass print('모든 업로드 작업을 마쳤습니다.') exit_enter() def update_youtube_urls(my_channel_id=None): path = get_execution_path() input_path = get_input_path_or_exit() cookie_path = os.path.join(get_execution_path(), 'cookies.txt') if not my_channel_id: mychannelid_path = os.path.join(path, '.mychannelid') if os.path.isfile(mychannelid_path): with open(mychannelid_path, 'r') as f: my_channel_id = f.read() else: print('youtube upload 를 먼저 실행해주세요') exit_enter(1) yn = input('기존에 존재하는 youtube_url.txt 도 덮어쓰시겠습니까? [y/n]: ') overwrite = yn == 'y' video_dirs = {} for cur_dir, _, files in os.walk(input_path): if not overwrite and os.path.isfile(os.path.join(cur_dir, 'youtube_url.txt')): continue dir_name = os.path.basename(cur_dir) for filename in files: name, ext = os.path.splitext(filename) if ext == '.mp4' and dir_name.isnumeric(): video_dirs[name] = cur_dir yt = youtube_dl.YoutubeDL(dict(cookiefile=cookie_path)) my_channel_playlist = yt.extract_info( f'https://www.youtube.com/channel/{my_channel_id}', download=False, process=False ).get('url') is_created = False video_urls = {} for video in yt.extract_info(my_channel_playlist, download=False, process=False).get('entries'): title = video['title'] if title.isnumeric() and video_dirs.get(title): is_created = True video_urls[int(title)] = (title, f"https://www.youtube.com/watch?v={video['id']}") if not is_created: print('새로 업로드 된 동영상이 없거나, 아직 업로드가 완전히 완료되지 않았습니다.') print('잠시 후 다시 시도해주세요.') else: for _, (title, url) in sorted(video_urls.items(), key=itemgetter(0)): print(f'make youtube_url.txt: {title}') with open(os.path.join(video_dirs[title], 'youtube_url.txt'), 'w') as f: f.write(url) exit_enter() def qrcode(): input_path = get_input_path_or_exit() chrome_driver_path = get_chrome_driver_path_or_exit() resource_path = get_resource_path() if not os.path.isfile(os.path.join(resource_path, 'DXGulimB-KSCpc-EUC-H.ttf')): print('폰트 파일을 찾을 수 없습니다.') print('DXGulimB-KSCpc-EUC-H.ttf 파일을 "font/" 안에 넣어주세요!') exit_enter(1) naver_qr: Optional[NaverQrCode] = None def walk_dir(): walk_dirs = {} for cur_dir, dirs, files in os.walk(input_path): dir_name = os.path.basename(cur_dir) if not dir_name.isnumeric(): continue if 'youtube_url.txt' not in files: continue if 'qrcode.html' in files: print(f'already created: {dir_name}') continue walk_dirs[int(dir_name)] = (cur_dir, dirs, files) return (v for k, v in sorted(walk_dirs.items(), key=itemgetter(0))) for cur_dir, _, files in walk_dir(): dir_name = os.path.basename(cur_dir) idx = int(dir_name) idx_text = f'{idx:04}' with open(os.path.join(cur_dir, 'youtube_url.txt'), 'r') as f: youtube_url = f.read() if not naver_qr: naver_qr = NaverQrCode() naver_qr.init(chrome_driver_path) print('waiting login ...') naver_qr.login() print('login success') qr_data = naver_qr.create_qr(idx_text, youtube_url).get('QRCodeData', {}) qr_url = qr_data.get('qrCodeUrl') qr_id = qr_data.get('qrcdNo') if not qr_url: print(f'{idx_text}: QR CODE 생성에 실패했습니다') continue with cd(resource_path): print(f'creating "{idx_text}.png"') image = make_qr_image(Size(591, 738), qr_url, idx) # 5cm x 6.25cm (300dpi) with open(os.path.join(cur_dir, f'{idx_text}.png'), 'wb') as f: image.save(f, format='PNG', dpi=(300, 300)) make_redirect_html( os.path.join(cur_dir, 'qrcode.html'), f'https://qr.naver.com/code/updateForm.nhn?qrcdNo={qr_id}' ) if naver_qr: naver_qr.visit_admin_page() print('모든 작업이 끝났습니다.') input('press enter to exit...') if naver_qr: naver_qr.close() sys.exit(0) def organize(): input_path = get_input_path_or_exit() for filename in os.listdir(input_path): filepath = os.path.join(input_path, filename) if not os.path.isfile(filepath) or 'README.txt' == filename: continue name, _ = os.path.splitext(filename) if not name.isnumeric(): print(f'pass: "{filename}" 은 숫자로 이루어진 이름이 아닙니다') continue dir_path = os.path.join(input_path, f'{int(name)}') os.makedirs(dir_path, exist_ok=True) try: os.rename(filepath, os.path.join(dir_path, filename)) print(f'move "{filename}" to "{int(name)}/{filename}"') except Exception as e: print(f'"{filename}" 을 옮기는데 실패하였습니다: {e}') if __name__ == '__main__': parser = argparse.ArgumentParser(description='Chip Helper') subparsers = parser.add_subparsers(help='commands', dest='command', required=True) subparsers.add_parser('makedirs', help='Create dirs like "nnnn" format in a specific path') subparsers.add_parser('organize', help='Create numeric dirs and move video files in it') subparsers.add_parser('thumbnail', help='Create thumbnails') subparsers.add_parser('upload', help='Upload videos to youtube') subparsers.add_parser('youtube-url', help='Make youtube_url.txt in input dirs') subparsers.add_parser('qrcode', help='Generate Naver QR and composite qr image') args = parser.parse_args() func = { 'makedirs': make_dirs, 'thumbnail': make_thumbnail, 'upload': upload_videos, 'youtube-url': update_youtube_urls, 'qrcode': qrcode, 'organize': organize, }.get(args.command) func() print('모든 작업이 완료되었습니다.') exit_enter()
11,162
4,333
# -*- coding: utf-8 -*- import os import io import urllib2 import string from BeautifulSoup import BeautifulSoup import pandas as pd import sys city_url = 'http://twblg.dict.edu.tw/holodict_new/index/xiangzhen_level1.jsp?county=1' def extract_items(base_url): html = urllib2.urlopen(base_url).read() soup = BeautifulSoup(html) #print(soup.prettify()) data = [] table = soup.findAll('tr', attrs={'class':['all_space1', 'all_space2']}) for row in table: cols = row.findAll('td') cols = [ele.text.strip() for ele in cols] data.append([ele for ele in cols if ele]) # Get rid of empty values return data def get_area_url(): base_url = 'http://twblg.dict.edu.tw/holodict_new/index/xiangzhen_level1.jsp?county=%s' url = [] for i in string.ascii_uppercase: url.append(base_url % i) return url if __name__=='__main__': # 縣市名稱 data = extract_items(city_url) data.pop() # ignore data from '其他' print 'Cities and countries are done.' # 鄉鎮區名稱 area_url = get_area_url() for i in area_url: area_data = extract_items(i) data.extend(area_data) print 'Townships are done.' #df = pd.DataFrame(data, columns=['name', 'holo']) df = pd.DataFrame(data) df.to_csv('moe_mapping.csv', encoding='utf-8', index=False, header=0) print 'csv file done.'
1,338
589
import csv #global variables for teams: sharks = [] dragons = [] raptors = [] #read the csv file with the player info and create a player dictionary: def read_players(): player_reader = csv.reader(open('soccer_players.csv')) player_dictionary = {} for row in player_reader: key = row[0] player_dictionary[key] = row[1:] #deleting the first row with column titles: del player_dictionary['Name'] return player_dictionary #distribute kids based on experience: def experienced_players(): exp_kids = [] #calling the previous function to use the player dictionary: player_list = read_players() for keys, values in player_list.items(): if values[1] == "YES": exp_kids.append(keys) return exp_kids def inexperienced_players(): inexp_kids = [] #calling the previous function to use the player dictionary: player_list = read_players() for keys, values in player_list.items(): if values[1] == "NO": inexp_kids.append(keys) return inexp_kids #finalize teams: def make_teams(): #calling a previous function to generate 2 separate lists for experienced and inexperienced kids: ek = experienced_players() iek = inexperienced_players() sharks.extend(ek[0:3]) sharks.extend(iek[0:3]) dragons.extend(ek[3:6]) dragons.extend(iek[3:6]) raptors.extend(ek[6:9]) raptors.extend(iek[6:9]) return sharks, dragons, raptors #update the player dictionary to include the assigned teams: def final_league(): #calling the function to create a player dictionary letter_info = read_players() for keys, values in letter_info.items(): if keys in sharks: values.append("Sharks") if keys in dragons: values.append("Dragons") if keys in raptors: values.append("Raptors") return letter_info #write the league info into the text file: def create_textfile(): files = final_league() with open("teams.txt", "w") as textfile: textfile.write("Sharks" + "\n") for keys,values in files.items(): if values[3] == "Sharks": textfile.write(str(keys) + ", " + str(values[1]) + ", " + str(values[2]) + "\n") textfile.write("\n") textfile.write("Dragons" + "\n") for keys, values in files.items(): if values[3] == "Dragons": textfile.write(str(keys) + ", " + str(values[1]) + ", " + str(values[2]) + "\n") textfile.write("\n") textfile.write("Raptors" + "\n") for keys, values in files.items(): if values[3] == "Raptors": textfile.write(str(keys) + ", " + str(values[1]) + ", " + str(values[2]) + "\n") textfile.write("\n") #generate letters to send the guardians: def letter_generator(): letters = final_league() for keys, values in letters.items(): if values[3] == "Sharks": practice_start = '8am' if values[3] == "Dragons": practice_start = '9am' if values[3] == "Raptors": practice_start = '10am' for keys, values in letters.items(): letter = ("Dear {},\n" "Congratulations! Your child, {}, is selected to play on the {} team. " "The first practice is at {} next Saturday, July 1st." "Thanks,\n" "Coach Deniz".format(values[2],keys,values[3],practice_start)) with open("{}.txt".format((keys.lower()).replace(" ", "_")), "w") as textfile: textfile.write(letter) if __name__ == "__main__": read_players() experienced_players() inexperienced_players() make_teams() create_textfile() final_league() letter_generator()
3,729
1,173
import wpilib import wpilib.drive import ctre import robotmap from wpilib.interfaces import GenericHID RIGHT_HAND = GenericHID.Hand.kRight LEFT_HAND = GenericHID.Hand.kLeft class Robot(wpilib.TimedRobot): def robotInit(self): front_left_motor = ctre.WPI_TalonSRX(robotmap.mecanum['front_left_motor']) back_left_motor = ctre.WPI_TalonSRX(robotmap.mecanum['back_left_motor']) front_right_motor = ctre.WPI_TalonSRX(robotmap.mecanum['front_right_motor']) back_right_motor = ctre.WPI_TalonSRX(robotmap.mecanum['back_right_motor']) front_left_motor.setInverted(True) #back_left_motor.setInverted(True) self.drive = wpilib.drive.MecanumDrive( front_left_motor, back_left_motor, front_right_motor, back_right_motor ) self.drive.setExpiration(0.1) self.lstick = wpilib.XboxController(0) self.rstick = wpilib.XboxController(1) self.gyro = wpilib.AnalogGyro(1) #def teleopInit(self): # self.xforward = 0 # self.yforward = 0 """def operatorControl(self): Called when operation control mode is enabled while self.isOperatorControl() and self.isEnabled(): self.drive.driveCartesian( self.lstick.getX(), self.lstick.getY(), self.rstick.getX(), 0 ) wpilib.Timer.delay(0.04) """ def teleopPeriodic(self): """Called when operation control mode is enabled""" if not self.rstick.getXButton() or not self.lstick.getXButton(): lspeed = deadzone(self.lstick.getX(LEFT_HAND), 0.2) rspeed = deadzone(self.lstick.getY(LEFT_HAND), 0.2) rotate = self.lstick.getX(RIGHT_HAND) else: rotate = 0 lspeed = 0 rspeed = 0 self.drive.driveCartesian( lspeed, rspeed, rotate, self.gyro.getAngle() ) def deadzone(val, deadzone): if abs(val) < deadzone: return 0 elif val < (0): x = ((abs(val) - deadzone)/(1-deadzone)) return (-x) else: x = ((val - deadzone)/(1-deadzone)) return (x) if __name__ == "__main__": wpilib.run(Robot,physics_enabled=True)
2,312
833
from django.contrib import admin from django.urls import path from .views import blog urlpatterns = [ path('', blog, name='blog'), ]
139
42
#!/usr/bin/env python from iris_sdk.models.maps.base_map import BaseMap class LocalRateCenterListMap(BaseMap): rate_center_id = None
139
48
import re import traceback import urllib2 import pandas as pd import json,random,time,datetime from bs4 import BeautifulSoup from pandas.tseries.offsets import YearEnd from sqlalchemy import text from webapp import db, app from webapp.models import FinanceBasic headers = {'User-Agent':'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.1.6) Gecko/20091201 Firefox/3.5.6'} def updateFinanceBasic(code): sql = "select max(report_type) from hexun_finance_basic where code=:code"; resultProxy = db.session.execute(text(sql), {'code': code}) s_date = resultProxy.scalar() if (s_date == None): s_date = '2000.01.01' #获得2000年以来的所有会计年度 index = pd.date_range(s_date, datetime.today(), freq='Q') periods = [x.strftime('%Y.%m.%d') for x in index] #为避免网络访问频繁,分2年为一个阶段分别读取数据取值 l = len(periods) i = 0 while (i <= l): ri = random.randint(2, 5) time.sleep(ri) #延迟执行 updateFinanceBasicByPeriod(periods[i:i+16],code) i = i + 16 def updateFinanceBasicByPeriod(periods,code): df = pd.DataFrame() for x in periods: fb = db.session.query(FinanceBasic).filter_by(code=code, report_type=x).first() if (not fb): fd = findFinanceData(code, x) if not fd.empty: df = df.append(fd) if df.size > 0: df.columns = ['code', 'report_type', 'yysr', 'jlr', 'lrze', 'kjlr', 'zzc', 'gdqy', 'jyjxjl', 'mgsy', 'roe', 'mgjyxjl', 'mgjzc'] tpd = pd.DataFrame({ 'mgsy_ttm': '', 'mgjyxjl_ttm': '' }, index=df.index) df1 = pd.concat([df, tpd], axis=1) ndf = calculateTTMValue(df1, code) ndf.to_sql('hexun_finance_basic', db.engine, if_exists='append', index=False, chunksize=1000) app.logger.info(code + ' finance update done...') return True else: return False def findFinanceData(code,period): period = re.sub('03.31', '03.15', period) url = "http://stockdata.stock.hexun.com/2008/zxcwzb.aspx?stockid="+code+"&accountdate="+period app.logger.info('query stock(' + code + ') finance data url is:' + url) req = urllib2.Request(url =url,headers = headers) feeddata = urllib2.urlopen(req).read() soup = BeautifulSoup(feeddata, "html5lib") paper_name = soup.html.body.find(id="zaiyaocontent").table.tbody if paper_name is None: return pd.DataFrame() else: paper_name = paper_name.find_all('tr') data = [code] for e in paper_name: s = e.find_all('td') i = '0' if s[1].div.text == '--' else re.sub(',', '',s[1].div.text) data.append(i) df = pd.DataFrame(data).T #如果无会计日期,丢弃数据 if df.iat[0, 1]=='': return pd.DataFrame() else: return df.iloc[:,:13] #更新ttm数据 def calculateTTMValue(in_df,code): in_df_date = in_df['report_type'].map(lambda x: pd.to_datetime(x)) df = pd.read_sql_query( "select code,report_type,yysr,jlr,lrze,kjlr,zzc,gdqy,jyjxjl,mgsy,roe,mgjyxjl,mgjzc,mgsy_ttm,mgjyxjl_ttm \ from hexun_finance_basic \ where code=%(name)s", db.engine, params={'name': code}) df = df.append(in_df) i = df['report_type'].map(lambda x: pd.to_datetime(x)) df3 = df.set_index(i) for index, row in df3.iterrows(): if row.mgjyxjl_ttm is None or row.mgjyxjl_ttm == '': # 去年年底 lastYearEnd = YearEnd().rollback(index) # offset = offset.strftime('%Y-%m-%d') lastYearQuart = index - pd.DateOffset(months=12) app.logger.debug(index.strftime('%Y-%m-%d') + ':' + lastYearEnd.strftime('%Y-%m-%d') + ':' + lastYearQuart.strftime( '%Y-%m-%d')) try: if index.quarter != 4: n_mgsy = float(df3.loc[lastYearEnd].mgsy) - float(df3.loc[lastYearQuart].mgsy) + float(row.mgsy) n_mgjyxjl = float(df3.loc[lastYearEnd].mgjyxjl) - float(df3.loc[lastYearQuart].mgjyxjl) + float( row.mgjyxjl) else: n_mgsy = float(row.mgsy) n_mgjyxjl = float(row.mgjyxjl) df3.mgsy_ttm.loc[index] = n_mgsy df3.mgjyxjl_ttm.loc[index] = n_mgjyxjl except Exception, ex: app.logger.warn(traceback.format_exc()) df3.mgsy_ttm.loc[index] = float(row.mgsy) df3.mgjyxjl_ttm.loc[index] = float(row.mgjyxjl) #数据位截取 v_mgsy_ttm = round(df3.mgsy_ttm.loc[index],2) v_mgjyxjl_ttm = round(df3.mgjyxjl_ttm.loc[index],2) #零值处理 v_mgsy_ttm = 0.01 if v_mgsy_ttm==0 else v_mgsy_ttm v_mgjyxjl_ttm = 0.01 if v_mgjyxjl_ttm == 0 else v_mgjyxjl_ttm df3.mgsy_ttm.loc[index] = v_mgsy_ttm df3.mgjyxjl_ttm.loc[index] = v_mgjyxjl_ttm return df3.iloc[df3.index.isin(in_df_date)] #获取每股收益,每股净资产,每股经营现金流 def getPerStockRevenue(): df = pd.read_sql_query("select code,report_type,mgsy_ttm,mgjzc,mgjyxjl_ttm from hexun_finance_basic", db.engine) i = df['report_type'].map(lambda x: pd.to_datetime(x)) df = df.set_index(i) df = df.sort_index(ascending=False) df = df.groupby([df['code']]).first() df = df.reset_index() return df.set_index(df['code'])
5,374
2,255
import tensorflow as tf from tensorflow.examples.tutorials.mnist import input_data mnist = input_data.read_data_sets('MNIST_data', one_hot = True) # Network hyperparameters learning_rate = 0.0001 # 1.95 for sigmoid activation function batch_size = 10 update_step = 10 input_nodes = 784 # 28x38 images as input layer_1_nodes = 500 layer_2_nodes = 500 layer_3_nodes = 500 output_nodes = 10 network_input = tf.placeholder(tf.float32, [None, input_nodes]) target_output = tf.placeholder(tf.float32, [None, output_nodes]) # Network model, weights and biases layer_1 = tf.Variable(tf.random_normal([input_nodes, layer_1_nodes])) layer_2 = tf.Variable(tf.random_normal([layer_1_nodes, layer_2_nodes])) layer_3 = tf.Variable(tf.random_normal([layer_2_nodes, layer_3_nodes])) output_layer = tf.Variable(tf.random_normal([layer_3_nodes, output_nodes])) layer_1_bias = tf.Variable(tf.random_normal([layer_1_nodes])) layer_2_bias = tf.Variable(tf.random_normal([layer_2_nodes])) layer_3_bias = tf.Variable(tf.random_normal([layer_3_nodes])) output_layer_bias = tf.Variable(tf.random_normal([output_nodes])) # Feedforward calculations layer_1_out = tf.nn.relu(tf.matmul(network_input, layer_1) + layer_1_bias) layer_2_out = tf.nn.relu(tf.matmul(layer_1_out, layer_2) + layer_2_bias) layer_3_out = tf.nn.relu(tf.matmul(layer_2_out, layer_3) + layer_3_bias) network_out_1 = tf.matmul(layer_3_out, output_layer) + output_layer_bias network_out_2 = tf.nn.softmax(network_out_1) cost_function = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits = network_out_1, labels = target_output)) training_step = tf.train.GradientDescentOptimizer(learning_rate).minimize(cost_function) predicitions = tf.equal(tf.argmax(network_out_2, 1), tf.argmax(target_output, 1)) accuracy = tf.reduce_mean(tf.cast(predicitions, tf.float32)) # Running the neural network with tf.Session() as session: session.run(tf.global_variables_initializer()) no_epochs = 10 for epoch in range(no_epochs): total_cost = 0 no_batches = int(mnist.train.num_examples / batch_size) for batch in range(no_batches): input_data, labels = mnist.train.next_batch(batch_size) step, cost = session.run([training_step, cost_function], feed_dict = {network_input: input_data, target_output: labels}) total_cost += cost print('Epoch {} out of {} completed, loss: {}'.format(epoch, no_epochs, total_cost)) print('Accuracy: {}'.format(accuracy.eval({network_input: mnist.test.images, target_output: mnist.test.labels})))
2,555
942
class Solution(object): def isHappy(self, n): """ :type n: int :rtype: bool """ seen = set() while (True): sumOfSquaredDigits = 0 while n > 0: digit = n % 10 n = n // 10 sumOfSquaredDigits += digit**2 if sumOfSquaredDigits == 1: return True if sumOfSquaredDigits in seen: return False seen.add(sumOfSquaredDigits) n = sumOfSquaredDigits # seen = [] # while (True): # print (seen) # digits = self.getDigits(n) # total = 0 # print ("Digits: " + str(digits)) # for i in digits: # total += int(pow(i,2)) # if (total in seen): # return False # if (total == 1): # return True # seen.append(total) # n = total # def getDigits(self, n): # digits = [] # while (n > 0): # digits.append(n%10) # n = n//10 # return digits
1,157
340
from dataclasses import dataclass, field from dataclazzes.track import Track @dataclass class Playlist: """Class for storing playlist information.""" id: str title: str tracks: [Track] = field(default_factory=list) @staticmethod def from_raw(raw_playlists: list): return [ Playlist(id=playlist['playlistId'], title=playlist['title']) for playlist in raw_playlists ]
434
127
#!/usr/bin/env python3 # pylint: disable=C0103 """ Définie la classe entity Permet de modeliser le personnage et des monstre """ from random import choice from vect import Vect from astar import calc_path_astart import chars class Player(): """ Classe Player : """ BULLET_MAX = 10 HP_MAX = 10 START_MONEY = 0 def __init__(self): """ Personnage """ self.pos = Vect(0, 0) self.direction = Vect(1, 0) self.distance_view = 7 self.bullet = self.BULLET_MAX self.hp = self.HP_MAX self.level = 0 self.money = self.START_MONEY self.sword_damage = 1 self.gun_damage = 2 def level_up(self, pos): """ Le personnage gagne un level """ self.level += 1 self.pos = pos def g_case_visible(self, mat_collide): """ retourne sur toutes les cases visibles par self dans mat_collide """ # Nb : prend les segments depuis un cercle et non un rect # n'est pas OK border = self.pos.g_rect(Vect(self.distance_view, self.distance_view)) for bordure_pos in border: for pos in self.pos.g_bresenham_line(bordure_pos): if self.pos.distance(pos) >= self.distance_view: break if not Vect(0, 0) <= pos < Vect(len(mat_collide), len(mat_collide[0])): break if not mat_collide[pos.x][pos.y]: yield pos break yield pos def shoot(self): """ Tire une nouvelle balle """ self.bullet -= 1 return Bullet(self.pos, self.direction, self.gun_damage) def strike(self, mat_collide): """ Donne un coup d'épée """ return Sword(self.pos + self.direction, self.sword_damage) def add_money(self, value): """ Ajoute des pièces au Player """ assert value >= 0 self.money += value return True def add_hp(self, value): """ Ajoute des HP au Player """ assert value >= 0 if self.hp == self.HP_MAX: return False self.hp = min(self.hp + value, self.HP_MAX) return True def add_bullets(self, value): """ Ajoute des balles au Player """ assert value >= 0 if self.bullet == self.BULLET_MAX: return False self.bullet = min(self.bullet + value, self.BULLET_MAX) return True def update(self, mat_collide, depl_vect): """ Met à jour la position du personnage en fonction des evenements et de mat_collide """ if depl_vect != Vect(0, 0): self.direction = depl_vect new_pos = self.pos + depl_vect # Tests de collision (Diagonales) if mat_collide[new_pos.x][self.pos.y]: # premier chemin libre en x if mat_collide[new_pos.x][new_pos.y]: # deuxieme chemin libre en y self.pos = new_pos else: # deuxieme chemin bloque en y self.pos.x = new_pos.x elif mat_collide[self.pos.x][new_pos.y]: # premier chemin libre en y if mat_collide[new_pos.x][new_pos.y]: # deuxieme chemin libre en x self.pos = new_pos else: # deuxieme chemin bloque en x self.pos.y = new_pos.y else: # Aucun chemin libre # Do nothind pass def render(self): """ Retourne le char à afficher """ return chars.C_PLAYER def __str__(self): """ Retourne une chaine d'affichage """ heal_str = ('\u2665' * int(self.hp / self.HP_MAX * 10) + ' ' * (10-int(self.hp / self.HP_MAX * 10))) bullet_str = ('|' * int(self.bullet) + ' ' * int(self.BULLET_MAX - self.bullet)) return ('Position : {} | HP : [' + heal_str + '] | Bullets [' + bullet_str + ']') class Bullet: """ Classe Bullet : """ def __init__(self, pos, directions, dammage): """ Personnage """ self.pos = pos self.direction = directions self.dammage = dammage def update(self, mat_collide): """ Met à jour la balle Retourne 1 si elle touche un obstacle """ self.pos += self.direction return mat_collide[self.pos.x][self.pos.y] def render(self): """ Retourne le char à afficher """ return chars.C_BULLETS[int(self.direction.angle()/2/3.1415 * 8)] def __str__(self): return "(*:{})".format(self.pos) class Monster: """ Monster """ IDLE = 0 RUN = 1 DECOMPOSITION = 2 SHOCKED = 3 def __init__(self, pos, dammage, health): """ Personnage """ self.pos = pos self.dammage = dammage self.health = health self.shocked = 0 self.state = self.IDLE self.ttd = 8 # TIme to die # TEMP # Le chemin du monstre au joueur self.path = [] def update(self, mat_collide, player_pos): """ Met à jour l'enemie """ if self.state == self.SHOCKED: if self.shocked: self.shocked -= 1 else: self.state = self.IDLE return False if self.state == self.IDLE or self.state == self.RUN: if self.pos.distance(player_pos) <= 10: self.state = self.RUN else: self.state = self.IDLE if self.state == self.RUN: self.path = calc_path_astart(mat_collide, self.pos, player_pos) if self.path != []: self.pos = self.path[0] if self.state == self.IDLE: # TODO: Depl aléatoire pass return False self.ttd -= 1 return self.ttd == 0 # Mort def render(self): """ Retourne le char à afficher """ if self.state == self.RUN: return chars.C_MONSTER_RUN return chars.C_MONSTERS[self.ttd % len(chars.C_MONSTERS)] def kill(self): """ Elimine le mechant """ if self.health <= 0: self.state = self.DECOMPOSITION else: self.state = self.SHOCKED self.shocked = 4 def __str__(self): return "(*:{})".format(self.pos) class Treasure: """ Trésor, peut contenir 3 types d'objet différents : * des sous * des munitions * de la vie """ HEART = 0 BULLET = 1 GOLD = 2 STRENGH = 3 POWER = 4 CHARS = [chars.C_HEART, chars.C_BULLET_CHRG, chars.C_MONEY] def __init__(self, pos, value): """ Init """ self.pos = pos if value == 1: self.object = choice([self.HEART, self.BULLET, self.GOLD]) else: self.cpt = -1 if value == 2: self.object = self.STRENGH else: self.object = self.POWER self.value = value def render(self): """ Render """ if self.value == 1: return self.CHARS[self.object] self.cpt += 1 if self.value == 2: return chars.C_TRE_WEAPON[self.cpt % len(chars.C_TRE_WEAPON)] return chars.C_TRE_GUN[self.cpt % len(chars.C_TRE_GUN)] def get_value(self): """ Retourne la valeur du contenue du coffre """ return self.value class Sword: """ coup d'épée venant du joueur """ DELTA_POSS = list(Vect(0, 0).g_rect(Vect(1, 1))) def __init__(self, pos, dammage): self.pos = pos self.cpt = len(self.DELTA_POSS)-1 self.dammage = dammage def update(self, mat_collide, player_pos): """ Met à jour l'enemie """ if self.cpt < 0: return True self.pos = player_pos + self.DELTA_POSS[self.cpt] self.cpt -= 1 return False def render(self): """ render """ return chars.C_SWORDS[- self.cpt % len(chars.C_SWORDS)] class Door: """ La porte de sortie """ def __init__(self, pos): """ Init """ self.pos = pos self.cpt = 0 def update(self): """ Update """ def render(self): """ Render """ self.cpt += 1 return chars.C_DOORS[self.cpt % len(chars.C_DOORS)] def main(): """ Test unitaire """ if __name__ == '__main__': main()
9,091
2,962
import streamlit as st def app(): import plotly.express as px import plotly.graph_objects as go from textblob import TextBlob import tweepy import sys import pandas as pd api_key = 'q7QHHHAKEwd5igoUvVrx5sCiw' api_secret_key = 'i7uhcFirM38bnbYscv32beJnMpsmMxFdYSHitwfSCPIeMj7Lcs' access_token = '916414257993879552-kWKlelyL9e6HGH40wcdawT8CiCvO3Hz' access_token_secret= 'zYflOPrxysrdOsQiAhp8gmJjAtwRMUcSyX6KlexMk03eB' auth_handler = tweepy.OAuthHandler(consumer_key = api_key,consumer_secret=api_secret_key) auth_handler.set_access_token(access_token,access_token_secret) api = tweepy.API(auth_handler) searchwhat = st.sidebar.text_input("Search Term", 'Zomato IPO') tweet_amount = int(st.sidebar.text_input('Tweet Amount', '50')) tweets = tweepy.Cursor(api.search,q=searchwhat,lang='en').items(tweet_amount) polarity,positive,negative,neutral = 0,0,0,0 tweetlist = [] polaritylist = [] for tweet in tweets: final_text = tweet.text.replace('RT','') if final_text.startswith(' @'): position = final_text.index(':') final_text = final_text[position+2:] if final_text.startswith('@'): position = final_text.index('') final_text = final_text[position+2:] analysis = TextBlob(final_text) tweet_polarity = analysis.polarity if tweet_polarity>0: positive+=1 polaritylist.append('positive') elif tweet_polarity<0: negative+=1 polaritylist.append('negative') else: neutral+=1 polaritylist.append('neutral') polarity += tweet_polarity tweetlist.append(final_text) labels = ['Positive','Negative','Neutral'] values = [positive,negative,neutral] st.write(f'The Sentiment Analysis for Search Term : {searchwhat}') # Use `hole` to create a donut-like pie chart fig = go.Figure(data=[go.Pie(labels=labels, values=values, hole=.5)]) st.plotly_chart(fig) # tweetcontainer = pd.DataFrame(list(zip(tweetlist,polaritylist)),columns=['Tweets','Sentiment']) # st.write(tweetcontainer) if len(tweetlist)<10: def showTweets(index=0,limit=len(tweetlist)): while(index<limit): st.write(tweetlist[index]) st.write(polaritylist[index]) index+=1 with st.beta_expander('See the Tweets'): showTweets() else: def showTweets(index=0,limit=10): while(index<limit): st.write(tweetlist[index]) st.write(polaritylist[index]) index+=1 with st.beta_expander('See the Tweets'): showTweets() st.subheader('RHP Analysis') #st.write('Enter any text from the RHP here') user_input = st.text_area('Enter any section from the RHP here') rhpanalysis = TextBlob(user_input) rhppolarity = rhpanalysis.polarity if rhppolarity>0: ipo_outlook = 'Positve' elif rhppolarity<0: ipo_outlook = 'Negative' else: ipo_outlook = 'Neutral' with st.beta_expander('See the Analysis'): st.subheader(f'Outlook of the Particular IPO is {ipo_outlook} with polarity :{rhppolarity}')
3,488
1,276
import unittest import logging from varfilter import varfilter, filter class TestVarfilter(unittest.TestCase): def setUp(self): print("Preparando el contexto") self.source1 = {"teststr1": "Value teststr1", "testint1": 10} self.source2 = {"teststr2": "Value teststr2", "testint2": 10, "teststr": "Value str"} def test_fVar_default_nonexistent(self): print("fVar con valor por defecto no existente") t = varfilter.fVar("test1", "default value", None, self.source1) self.assertEqual(t, "default value") def test_fVar_default_nonexistent_int(self): print("fVar con valor por defecto no existente de tipo int") t = varfilter.fVar("test1", 0, "int", self.source1) self.assertEqual(t, 0) def test_fVar_default_existent(self): print("fVar con valor por defecto") t = varfilter.fVar("teststr1", "default value", None, self.source1) self.assertEqual(t, "Value teststr1") def test_fVar_default_existent_int(self): print("fVar con valor por defecto existente de tipo int") t = varfilter.fVar("testint1", 0, "int", self.source1) self.assertEqual(t, 10) def test_fVar_default_existent_int_filtererror(self): print("fVar con valor por defecto erroneo existente de tipo int") self.assertRaises(filter.IntegerError, varfilter.fVar, 'test1', 'error', 'int', self.source1) def test_fVar_existent_severalsources(self): print("fVar con valor existente varios sources") t = varfilter.fVar("teststr", "default value", None, self.source1, self.source2) self.assertEqual(t, "Value str") if __name__ == '__main__': logging.basicConfig(format='%(asctime)s - %(message)s', level=logging.DEBUG) unittest.main()
2,447
668
"""Test cases for Arithmatex.""" from .. import util class TestArithmatexLimit(util.MdCase): """Test limiting Arithmatex inline and block inputs.""" extension = [ 'pymdownx.arithmatex' ] extension_configs = {'pymdownx.arithmatex': {'inline_syntax': ['round'], 'block_syntax': ['square']}} def test_round_only(self): """Test round only.""" self.check_markdown( "\\(1 + 2 + 3\\)", """<p><span class="arithmatex"><span class="MathJax_Preview">1 + 2 + 3</span><script type="math/tex">1 + 2 + 3</script></span></p>""" # noqa: E501 ) def test_square_only(self): """Test square only.""" self.check_markdown( r""" \[ 1 + 2 + 3 \] """, r""" <div class="arithmatex"> <div class="MathJax_Preview"> 1 + 2 + 3 </div> <script type="math/tex; mode=display"> 1 + 2 + 3 </script> </div> """, True ) class TestArithmatexBlockEscapes(util.MdCase): """Test escaping cases for Arithmatex blocks.""" extension = [ 'pymdownx.arithmatex' ] extension_configs = {} def test_escaped_dollar_block(self): """Test escaping a dollar.""" self.check_markdown( r''' $$3+2\$$ ''', r''' <p>$<span class="arithmatex"><span class="MathJax_Preview">3+2\$</span><script type="math/tex">3+2\$</script></span></p> ''', # noqa: E501 True ) def test_escaped_dollar_dollar_block(self): """Test escaping both dollars.""" self.check_markdown( r''' $$3+2\$\$ ''', r''' <p>$$3+2$$</p> ''', True ) def test_double_escaped_dollar_block(self): """Test double escaping a dollar.""" self.check_markdown( r''' $$3+2\\$$ ''', r''' <div class="arithmatex"> <div class="MathJax_Preview">3+2\\</div> <script type="math/tex; mode=display">3+2\\</script> </div> ''', True ) def test_escaped_end_block(self): """Test escaping an end.""" self.check_markdown( r''' \begin{align}3+2\\end{align} ''', r''' <p>\begin{align}3+2\end{align}</p> ''', True ) def test_double_escaped_end_block(self): """Test double escaping an end.""" self.check_markdown( r''' \begin{align}3+2\\\end{align} ''', r''' <div class="arithmatex"> <div class="MathJax_Preview">\begin{align}3+2\\\end{align}</div> <script type="math/tex; mode=display">\begin{align}3+2\\\end{align}</script> </div> ''', True ) def test_escaped_bracket_block(self): """Test escaping a bracket.""" self.check_markdown( r''' \[3+2\\] ''', r''' <p>[3+2\]</p> ''', True ) def test_double_escaped_bracket_block(self): """Test double escaping a bracket.""" self.check_markdown( r''' \[3+2\\\] ''', r''' <div class="arithmatex"> <div class="MathJax_Preview">3+2\\</div> <script type="math/tex; mode=display">3+2\\</script> </div> ''', True ) class TestArithmatexHang(util.MdCase): """Test hang cases.""" def test_hang_dollar(self): """ We are just making sure this works. Previously this pattern would hang. It isn't supposed to match due to the space before the last dollar, but it definitely shouldn't hang the process. """ self.check_markdown( r''' $z^{[1]} = \begin{bmatrix}w^{[1]T}_1 \\ w^{[1]T}_2 \\ w^{[1]T}_3 \\ w^{[1]T}_4 \end{bmatrix} \begin{bmatrix}x_1 \\ x_2 \\ x_3 \end{bmatrix} + \begin{bmatrix}b^{[1]}_1 \\ b^{[1]}_2 \\ b^{[1]}_3 \\ b^{[1]}_4 \end{bmatrix}= \begin{bmatrix}w^{[1]T}_1 x + b^{[1]}_1 \\ w^{[1]T}_2 x + b^{[1]}_2\\ w^{[1]T}_3 x + b^{[1]}_3 \\ w^{[1]T}_4 x + b^{[1]}_4 \end{bmatrix} $ ''', # noqa: E501 r''' <p>$z^{[1]} = \begin{bmatrix}w^{[1]T}_1 \ w^{[1]T}_2 \ w^{[1]T}_3 \ w^{[1]T}_4 \end{bmatrix} \begin{bmatrix}x_1 \ x_2 \ x_3 \end{bmatrix} + \begin{bmatrix}b^{[1]}_1 \ b^{[1]}_2 \ b^{[1]}_3 \ b^{[1]}_4 \end{bmatrix}= \begin{bmatrix}w^{[1]T}_1 x + b^{[1]}_1 \ w^{[1]T}_2 x + b^{[1]}_2\ w^{[1]T}_3 x + b^{[1]}_3 \ w^{[1]T}_4 x + b^{[1]}_4 \end{bmatrix} $</p> ''', # noqa: E501 True )
4,989
1,756
# -*- coding: utf-8 -*- from setuptools import setup setup( name="UrRtde", packages=["rtde"], version=1.0, description="Real-Time Data Exchange (RTDE) python client + examples", )
197
71
""" Task: Given a roman numeral, convert it to an integer. Input is guaranteed to be within the range from 1 to 3999. Symbol Value I 1 (unus) V 5 (quinque) X 10 (decem) L 50 (quinquaginta) C 100 (centum) D 500 (quingenti) M 1,000 (mille) >>> roman_to_int("DCXXI") 621 >>> roman_to_int("VI") 6 >>> roman_to_int("LXXVI") 76 >>> roman_to_int("XIII") 13 >>> roman_to_int("MMMCMXCIX") 3999 >>> roman_to_int("") 0 """ def roman_to_int(s): values = {'M': 1000, 'D': 500, 'C': 100, 'L': 50, 'X': 10, 'V': 5, 'I': 1} result = 0 for i in range(len(s)): if i == len(s) - 1: result += values[s[i]] else: if values[s[i]] < values[s[i + 1]]: result -= values[s[i]] else: result += values[s[i]] return result if __name__ == '__main__': import doctest doctest.testmod()
891
401
import mock import pytest from rump import parser, Server, Upstream, Selection, exc @pytest.fixture def parse(): return parser.for_upstream() def test_upstream_valid(parse): cases = [ ('hi', Upstream(Selection(Server('http', 'hi'), 1))), ('https://hi', Upstream(Selection(Server('https', 'hi'), 1))), ('https://hi,23', Upstream(Selection(Server('https', 'hi'), 23))), ('https://hi,23 http://there/you,33 bye', Upstream([ Selection(Server('https', 'hi'), 23), Selection(Server('http', 'there/you'), 33), Selection(Server('http', 'bye'), 1), ])), ] for raw, expected in cases: parsed = parse(raw) assert expected == parsed def test_upstream_invalid(parse): cases = [ ('hi,nope', exc.ParseException), ('hi,-1', exc.ParseException), ('', exc.ParseException), ] for raw, ex in cases: with pytest.raises(ex): parse(raw) def test_upstream_select(parse): upstream = parse('http://1:81,1 http://2:82,5 https://4:84,3') assert not upstream.uniform servers = [selection.server for selection in upstream] for _ in xrange(20): server = upstream() assert server in servers def test_upstream_select_uniform(parse): upstream = parse('http://1:81 http://2:82 https://4:84') assert [ Server('http', '1:81'), Server('http', '2:82'), Server('https', '4:84') ] == upstream.servers assert upstream.uniform servers = [selection.server for selection in upstream] for _ in xrange(20): server = upstream() assert server in servers def test_upstream_select_impossible(parse): upstream = parse('http://1:81,1 http://2:825 https://4:84,3') with mock.patch('rump.upstream.random.randint') as randint: randint.return_value = upstream.total with pytest.raises(Exception): upstream() randint.return_value = upstream.total - 1 upstream()
2,082
669
from django.contrib import admin from places.models import County, Constituency, Province, District from places.models import Division, Location, SubLocation, SchoolZone class CountyAdmin(admin.ModelAdmin): search_fields = ['name'] class ProvinceAdmin(admin.ModelAdmin): search_fields = ['name'] class DistrictAdmin(admin.ModelAdmin): search_fields = ['name'] class DivisionAdmin(admin.ModelAdmin): search_fields = ['name'] class ConstituencyAdmin(admin.ModelAdmin): search_fields = ['name'] class LocationAdmin(admin.ModelAdmin): search_fields = ['name'] class SubLocationAdmin(admin.ModelAdmin): search_fields = ['name'] class SchoolZoneAdmin(admin.ModelAdmin): search_fields = ['name'] admin.site.register(County, CountyAdmin) admin.site.register(Province, ProvinceAdmin) admin.site.register(District, DistrictAdmin) admin.site.register(Division, DivisionAdmin) admin.site.register(Constituency, ConstituencyAdmin) admin.site.register(Location, LocationAdmin) admin.site.register(SubLocation, SubLocationAdmin) admin.site.register(SchoolZone, SchoolZoneAdmin)
1,113
322
"""pyGEEMs: Geotechnical earthquake engineering models implemented in Python.""" import pathlib from pkg_resources import get_distribution import scipy.constants FPATH_DATA = pathlib.Path(__file__).parent / "data" KPA_TO_ATM = scipy.constants.kilo / scipy.constants.atm __author__ = "Albert Kottke" __copyright__ = "Copyright 2018 Albert Kottke" __license__ = "MIT" __title__ = "pygeems" __version__ = get_distribution("pygeems").version from . import dyn_props from . import ground_motion from . import slope_disp
520
181
from tkinter import * from tkinter.ttk import * # creando ventana tkinter root = Tk() # Agregando herramientas a la ventana Label(root, text = 'PuntosExtra', font =( 'Verdana', 15)).pack(side = TOP, pady = 10) # Insertando la imagen de login foto = PhotoImage(file = r"C:\Users\david\OneDrive\Imágenes\login.png") Button(root, text = 'Click Me !', image = foto).pack(side = TOP) mainloop()
414
170
from django.conf.urls import patterns, url, include from corehq.apps.api.urls import CommCareHqApi from custom.ewsghana.resources.v0_1 import EWSLocationResource from custom.ewsghana.views import EWSConfigView, EWSGlobalStats, InputStockView, EWSUserExtensionView, \ DashboardRedirectReportView hq_api = CommCareHqApi(api_name='v0.3') hq_api.register(EWSLocationResource()) urlpatterns = patterns('custom.ewsghana.views', url(r'^ews_config/$', EWSConfigView.as_view(), name=EWSConfigView.urlname), url(r'^sync_ewsghana/$', 'sync_ewsghana', name='sync_ewsghana'), url(r'^global_stats/$', EWSGlobalStats.as_view(), name=EWSGlobalStats.urlname), # for testing purposes url(r'^ews_sync_stock_data/$', 'ews_sync_stock_data', name='ews_sync_stock_data'), url(r'^ews_clear_stock_data/$', 'ews_clear_stock_data', name='ews_clear_stock_data'), url(r'^configure_in_charge/$', 'configure_in_charge', name='configure_in_charge'), url(r'^ews_resync_web_users/$', 'ews_resync_web_users', name='ews_resync_web_users'), url(r'^inventory_managment/$', 'inventory_management', name='inventory_managment'), url(r'^stockouts_product/$', 'stockouts_product', name='stockouts_product'), url(r'^ews_fix_locations/$', 'ews_fix_locations', name='ews_fix_locations'), url(r'^ews_add_products_to_locs/$', 'ews_add_products_to_locs', name='ews_add_products_to_locs'), url(r'^migrate_email_settings/$', 'migrate_email_settings_view', name='migrate_email_settings'), url(r'^fix_sms_users/$', 'fix_sms_users', name='fix_sms_users'), url(r'^delete_last_stock_data/$', 'delete_last_stock_data', name='delete_last_stock_data'), url(r'^(?P<site_code>\w+)/input_stock/$', InputStockView.as_view(), name='input_stock'), url(r'^', include(hq_api.urls)), url(r'^convert_user_data_fields/$', 'convert_user_data_fields', name='convert_user_data_fields'), url(r'^non_administrative_locations/$', 'non_administrative_locations_for_select2'), url(r'^user_settings/(?P<user_id>[ \w-]+)/$', EWSUserExtensionView.as_view(), name='ews_user_settings'), url(r'^dashboard/(?P<site_code>\w+)/', DashboardRedirectReportView.as_view(), name='dashboard_report') )
2,200
839
import numpy as np from scipy.spatial.distance import cdist import sys from plot_area import plot_area COLOR = ['tab:blue', 'tab:orange', 'tab:green'] class BaseClassifier: d = -1 c = -1 def __init__(self, d): super().__init__() if (d <= 0): raise RuntimeError("Classifier.D/C cannot be negative or zero") self.d = d def evaluate(self, feature): if (np.shape(feature)[1] != self.d): raise RuntimeError("Size of feature does not fit") if (self.c < 0): raise RuntimeError("Evaluate with an untrained classifier") return 0 def train(self, data): return self class NearestMeanClassifier(BaseClassifier): training_features = np.array([]) training_classifications = np.array([]) feature_means = np.array([]) classifications = np.array([]) num_of_classifications = np.array([]) weights_group = np.array([]) def __init__(self, d): super().__init__(d) def evaluate(self, feature): if (super().evaluate(feature) or np.size(self.weights_group) == 0): return None length = np.shape(feature)[0] min_distance = np.full(length, sys.maxsize) classification = np.full(length, None) for i in range(self.c): weights = self.weights_group[i] result = np.matmul(weights, np.transpose(np.insert(feature, 0, 1, axis = 1))) classification[np.all(result > 0, axis = 0)] = self.classifications[i] return classification def test(self, feature, classification): eval_classification = self.evaluate(feature) return np.count_nonzero(eval_classification == classification) / \ np.size(classification) def train_by_weights_group(self, weights_group, classifications): self.weights_group = weights_group self.classifications = classifications self.c = len(self.classifications) def train(self, data, training_type = 'ovo'): self.training_features = data[:, :-1] self.training_classifications = data[:, -1] if (np.shape(self.training_features)[1] != self.d): raise RuntimeError("Size of training data does not fit") self.classifications = np.unique(self.training_classifications) self.c = len(self.classifications) self.feature_means = np.zeros([self.c, self.d]) self.num_of_classifications, _ = np.histogram(self.training_classifications, \ bins = np.append(self.classifications, np.max(self.classifications) + 1)) for i in range(self.c): self.feature_means[i] = np.mean(\ self.training_features[self.training_classifications == self.classifications[i]], \ axis = 0) if (training_type == 'ovo' or training_type == 'mvm'): weights_group = np.array([]) for i in range(self.c): classification = self.classifications[i] weights = np.array([]) for j in range(self.c): if (i == j): continue another_classification = self.classifications[j] weight = self.get_boundary_weight_by_ovo(classification, another_classification) if (np.size(weights) == 0): weights = np.array([weight]) else: weights = np.append(weights, [weight], axis = 0) if (np.size(weights_group) == 0): weights_group = np.array([weights]) else: weights_group = np.append(weights_group, [weights], axis = 0) self.weights_group = weights_group elif (training_type == 'ovr'): weights = np.array([]) for classification in self.classifications: weight = self.get_boundary_weight_by_ovr(classification) if (np.size(weights) == 0): weights = np.array([weight]) else: weights = np.append(weights, [weight], axis = 0) weights_group = np.tile(-weights, (self.c, 1, 1)) for i in range(self.c): weights_group[i][i] = -weights_group[i][i] self.weights_group = weights_group return self def get_boundary_weight_by_ovo(self, classification, another_classification): if (self.c < 0): raise RuntimeError("Get boundary with an untrained classifier") return self._get_boundary_weight_by_two_means(\ self.feature_means[self.classifications == classification], \ self.feature_means[self.classifications == another_classification]) def get_boundary_weight_by_ovr(self, classification): if (self.c < 0): raise RuntimeError("Get boundary with an untrained classifier") i = np.argwhere(self.classifications == classification) feature_mean = self.feature_means[i] non_feature_means = np.delete(self.feature_means, i, axis = 0) num_of_non_classifications = np.delete(self.num_of_classifications, i) non_feature_mean = np.average(non_feature_means, weights = num_of_non_classifications, axis = 0) return self._get_boundary_weight_by_two_means(feature_mean, non_feature_mean) def plot_features(self, ax): if (np.size(self.training_features) == 0 or self.c < 0): raise RuntimeError('Plot with an untrained classifier or not applicable') for i in range(self.c): features = self.training_features[self.training_classifications == self.classifications[i], :] ax.scatter(features[:, 0], features[:, 1], label = 'Class ' + str(int(self.classifications[i])), \ marker = 'x', alpha = 0.8, color = COLOR[i]) def plot_means(self, ax): if (np.size(self.feature_means) == 0 or self.c < 0): raise RuntimeError('Plot with an untrained classifier or not applicable') for i in range(self.c): ax.scatter(self.feature_means[i][0], self.feature_means[i][1], label = 'Class ' + str(self.classifications[i]) + ' Mean', \ linewidth = 8, marker = 'o', color = COLOR[i]) def plot_decision_regions_boundaries(self, ax): if (self.c < 0): raise RuntimeError('Plot with an untrained classifier') for i in range(self.c): plot_area(ax, [self.weights_group[i]], self.get_plot_linespace(), \ label = 'Gamma_' + str(int(self.classifications[i])), \ color = COLOR[i], alpha = 0.6) def plot_indeterminate_regions(self, ax): if (self.c < 0): raise RuntimeError('Plot with an untrained classifier') indeterminate_weights_group = self._get_complement_weights_group(self.weights_group) plot_area(ax, indeterminate_weights_group, self.get_plot_linespace(), \ label = 'Indeterminate', \ color = 'tab:grey') def get_plot_linespace(self): linespace_lim = self._get_plot_limit()[0] linespace = np.arange(linespace_lim[0], linespace_lim[1], 0.002) return linespace def plot_setlim(self, ax): plot_limit = self._get_plot_limit() ax.set_xlim(plot_limit[0][0], plot_limit[0][1]) ax.set_ylim(plot_limit[1][0], plot_limit[1][1]) def plot_legend(self, ax, **kwargs): handles, labels = ax.get_legend_handles_labels() newLabels, newHandles = [], [] for handle, label in zip(handles, labels): if label not in newLabels: newLabels.append(label) newHandles.append(handle) ax.legend(newHandles, newLabels, **kwargs) def _get_plot_limit(self): if (np.size(self.training_features) == 0): return np.array([[0, 6], [0, 6]]) min = np.min(self.training_features, axis = 0) max = np.max(self.training_features, axis = 0) min = min - np.maximum((max - min) * 0.25, [0.5, 0.5]) max = max + np.maximum((max - min) * 0.25, [0.5, 0.5]) return np.transpose([min, max]) def _get_boundary_weight_by_two_means(self, a, b): return np.append(np.sum(np.square(b)) - np.sum(np.square(a)), 2 * (a - b)) def _get_complement_weights_group(self, weights_group, complement_weights_group = np.array([]), weights = np.array([])): _weights = weights for weight in weights_group[0]: if (np.size(_weights) == 0): weights = np.array([-weight]) else: weights = np.append(_weights, [-weight], axis = 0) if (np.shape(weights_group)[0] > 1): complement_weights_group = self._get_complement_weights_group(weights_group = weights_group[1:, :, :], \ complement_weights_group = complement_weights_group, weights = weights) else: if (np.size(complement_weights_group) == 0): complement_weights_group = np.array([weights]) else: complement_weights_group = np.append(complement_weights_group, [weights], axis = 0) return complement_weights_group
9,288
2,865
import sys from utils import * class DiscussManager: def add_discuss(self, problem_id: int, username: str, data: str): db = db_connect() cursor = db.cursor() try: cursor.execute("INSERT INTO Discuss(Problem_ID, Username, Data, Time) VALUES(%s, %s, %s, %s)", (problem_id, username, data, unix_nano())) db.commit() except pymysql.Error: db.rollback() sys.stderr.write("SQL Error in DiscussManager: Add_Discuss\n") db.close() return def modify_discuss(self, discuss_id: int, new_data: str): db = db_connect() cursor = db.cursor() try: cursor.execute("UPDATE Discuss SET DATA = %s WHERE ID = %s", (new_data, discuss_id)) db.commit() except pymysql.Error: db.rollback() sys.stderr.write("SQL Error in DiscussManager: Modify_Discuss\n") db.close() return def get_author(self, discuss_id: int): db = db_connect() cursor = db.cursor() cursor.execute("SELECT Username FROM Discuss WHERE ID = %s", discuss_id) ret = cursor.fetchone() return ret[0] def get_discuss_for_problem(self, problem_id: int): db = db_connect() cursor = db.cursor() cursor.execute("SELECT ID, Username, DATA, Time FROM Discuss WHERE Problem_ID = %s", problem_id) ret = cursor.fetchall() db.close() return ret def delete_discuss(self, discuss_id: int): db = db_connect() cursor = db.cursor() try: cursor.execute("DELETE FROM Discuss WHERE ID = %s", discuss_id) db.commit() except pymysql.Error: db.rollback() sys.stderr.write("SQL Error in DiscussManager: Erase_Discuss\n") db.close() return Discuss_Manager = DiscussManager()
1,920
581
import math, sys, random, mcint from scipy import integrate import numpy as np gap = float(sys.argv[1]) lam = float(sys.argv[2]) print(gap, lam) ## calculate the yukawa force over a distributed test mass assumed to be cube D = 5 # diameter of bead (um) rhob = 2e3 # density bead (kg/m^3) rhoa = 19.3e3 # density attractor rhosi = 2.3e3 # density attractor a = 10 # length of attractor cube side (um) a_depth = 200 # depth of attractor cube side (um) au_thick = 0.2 # shield layer thickness (um) def dV(phi,theta,r): return r**2 * math.sin(theta) alpha = 1.0 G = 6.67398e-11 #def Fg(phi, theta, r, currx, curry, currz): def integrand(xin): r = xin[0] theta = xin[1] phi = xin[2] currx = xin[3] curry = xin[4] currz = xin[5] ## distance between r,theta,phi point and currx,curry,currz measured relative to center of cube at (gap + a/2, 0, 0) dx = r*math.sin(theta)*math.cos(phi) - (gap + a_depth/2.0 + currx) dy = r*math.sin(theta)*math.sin(phi) - curry dz = r*math.cos(theta) - currz dist = math.sqrt( dx**2 + dy**2 + dz**2 ) ##only want component in x direction (all others cancel) rhat_dot_xhat = abs(dx)/dist return (alpha*G*(rhoa-rhosi)/dist**2)*math.exp(-dist/lam)*(1.0 + dist/lam)*rhob*dV(phi,theta,r)*rhat_dot_xhat def sampler(): while True: r = random.uniform(0.,D/2.) theta = random.uniform(0.,2.*math.pi) phi = random.uniform(0.,math.pi) x = random.uniform(-a_depth/2.0+au_thick,a_depth/2.0+au_thick) y = random.uniform(-a/2., a/2.) z = random.uniform(-a/2., a/2.) yield (r, theta, phi, x, y, z) nmc = 100000000 domainsize = D * math.pi**2 * a_depth * a**2 random.seed(1) result, error = mcint.integrate(integrand, sampler(), measure=domainsize, n=nmc) print("integral is: ", result, error) #fname = 'data/lam_arr_%.3f_%.3f.npy' % (gap*1e6,lam*1e6) #np.save(fname,intval)
1,977
834
import logging from typing import ( Dict, Awaitable, Callable, Any, Set, List, Optional, TYPE_CHECKING) from opentrons.types import Mount, Point, Location from opentrons.config import feature_flags as ff from opentrons.hardware_control import ThreadManager, CriticalPoint from opentrons.protocol_api import labware from opentrons.protocols.geometry import deck from robot_server.robot.calibration import util from robot_server.service.errors import RobotServerError from robot_server.service.session.models.command import CalibrationCommand from ..errors import CalibrationError from ..helper_classes import RequiredLabware, AttachedPipette from ..constants import ( TIP_RACK_LOOKUP_BY_MAX_VOL, SHORT_TRASH_DECK, STANDARD_DECK, CAL_BLOCK_SETUP_BY_MOUNT, MOVE_TO_TIP_RACK_SAFETY_BUFFER, ) from .constants import TipCalibrationState as State, TIP_RACK_SLOT from .state_machine import TipCalibrationStateMachine if TYPE_CHECKING: from opentrons_shared_data.labware import LabwareDefinition MODULE_LOG = logging.getLogger(__name__) """ A collection of functions that allow a consumer to prepare and update calibration data associated with the combination of a pipette tip type and a unique (by serial number) physical pipette. """ # TODO: BC 2020-07-08: type all command logic here with actual Model type COMMAND_HANDLER = Callable[..., Awaitable] COMMAND_MAP = Dict[str, COMMAND_HANDLER] class TipCalibrationUserFlow: def __init__(self, hardware: ThreadManager, mount: Mount, has_calibration_block: bool, tip_rack: 'LabwareDefinition'): self._hardware = hardware self._mount = mount self._has_calibration_block = has_calibration_block self._hw_pipette = self._hardware._attached_instruments[mount] if not self._hw_pipette: raise RobotServerError( definition=CalibrationError.NO_PIPETTE_ON_MOUNT, mount=mount) self._tip_origin_pt: Optional[Point] = None self._nozzle_height_at_reference: Optional[float] = None deck_load_name = SHORT_TRASH_DECK if ff.short_fixed_trash() \ else STANDARD_DECK self._deck = deck.Deck(load_name=deck_load_name) self._tip_rack = self._get_tip_rack_lw(tip_rack) self._initialize_deck() self._current_state = State.sessionStarted self._state_machine = TipCalibrationStateMachine() self._command_map: COMMAND_MAP = { CalibrationCommand.load_labware: self.load_labware, CalibrationCommand.jog: self.jog, CalibrationCommand.pick_up_tip: self.pick_up_tip, CalibrationCommand.invalidate_tip: self.invalidate_tip, CalibrationCommand.save_offset: self.save_offset, CalibrationCommand.move_to_reference_point: self.move_to_reference_point, # noqa: E501 CalibrationCommand.move_to_tip_rack: self.move_to_tip_rack, # noqa: E501 CalibrationCommand.exit: self.exit_session, } def _set_current_state(self, to_state: State): self._current_state = to_state @property def current_state(self) -> State: return self._current_state def get_pipette(self) -> AttachedPipette: # TODO(mc, 2020-09-17): s/tip_length/tipLength return AttachedPipette( # type: ignore[call-arg] model=self._hw_pipette.model, name=self._hw_pipette.name, tip_length=self._hw_pipette.config.tip_length, mount=str(self._mount), serial=self._hw_pipette.pipette_id ) def get_required_labware(self) -> List[RequiredLabware]: slots = self._deck.get_non_fixture_slots() lw_by_slot = {s: self._deck[s] for s in slots if self._deck[s]} return [ RequiredLabware.from_lw(lw, s) # type: ignore for s, lw in lw_by_slot.items()] async def handle_command(self, name: Any, data: Dict[Any, Any]): """ Handle a client command :param name: Name of the command :param data: Data supplied in command :return: None """ next_state = self._state_machine.get_next_state(self._current_state, name) handler = self._command_map.get(name) if handler is not None: await handler(**data) self._set_current_state(next_state) MODULE_LOG.debug(f'TipCalUserFlow handled command {name}, transitioned' f'from {self._current_state} to {next_state}') async def load_labware(self): pass async def move_to_tip_rack(self): # point safely above target tip well in tip rack pt_above_well = self._tip_rack.wells()[0].top().point + \ MOVE_TO_TIP_RACK_SAFETY_BUFFER if self._tip_origin_pt is not None: # use jogged to x and y offsets only if returning tip to rack await self._move(Location(Point(self._tip_origin_pt.x, self._tip_origin_pt.y, pt_above_well.z), None)) else: await self._move(Location(pt_above_well, None)) async def save_offset(self): if self._current_state == State.measuringNozzleOffset: # critical point would default to nozzle for z height cur_pt = await self._get_current_point( critical_point=None) self._nozzle_height_at_reference = cur_pt.z elif self._current_state == State.measuringTipOffset: assert self._hw_pipette.has_tip assert self._nozzle_height_at_reference is not None # set critical point explicitly to nozzle cur_pt = await self._get_current_point( critical_point=CriticalPoint.NOZZLE) util.save_tip_length_calibration( pipette_id=self._hw_pipette.pipette_id, tip_length_offset=cur_pt.z - self._nozzle_height_at_reference, tip_rack=self._tip_rack) def _get_default_tip_length(self) -> float: tiprack: labware.Labware = self._deck[TIP_RACK_SLOT] # type: ignore full_length = tiprack.tip_length overlap_dict: Dict = \ self._hw_pipette.config.tip_overlap default = overlap_dict['default'] overlap = overlap_dict.get(tiprack.uri, default) return full_length - overlap def _get_critical_point_override(self) -> Optional[CriticalPoint]: return (CriticalPoint.FRONT_NOZZLE if self._hw_pipette.config.channels == 8 else None) async def _get_current_point( self, critical_point: CriticalPoint = None) -> Point: return await self._hardware.gantry_position(self._mount, critical_point) async def jog(self, vector): await self._hardware.move_rel(mount=self._mount, delta=Point(*vector)) async def move_to_reference_point(self): ref_loc = util.get_reference_location( mount=self._mount, deck=self._deck, has_calibration_block=self._has_calibration_block) await self._move(ref_loc) async def pick_up_tip(self): await util.pick_up_tip(self, tip_length=self._get_default_tip_length()) async def invalidate_tip(self): await util.invalidate_tip(self) async def exit_session(self): await self.move_to_tip_rack() await self._return_tip() def _get_tip_rack_lw(self, tip_rack_def: 'LabwareDefinition') -> labware.Labware: try: return labware.load_from_definition( tip_rack_def, self._deck.position_for(TIP_RACK_SLOT)) except Exception: raise RobotServerError(definition=CalibrationError.BAD_LABWARE_DEF) def _get_alt_tip_racks(self) -> Set[str]: pip_vol = self._hw_pipette.config.max_volume return set(TIP_RACK_LOOKUP_BY_MAX_VOL[str(pip_vol)].alternatives) def _initialize_deck(self): self._deck[TIP_RACK_SLOT] = self._tip_rack if self._has_calibration_block: cb_setup = CAL_BLOCK_SETUP_BY_MOUNT[self._mount] self._deck[cb_setup.slot] = labware.load( cb_setup.load_name, self._deck.position_for(cb_setup.slot)) async def _return_tip(self): await util.return_tip(self, tip_length=self._get_default_tip_length()) async def _move(self, to_loc: Location): await util.move(self, to_loc)
8,823
2,697
'''Autogenerated by xml_generate script, do not edit!''' from OpenGL import platform as _p, arrays # Code generation uses this from OpenGL.raw.WGL import _types as _cs # End users want this... from OpenGL.raw.WGL._types import * from OpenGL.raw.WGL import _errors from OpenGL.constant import Constant as _C import ctypes _EXTENSION_NAME = 'WGL_ARB_pbuffer' def _f( function ): return _p.createFunction( function,_p.PLATFORM.WGL,'WGL_ARB_pbuffer',error_checker=_errors._error_checker) WGL_DRAW_TO_PBUFFER_ARB=_C('WGL_DRAW_TO_PBUFFER_ARB',0x202D) WGL_MAX_PBUFFER_HEIGHT_ARB=_C('WGL_MAX_PBUFFER_HEIGHT_ARB',0x2030) WGL_MAX_PBUFFER_PIXELS_ARB=_C('WGL_MAX_PBUFFER_PIXELS_ARB',0x202E) WGL_MAX_PBUFFER_WIDTH_ARB=_C('WGL_MAX_PBUFFER_WIDTH_ARB',0x202F) WGL_PBUFFER_HEIGHT_ARB=_C('WGL_PBUFFER_HEIGHT_ARB',0x2035) WGL_PBUFFER_LARGEST_ARB=_C('WGL_PBUFFER_LARGEST_ARB',0x2033) WGL_PBUFFER_LOST_ARB=_C('WGL_PBUFFER_LOST_ARB',0x2036) WGL_PBUFFER_WIDTH_ARB=_C('WGL_PBUFFER_WIDTH_ARB',0x2034) @_f @_p.types(_cs.HPBUFFERARB,_cs.HDC,_cs.c_int,_cs.c_int,_cs.c_int,ctypes.POINTER(_cs.c_int)) def wglCreatePbufferARB(hDC,iPixelFormat,iWidth,iHeight,piAttribList):pass @_f @_p.types(_cs.BOOL,_cs.HPBUFFERARB) def wglDestroyPbufferARB(hPbuffer):pass @_f @_p.types(_cs.HDC,_cs.HPBUFFERARB) def wglGetPbufferDCARB(hPbuffer):pass @_f @_p.types(_cs.BOOL,_cs.HPBUFFERARB,_cs.c_int,ctypes.POINTER(_cs.c_int)) def wglQueryPbufferARB(hPbuffer,iAttribute,piValue):pass @_f @_p.types(_cs.c_int,_cs.HPBUFFERARB,_cs.HDC) def wglReleasePbufferDCARB(hPbuffer,hDC):pass
1,572
789
from collections import namedtuple import inspect from django import forms FormFieldWarning = namedtuple("FormFieldWarning", ["message", "description"]) class WarningFormMixin: """Classes using WarningFormMixin should implement methods to catch warnings >>> def warning_mailboxes(self) -> List[FormFieldWarning]: if some condition: return FormFieldWarning(message, description) """ def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.fields["ignore_warnings"] = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput() ) self.warnings = [] def get_warnings(self): self.warnings = [] if self.cleaned_data.get("ignore_warnings"): return [] else: warning_methods = [ (method_name, method) for method_name, method in inspect.getmembers( self, predicate=inspect.ismethod ) if method_name.startswith("warning_") ] for method_name, method in warning_methods: warnings = method() if warnings: self.warnings.extend( [ { "field": method_name.split("warning_")[1], "message": warning.message, "description": warning.description, } for warning in warnings ] ) self.data._mutable = True self.data["ignore_warnings"] = True self.data._mutable = False return self.warnings def is_valid(self): return super().is_valid() and not self.get_warnings()
1,885
445
print("Hello, World! Again!")
29
10
from numpy import array, arange, argmax from numpy.random import choice from itertools import product from ai.NN import NN import pickle from random import random, randrange, randint from collections import deque from math import floor class AI: MEMORY_DIMENSION = 90000 def __init__(self, load: bool = False, state_size: int = 0, rods_number: int = 0, offset=None, angle_velocity=None, hidden_layers=(100, 100), log_size=10, nn_file: str = "save.model", actions_file: str = "save.actions"): """ :param load: specify if model should be load :param state_size: how many attributes has a state :param rods_number: number of rods of a player :param offset: represents a vector with scalars that will move rod to left/right :param angle_velocity: represents a vector with rates of change of a rodsman angle :param hidden_layers: a vector of values that represents how many units have each layer :param nn_file: file to save neural network :param actions_file: file to save actions_file """ self.actions = None self.model = None self.last_predictions = deque(maxlen=2 * log_size) self.last_states = deque(maxlen=2 * log_size) self.last_actions_index = deque(maxlen=2 * log_size) self.last_reward_sums = deque([0, 0], maxlen=2 * log_size + 2) self.log_size = log_size self.lamda = 0.6 self.alpha = 0.9 self.epsilon = 0.5 # greedy policy self.__epsilon_backup = self.epsilon # decreasing_rate will decrease epsilon such that in the future, when nn learned something # to not make anymore random choices self.__decreasing_rate = 0.99997 # memory replay self.memory_state = deque(maxlen=AI.MEMORY_DIMENSION) self.memory_target = deque(maxlen=AI.MEMORY_DIMENSION) # with save_probability save a memory with consist of a state and a target self.save_probability = 0.3 if load: self.__load(nn_file, actions_file) return self.rods_number = rods_number self.actions = array( [action for action in product(arange(rods_number), offset, angle_velocity)] ) self.batch_size = int(floor((2 * log_size) ** 0.5)) self.model = NN(input_dim=state_size, hidden_layers=hidden_layers, output_dim=len(self.actions), batch_size=self.batch_size) self.model.compile() def __load(self, nn_file, actions_file): self.model = NN(load_file=nn_file) fd = open(actions_file, "rb") self.rods_number, self.actions, self.epsilon, self.lamda, self.batch_size = pickle.load(fd) self.model.batch_size = self.batch_size fd.close() def save(self, nn_file: str="save.model", actions_file: str="save.actions"): self.model.save(nn_file) print("saving ai...") fd = open(actions_file, "wb") to_save = (self.rods_number, self.actions, self.epsilon, self.lamda, self.batch_size) pickle.dump(to_save, fd, protocol=0) # protocol 0 for compatibility fd.close() def __compute_and_backup(self, state): self.last_predictions.append(self.model.predict_action(state)) self.last_states.append(state) # noinspection PyMethodMayBeStatic def one_action(self, q_values): return [argmax(q_values)] def multiple_actions(self, q_values): actions_idxs = [] slice_size = int(len(self.actions) // self.rods_number) for i in range(self.rods_number): actions_idxs.append(i * slice_size + argmax(q_values[i * slice_size:(i + 1) * slice_size])) return actions_idxs def get_action(self, state, action_selector): """ :param state: a state of the current game :param action_selector: may be one of the following functions: ane_action, multiple_actions :return: """ self.__compute_and_backup(state) self.last_actions_index.append(action_selector(self.last_predictions[-1])) return [self.actions[i] for i in self.last_actions_index[-1]] def one_action_off_policy(self, rand, q_values): if rand: return [randrange(0, len(self.actions))] else: return self.one_action(q_values) def multiple_actions_off_policy(self, rand, q_values): slice_size = int(len(self.actions) // self.rods_number) if rand: return [i * slice_size + randrange(0, slice_size) for i in range(self.rods_number)] else: return self.multiple_actions(q_values) def get_action_off_policy(self, state, action_selector): self.__compute_and_backup(state) if random() < self.epsilon: # should choose an action random self.epsilon *= self.__decreasing_rate self.last_actions_index.append(action_selector(True, None)) return [self.actions[i] for i in self.last_actions_index[-1]] self.last_actions_index.append(action_selector(False, self.last_predictions[-1])) return [self.actions[i] for i in self.last_actions_index[-1]] def update(self, action_based_reward, new_states): assert len(action_based_reward) == len(new_states), "must have reward for each new_state" assert len(action_based_reward) == 2, "exactly 2 players supported ATM" for idx in range(0, len(self.last_reward_sums), 2): self.last_reward_sums[idx] += action_based_reward[0] self.last_reward_sums[idx + 1] += action_based_reward[1] self.last_reward_sums.extend(action_based_reward) q_values = [[self.last_predictions[i][j] for j in self.last_actions_index[i]] for i in range(len(self.last_actions_index))] action_selector = self.one_action if len(q_values[0]) == 1 else self.multiple_actions for i in range(len(q_values) - 2): next_q_values = q_values[i + 2] q_values_updated = [ # TODO: not sure about self.last_rewards[i] (1 - self.alpha) * q + self.alpha * (self.last_reward_sums[i] / ((len(q_values) + 1 - i) // 2) + (self.lamda * next_q)) for q, next_q in zip(q_values[i], next_q_values)] for j, update in zip(self.last_actions_index[i], q_values_updated): self.last_predictions[i][j] = update for i in range(len(new_states)): next_max_q_values = action_selector(self.model.predict_action(new_states[i])) q_values_updated = [ (1 - self.alpha) * q + self.alpha * (action_based_reward[i] + self.lamda * next_q) for q, next_q in zip(q_values[-2 + i], next_max_q_values)] for j, update in zip(self.last_actions_index[-2 + i], q_values_updated): self.last_predictions[-2 + i][j] = update if random() < self.save_probability: self.memory_state.append(self.last_states[-1]) self.memory_target.appendleft(self.last_predictions[-1]) if random() <= 0.5: self.model.update(self.last_states, self.last_predictions) else: self.from_memory_update() # we trust more in next move when network learn more self.lamda += self.lamda * 1.e-7 if self.lamda > 1: print("lambda = {}".format(self.lamda)) self.lamda = 1 def predict_action(self, state, action_selector): actions_idxs = action_selector(self.model.predict_action(state)) return [self.actions[i] for i in actions_idxs] def flush_last_actions(self): self.last_states.clear() self.last_actions_index.clear() self.last_predictions.clear() def switch_random_action(self, activate): if not activate: self.__epsilon_backup = self.epsilon self.epsilon = 0 else: self.epsilon = self.__epsilon_backup def from_memory_update(self): if len(self.memory_state) < 1000: return idxs = arange(len(self.memory_state)) size = randint(200, 1000) sample = choice(idxs, size, replace=False) self.model.update(array([self.memory_state[i] for i in sample]), array([self.memory_target[i] for i in sample]), False, 3)
8,791
2,721
print('=' * 12 + 'Desafio 49' + '=' * 12) numero = int(input('Digite o número para a tabuada: ')) print('=' * 13) print(f'Tabuada do {numero}') print('=' * 13) for i in range(1,11): print(f'{numero} x {i:2} = {numero * i}') print('=' * 13)
243
121
import pretty_midi import sys import numpy as np from tqdm import tqdm from collections import Counter sys.path.append('..') from Pre_Production.Midi_Pre_Processor import * from Shared_Files.Global_Util import * class MusicPallete: def __init__(self, pre_processor_obj): self.__all_possible_instr_note_pairs = pre_processor_obj.return_all_possible_instr_note_pairs() self.__centroid_instr_note = None self.__instr_note_pair_attributes = dict() set_test_eval = set() full_matrix = [] count = 0 for instr_note_pair_str in (self.__all_possible_instr_note_pairs): if "True" in instr_note_pair_str: count += 1 instr_note_pair_obj = convert_string_to_instr_note_pair(instr_note_pair_str) # Evaluate instr/note pair fluid_synth, synth, note_hz = extract_instr_note_pair_attributes(instr_note_pair_obj) # print(fluid_synth) # --- print(fluid_synth) if self.__centroid_instr_note is None: self.__centroid_instr_note = fluid_synth self.__instr_note_pair_attributes[ instr_note_pair_str] = float(calculate_wave_mse(self.__centroid_instr_note, fluid_synth)) set_test_eval.add(calculate_wave_mse(self.__centroid_instr_note, fluid_synth)) print(len(set_test_eval)) print(count) cnt = Counter([v for k,v in self.__instr_note_pair_attributes.items()]) print(cnt.most_common()[:20]) print() print([k for k,v in self.__instr_note_pair_attributes.items() if v in set([test[0] for test in cnt.most_common()[:20]])])
1,686
563
#!/usr/bin/env python # -*- coding: utf-8 -*- # # launches the unit testing # # Copyright (C) # Honda Research Institute Europe GmbH # Carl-Legien-Str. 30 # 63073 Offenbach/Main # Germany # # UNPUBLISHED PROPRIETARY MATERIAL. # ALL RIGHTS RESERVED. # # import os import tempfile import unittest from ToolBOSCore.Storage import SIT from ToolBOSCore.Storage import CopyTreeFilter from ToolBOSCore.Util import FastScript from ToolBOSCore.Util import Any class TestBootstrap( unittest.TestCase ): def setUp( self ): if not FastScript.getEnv( 'VERBOSE' ) == 'TRUE': Any.setDebugLevel( 1 ) def test_bootstrapSIT( self ): basePkgList = SIT.getMinRequirements() outputDir = tempfile.mkdtemp( prefix='test-' ) # create test SIT copyFilter = CopyTreeFilter.CopyTreeFilter( [] ) # skip copying binaries copyFilter.filterDocu = True SIT.bootstrap( outputDir, True, True, copyFilter.callback, False ) # check if all essential packages are available for package in basePkgList: Any.requireIsFile( os.path.join( outputDir, package, 'packageVar.cmake' ) ) # check for Module-Index directory Any.requireIsDir( os.path.join( outputDir, 'Modules', 'Index' ) ) # clean-up FastScript.remove( outputDir ) if __name__ == '__main__': unittest.main() # EOF
1,408
474
#!/usr/bin/env python # # Module for adding Google Calendar Functionality. # # by Peter Juett # References:https://developers.google.com/calendar/quickstart/python # # Copyright 2018 # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # import httplib2 import os from apiclient import discovery from oauth2client import client from oauth2client import tools from oauth2client.file import Storage import logging from logging.handlers import RotatingFileHandler import paho.mqtt.client as mqtt import DB import datetime import time import pytz import os from dateutil import parser SLEEP_TIME = 60 try: import argparse flags = argparse.ArgumentParser(parents=[tools.argparser]).parse_args() except ImportError: flags = None # If modifying these scopes, delete your previously saved credentials # at ~/.credentials/calendar-python-quickstart.json SCOPES = 'https://www.googleapis.com/auth/calendar.readonly' CLIENT_SECRET_FILE = 'client_secret.json' APPLICATION_NAME = 'Google Calendar API Python Quickstart' class Calendar(object): def __init__(self, main_app_log): if main_app_log is None: self.log_formatter = logging.Formatter('%(asctime)s %(levelname)s %(funcName)s(%(lineno)d) %(message)s') self.logFile = 'logs/GoogleCalendar.log' self.my_handler = RotatingFileHandler(self.logFile, mode='a', maxBytes=5*1024*1024, backupCount=1, encoding=None, delay=0) self.my_handler.setFormatter(self.log_formatter) self.my_handler.setLevel(logging.INFO) self.app_log = logging.getLogger('root') self.app_log.setLevel(logging.INFO) self.app_log.addHandler(self.my_handler) else: self.app_log = main_app_log self._db = DB.DB() self._db.load_settings() self.start_mosquito() self.process_loop() def on_connect(self, mosclient, userdata, flags, rc): self.app_log.info("Subscribing to topic: " + self._db.get_value("mostopic")) mosclient.subscribe(self._db.get_value("mostopic")) def on_disconnect(self, client, userdata, rc): if rc != 0: self.app_log.info("Unexpected disconnection") def on_publish(self, client, userdata, mid): self.app_log.info("on_publish - published " + str(mid)) def start_mosquito(self): self.mos_client = mqtt.Client() self.mos_client.on_connect = self.on_connect self.mos_client.on_disconnect = self.on_disconnect self.mos_client.on_publish = self.on_publish if len(self._db.get_value("mospassword"))>0: self.mos_client.username_pw_set(self._db.get_value("mosusername"),self._db.get_value("mospassword")) mos_broker_address = self._db.get_value("mosbrokeraddress") self.app_log.info("Connecting to: " + mos_broker_address) self.mos_client.connect(mos_broker_address, int(self._db.get_value("mosbrokerport")), 60) self.app_log.info("Connected") self.mos_client.loop_start() def broadcast_send(self, data_item, value): result = 0 mid = 0 if data_item is None: self.app_log.info("data_item is None") return if value is None: self.app_log.info("value is None") return self.app_log.info("publishing: " + data_item + " " + value) try: message = self._db.get_value("name") + "/" + data_item + "/" + value result, mid = self.mos_client.publish(self._db.get_value("mostopic"), message) if result == mqtt.MQTT_ERR_SUCCESS: self.app_log.info("published OK, Message ID = " + str(mid)) elif result == mqtt.MQTT_ERR_NO_CONN: self.app_log.info("publish failed, no connection") else: self.app_log.info("publish failed, result code = " + str(result)) except Exception as e: self.app_log.exception('Exception: %s', e) def process_loop(self): while(1): try: self.get() except Exception as e: self.app_log.exception('Exception: %s', e) finally: time.sleep(SLEEP_TIME) def get_credentials(self): """Gets valid user credentials from storage. If nothing has been stored, or if the stored credentials are invalid, the OAuth2 flow is completed to obtain the new credentials. Returns: Credentials, the obtained credential. """ home_dir = os.path.expanduser('~') credential_dir = os.path.join(home_dir, '.credentials') if not os.path.exists(credential_dir): os.makedirs(credential_dir) credential_path = os.path.join(credential_dir, 'calendar-python-quickstart.json') store = Storage(credential_path) credentials = store.get() if not credentials or credentials.invalid: flow = client.flow_from_clientsecrets(CLIENT_SECRET_FILE, SCOPES) flow.user_agent = APPLICATION_NAME if flags: credentials = tools.run_flow(flow, store, flags) else: # Needed only for compatibility with Python 2.6 credentials = tools.run(flow, store) print('Storing credentials to ' + credential_path) return credentials def get(self): """Shows basic usage of the Google Calendar API. Creates a Google Calendar API service object and outputs a list of the next 10 events on the user's calendar. """ credentials = self.get_credentials() http = credentials.authorize(httplib2.Http()) service = discovery.build('calendar', 'v3', http=http) now = datetime.datetime.utcnow().isoformat() + 'Z' # 'Z' indicates UTC time self.app_log.info('Getting the upcoming 5 events') eventsResult = service.events().list( calendarId='primary', timeMin=now, maxResults=5, singleEvents=True, orderBy='startTime').execute() events = eventsResult.get('items', []) x = 1 if not events: self.app_log.info('No upcoming events found.') for event in events: start = event['start'].get('dateTime', event['start'].get('date')) next_event = self.process_text(start, event['summary']) self.broadcast_send('Calendar' + str(x), next_event) if x == 1: next_two_events = next_event elif x == 2: next_two_events = next_two_events + ", " + next_event self.broadcast_send('Calendar1and2', next_two_events) x = x + 1 def process_text(self, start_time, summary): self.app_log.info(start_time) this_time = parser.parse(start_time) self.app_log.info(this_time.strftime("%a %b %d %Y %H:%M:%S --- %z")) tz_len = len(this_time.strftime("%z")) out_text = start_time[5:16].replace("T"," ") if tz_len > 0: tz = pytz.timezone(self._db.get_value("pytztimezone")) now = datetime.datetime.now(tz) else: now = datetime.datetime.now() #daily events can have no timezone info.... self.app_log.info (now.strftime("%a %b %d %Y %H:%M:%S --- %z")) today_date = now.strftime("%m-%d") tomorrow_date = (datetime.date.today() + datetime.timedelta(days=1)).strftime("%m-%d") delta = this_time - now #Replace with today or tomorrow, as applicable out_text = out_text.replace(today_date,"").replace(tomorrow_date,"Tomorrow") if (delta.days < 7): #if it is soon, then use the day of the week instead of date... out_text = out_text.replace(this_time.strftime("%m-%d"),this_time.strftime("%a")) out_text = out_text + " " + summary return out_text def run_program(main_app_log): Calendar(main_app_log) if __name__ == '__main__': run_program(None)
9,566
2,685
# -*- coding: utf-8 -*- """ For an example page see: http://www.jneurosci.org/content/23/10/4355.long#ref-list-1 Note that JNeuroscience seems to be part of a consortium with J Physiology, J Neurophysiology, PNS, etc perhaps check out HighWire Press? """ import requests from bs4 import BeautifulSoup class JNeuroscienceRef(object): def __init__(self): pass """ <li><span class="ref-label ref-label-empty"></span> <a class="rev-xref-ref" href="#xref-ref-40-1" id="ref-40" title="View reference in text">↵</a> <div class="cit ref-cit ref-other" id="cit-23.10.4355.40"> <div class="cit-metadata"><cite>Yoshimura N, Seki S, de Groat WC (<span class="cit-pub-date">2001b</span>) Nitric oxide modulates Ca<sup>2</sup><sup>+</sup> channels in dorsal root ganglion neurons innervating rat urinary bladder. <span class="cit-source">J Neurophysiol</span> <span class="cit-vol">86 </span>: <span class="cit-fpage">304</span>–311.</cite></div> <div class="cit-extra"><a class="cit-ref-sprinkles cit-ref-sprinkles-ijlinks" href="/cgi/ijlink?linkType=ABST&amp;journalCode=jn&amp;resid=86/1/304"> <span class="cit-reflinks-abstract">Abstract</span><span class="cit-sep cit-reflinks-variant-name-sep">/</span><span class="cit-reflinks-full-text"> <span class="free-full-text">FREE </span>Full Text</span></a></div> </div> </li> """ def get_references(url): r = requests.get(url) soup = BeautifulSoup(r.text) ref_section = soup.find('div',{'class':'section ref-list'}) ref_entries = ref_section.find_all('li') import pdb pdb.set_trace() pass
1,705
666
# coding: utf-8 from __future__ import annotations import re # noqa: F401 from datetime import date, datetime # noqa: F401 from typing import Any, Dict, List, Optional # noqa: F401 from pydantic import AnyUrl, BaseModel, EmailStr, validator # noqa: F401 class CreateMetric(BaseModel): """NOTE: This class is auto generated by OpenAPI Generator (https://openapi-generator.tech). Do not edit the class manually. CreateMetric - a model defined in OpenAPI processed_at: The processed_at of this CreateMetric [Optional]. event_at: The event_at of this CreateMetric [Optional]. machine_id: The machine_id of this CreateMetric. metric: The metric of this CreateMetric [Optional]. value: The value of this CreateMetric [Optional]. unit: The unit of this CreateMetric [Optional]. """ processed_at: Optional[datetime] = None event_at: Optional[datetime] = None machine_id: int metric: Optional[str] = None value: Optional[str] = None unit: Optional[str] = None CreateMetric.update_forward_refs()
1,089
327
from .la_initiator_bfm import *
32
15
#!/usr/bin/env python3 import h5py import numpy as np import argparse import os import seissolxdmf as sx import seissolxdmfwriter as sw # These 2 latter modules are on pypi (e.g. pip install seissolxdmf) def read_reshape2d(sx, dataname): """read seissol dataset and if there is only one time stamp create a second dimension of size 1""" myData = sx.ReadData(dataname) if len(myData.shape) == 1: myData = myData.reshape((1, myData.shape[0])) return myData def fuzzysort(arr, idx, dim=0, tol=1e-6): """ return indexes of sorted points robust to small perturbations of individual components. https://stackoverflow.com/questions/19072110/numpy-np-lexsort-with-fuzzy-tolerant-comparisons note that I added dim<arr.shape[0]-1 in some if statement (else it will crash sometimes) """ arrd = arr[dim] srtdidx = sorted(idx, key=arrd.__getitem__) i, ix = 0, srtdidx[0] for j, jx in enumerate(srtdidx[1:], start=1): if arrd[jx] - arrd[ix] >= tol: if j - i > 1 and dim < arr.shape[0] - 1: srtdidx[i:j] = fuzzysort(arr, srtdidx[i:j], dim + 1, tol) i, ix = j, jx if i != j and dim < arr.shape[0] - 1: srtdidx[i:] = fuzzysort(arr, srtdidx[i:], dim + 1, tol) return srtdidx def lookup_sorted_geom(geom): """return the indices to sort the geometry array by x, then y, then z and the associated inverse look-up table """ ind = fuzzysort(geom.T, list(range(0, geom.shape[0])), tol=1e-4) # generate inverse look-up table dic = {i: index for i, index in enumerate(ind)} ind_inv = np.zeros_like(ind) for k, v in dic.items(): ind_inv[v] = k return ind, ind_inv def read_geom_connect(sx): return sx.ReadGeometry(), sx.ReadConnect() def return_sorted_geom_connect(sx): """sort geom array and reindex connect array to match the new geom array""" geom, connect = read_geom_connect(sx) nv = geom.shape[0] try: import pymesh geom, connect, inf = pymesh.remove_duplicated_vertices_raw( geom, connect, tol=1e-4 ) print(f"removed {inf['num_vertex_merged']} duplicates out of {nv}") except ModuleNotFoundError: print("pymesh not found, trying trimesh...") import trimesh trimesh.tol.merge = 1e-4 mesh = trimesh.Trimesh(geom, connect) mesh.merge_vertices() geom = mesh.vertices connect = mesh.faces print(f"removed {nv-geom.shape[0]} duplicates out of {nv}") ind, ind_inv = lookup_sorted_geom(geom) geom = geom[ind, :] connect = np.array([ind_inv[x] for x in connect.flatten()]).reshape(connect.shape) # sort along line (then we can use multidim_intersect) connect = np.sort(connect, axis=1) return geom, connect def multidim_intersect(arr1, arr2): """find indexes of same triangles in 2 connect arrays (associated with the same geom array) generate 1D arrays of tuples and use numpy function https://stackoverflow.com/questions/9269681/intersection-of-2d-numpy-ndarrays """ arr1_view = arr1.view([("", arr1.dtype)] * arr1.shape[1]) arr2_view = arr2.view([("", arr2.dtype)] * arr2.shape[1]) intersected, ind1, ind2 = np.intersect1d(arr1_view, arr2_view, return_indices=True) ni, n1, n2 = intersected.shape[0], arr1.shape[0], arr2.shape[0] print( f"{ni} faces in common, n faces connect 1:{n1}, 2:{n2} (diff: {n1-ni}, {n2-ni})" ) return ind1, ind2 def same_geometry(sx1, sx2): geom1 = sx1.ReadGeometry() geom2 = sx2.ReadGeometry() if geom1.shape[0] != geom2.shape[0]: return False else: return np.all(np.isclose(geom1, geom2, rtol=1e-3, atol=1e-4)) parser = argparse.ArgumentParser( description="make difference between 2 (paraview) output files: f2-f1. \ The output must be from the same mesh, but the partionning may differ." ) parser.add_argument("xdmf_filename1", help="filename1") parser.add_argument("xdmf_filename2", help="filename2") parser.add_argument( "--idt", nargs="+", required=True, help="list of time step to differenciate (1st = 0); -1 = all", type=int, ) parser.add_argument( "--Data", nargs="+", required=True, metavar=("variable"), help="Data to differenciate (example SRs); all for all stored quantities", ) parser.add_argument( "--ratio", dest="ratio", default=False, action="store_true", help="compute relative ratio (f1-f2)/f1 instead of f2-f1", ) args = parser.parse_args() sx1 = sx.seissolxdmf(args.xdmf_filename1) sx2 = sx.seissolxdmf(args.xdmf_filename2) same_geom = same_geometry(sx1, sx2) if same_geom: print("same indexing detected, no need to reindex arrays") geom1, connect1 = read_geom_connect(sx1) geom2, connect2 = read_geom_connect(sx2) else: geom1, connect1 = return_sorted_geom_connect(sx1) geom2, connect2 = return_sorted_geom_connect(sx2) if not np.all(np.isclose(geom1, geom2, rtol=1e-3, atol=1e-4)): raise ValueError("geometry arrays differ") ind1, ind2 = multidim_intersect(connect1, connect2) connect1 = connect1[ind1, :] if args.idt[0] == -1: args.idt = list(range(0, sx1.ndt)) aData = [] if args.Data == ["all"]: variable_names = set() for elem in sx1.tree.iter(): if elem.tag == "Attribute": variable_names.add(elem.get("Name")) variable_names2 = set() for elem in sx2.tree.iter(): if elem.tag == "Attribute": variable_names2.add(elem.get("Name")) # return only variables in common variable_names = variable_names.intersection(variable_names2) for to_remove in ["partition", "locationFlag"]: if to_remove in variable_names: variable_names.remove(to_remove) else: variable_names = args.Data for dataname in variable_names: print(dataname) myData1 = read_reshape2d(sx1, dataname) myData2 = read_reshape2d(sx2, dataname) ndt = min(myData1.shape[0], myData2.shape[0]) if same_geom: myData = myData1[0:ndt, :] - myData2[0:ndt, :] if args.ratio: myData = myData / myData1[0:ndt, :] else: myData = myData1[0:ndt, ind1] - myData2[0:ndt, ind2] if args.ratio: myData = myData / myData1[0:ndt, ind1] for idt in args.idt: if idt < ndt: print(idt, np.amin(myData[idt, :]), np.amax(myData[idt, :])) else: print(f"removing idt={idt}>{ndt} from args.idt") args.idt.pop(idt) aData.append(myData) prefix, ext = os.path.splitext(args.xdmf_filename1) add2prefix = "ratio" if args.ratio else "diff" fname = f"{add2prefix}_{os.path.basename(prefix)}" try: dt = sx1.ReadTimeStep() except NameError: dt = 0.0 out_names = ["diff_" + name for name in variable_names] sw.write_seissol_output(fname, geom1, connect1, out_names, aData, dt, args.idt)
6,940
2,614
from django.urls import path from . import views urlpatterns = [ path('', views.home_P, name='home_P') ]
110
39
def count_bit_sum(N): bit = [0] * max_bitlen if N <= 0: return bit[:] for i in range(N.bit_length() - 1): bit[i] = 1 << (N.bit_length() - 2) bit[N.bit_length()-1] = N - (1 << (N.bit_length() - 1)) + 1 return [bit[i] + b for i, b in enumerate(count_bit_sum(N - (1 << (N.bit_length() - 1))))] max_bitlen = int(1<<31).bit_length() L, R = map(int, input().split()) ans = [(r - l) * 2 > (R - L + 1) for l, r in zip(count_bit_sum(L-1), count_bit_sum(R))] print(sum(x * (1 << i) for i, x in enumerate(ans)))
538
245
# Copyright 2012 Free Software Foundation, Inc. # # This file is part of GNU Radio # # GNU Radio is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3, or (at your option) # any later version. # # GNU Radio is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with GNU Radio; see the file COPYING. If not, write to # the Free Software Foundation, Inc., 51 Franklin Street, # Boston, MA 02110-1301, USA. # class ApiObject(): ''' Filter count variable if the filter design tool has to return multiple filter parameters in future e.g Cascaded Filters ''' def __init__(self, filtcount = 1): self.filtercount = filtcount self.restype = [''] * self.filtercount self.params = [''] * self.filtercount self.taps = [''] * self.filtercount ''' Updates params dictionary for the given filter number ''' def update_params(self, params, filtno): if (filtno <= self.filtercount): self.params[filtno - 1] = params ''' Updates filter type for the given filter number ''' def update_filttype(self, filttype, filtno): if (filtno <= self.filtercount): self.filttype[filtno - 1] = filttype ''' updates taps for the given filter number. taps will contain a list of coefficients in the case of fir design and (b,a) tuple in the case of iir design ''' def update_taps(self, taps, filtno): if (filtno <= self.filtercount): self.taps[filtno - 1] = taps ''' updates all of them in a single call ''' def update_all(self, filttype, params, taps, filtno): if (filtno <= self.filtercount): self.taps[filtno - 1] = taps self.params[filtno - 1] = params self.restype[filtno - 1] = filttype def get_filtercount(self): return self.filtercount def get_restype(self, filtno=1): if (filtno <= self.filtercount): return self.restype[filtno - 1] def get_params(self, filtno=1): if (filtno <= self.filtercount): return self.params[filtno - 1] def get_taps(self, filtno=1): if (filtno <= self.filtercount): return self.taps[filtno - 1]
2,639
832
# -*- coding: utf8 -*- # announcement # helper class for toprace # Alfredo Martin 2021 import pygame version = 'announcement.v.1.0.0' class Announcement: def __init__(self, screen=None, text='', time=5, color=(255, 0, 0)): """initiallizes the Announcement instance screen: pygame canvas instance text: str time: int (number of frames during which the announcement will be active color: color of the text""" _, _, sw, sh = screen.get_rect() self.t = 0 self.color = color self.time = time self.active = True style = pygame.font.SysFont('comicsans', 70) self.text = style.render(text, False, self.color) _, _, w, h = self.text.get_rect() text_x = (sw / 2) - (w / 2) text_y = (sh / 2) - (h / 2) self.pos = (text_x, text_y) def update(self, screen): """updates the announcements (whether they are active or not) and blits announcements in the screen screen: pygame canvas instance returns: screen after being updated""" self.t += 1 self.active = self.t <= self.time screen.blit(self.text, self.pos) return screen if __name__ == '__main__': print(version)
1,262
411
""" Implement regularization policies here """ import torch import torch.nn as nn from policies.policy import PolicyBase import logging def build_reg_from_config(model, reg_config): """ This function build regularizer given the model (only need for weigths typically) and regularizer configuration. """ reg_class = reg_config['class'] reg_args = {k: v for k, v in reg_config.items() if k != 'class'} reg = globals()[reg_class](model, **reg_args) return reg def build_regs_from_config(model, config): """ This function takes *general* config file for current run and model and returns a list of regularizers which are build by build_reg_from_config. Example config.yaml for pruner instances: >>> regularizers: >>> reg1: >>> class: Hoyer # regularization method to use >>> lambda: 1e-6 # regularization coefficient >>> modules: [net.0] # modules to apply regularization >>> weight_only: True # if regularizer is applied only to weights of module (no bias) >>> reg2: >>> class: HoyerSquare >>> lambda: 1e-6 >>> modules: [net.2] >>> weight_only: True """ if 'regularizers' not in config: return [] regs_config = config['regularizers'] regs = [build_reg_from_config(model, reg_config) for reg_config in regs_config.values()] return regs class Regularizer(PolicyBase): def __init__(self, model, **kwargs): self._model = model if not isinstance(self._model, nn.Module): raise ValueError('model should be an instance of nn.Module') modules_dict = dict(self._model.named_modules()) self._weight_only, self._lambda = kwargs['weight_only'], float(kwargs['lambda']) prefix = '' if isinstance(self._model, torch.nn.DataParallel): prefix = 'module.' self._module_names = [prefix + _name for _name in kwargs['modules']] self._modules = [ modules_dict[module_name] for module_name in self._module_names ] logging.debug(f'Constructed {self.__class__.__name__} with config:') logging.debug('\n'.join([f' -{k}:{v}' for k,v in kwargs.items()]) + '\n') def on_minibatch_begin(self, **kwargs): reg_loss = self._lambda * self._compute_penalty() return reg_loss def _compute_penalty(self): """ Base regularizer method which computes regularization penalty. """ raise ValueError('Implement in a child class') class Hoyer(Regularizer): def __init__(self, model, **kwargs): super(Hoyer, self).__init__(model, **kwargs) def _compute_penalty(self): penalty = 0. for module in self._modules: dim = module.weight.numel() vector = module.weight.view(-1) if not self._weight_only and module.bias: dim += self.module.bias.numel() vector = torch.cat([vector, module.bias]) penalty += (dim ** 0.5 - vector.abs().sum() / vector.norm()) / (dim ** 0.5 - 1) return penalty class SquaredHoyer(Regularizer): def __init__(self, model, **kwargs): super(SquaredHoyer, self).__init__(model, **kwargs) def _compute_penalty(self): penalty = 0. for module in self._modules: vector = module.weight.view(-1) if not self._weight_only and module.bias: vector = torch.cat([vector, module.bias]) penalty += vector.abs().sum().pow(2) / vector.norm().pow(2) return penalty if __name__ == '__main__': pass
3,633
1,095
from django.db import models from django.utils import timezone from stores.models import Store class Coupon(models.Model): SCOPE_CHOICES = [ ('I', 'Toàn bộ hoá đơn'), ('P', 'Sản phẩm'), ] BENEFICIARY_CHOICES = [ ('E', 'Mọi khách hàng'), ('N', 'Khách hàng hệ thống'), ('V', 'Khách hàng VIP') ] title = models.CharField(max_length=100, blank=False, null=False) scope = models.CharField(max_length=1, choices=SCOPE_CHOICES, blank=False, null=False) beneficiary = models.IntegerField(choices=BENEFICIARY_CHOICES, default='E') discount_percent = models.FloatField(blank=True) max_discount_money = models.PositiveIntegerField(blank=True) app_date = models.DateTimeField(default=timezone.now) exp_date = models.DateTimeField(blank=False, null=False) note = models.TextField(default='') store = models.ForeignKey(to=Store, on_delete=models.CASCADE)
938
346
#!/usr/bin/env python # encoding: utf-8 """ @author: zhanghe @software: PyCharm @file: inventory.py @time: 2018-06-28 00:22 """ from __future__ import unicode_literals from flask import jsonify, make_response from flask_restful import Resource, marshal, reqparse from web_api.bearings.outputs.inventory import fields_item_inventory from web_api.bearings.reqparsers.inventory import ( request_parser, request_parser_item_post, request_parser_item_put, structure_key_item, structure_key_items, ) from web_api.commons.exceptions import BadRequest, NotFound from web_api.bearings.apis.inventory import ( get_inventory_row_by_id, edit_inventory, delete_inventory, get_inventory_limit_rows_by_last_id, add_inventory, get_inventory_pagination, ) from web_api.commons.http_token_auth import token_auth from web_api import app SUCCESS_MSG = app.config['SUCCESS_MSG'] FAILURE_MSG = app.config['FAILURE_MSG'] class InventoryResource(Resource): """ InventoryResource """ decorators = [token_auth.login_required] def get(self, pk): """ Example: curl http://0.0.0.0:5000/bearings/inventory/1 :param pk: :return: """ data = get_inventory_row_by_id(pk) if not data: raise NotFound result = marshal(data, fields_item_inventory, envelope=structure_key_item) return jsonify(result) def delete(self, pk): """ Example: curl http://0.0.0.0:5000/bearings/inventory/1 -X DELETE :param pk: :return: """ result = delete_inventory(pk) if result: success_msg = SUCCESS_MSG.copy() return make_response(jsonify(success_msg), 204) else: failure_msg = FAILURE_MSG.copy() return make_response(jsonify(failure_msg), 400) def put(self, pk): """ Example: curl http://0.0.0.0:5000/bearings/inventory/1 -H "Content-Type: application/json" -X PUT -d ' { "inventory": { "product_id": 1, "warehouse_id": 1, "rack_id": 1, "stock_qty": 101 } }' :param pk: :return: """ request_args = request_parser.parse_args() request_item_args = request_parser_item_put.parse_args(req=request_args) if not request_item_args: raise BadRequest('Bad request.') request_data = request_item_args result = edit_inventory(pk, request_data) if result: success_msg = SUCCESS_MSG.copy() return make_response(jsonify(success_msg), 200) else: failure_msg = FAILURE_MSG.copy() return make_response(jsonify(failure_msg), 400) class InventoryListResource(Resource): """ InventoryListResource """ decorators = [token_auth.login_required] def get(self): """ Example: curl http://0.0.0.0:5000/bearings/inventories curl http://0.0.0.0:5000/bearings/inventories?last_pk=2&limit_num=2 :return: """ # 条件参数 filter_parser = reqparse.RequestParser(bundle_errors=True) filter_parser.add_argument('last_pk', type=int, default=0, location='args') filter_parser.add_argument('limit_num', type=int, default=20, location='args') filter_parser_args = filter_parser.parse_args() data = get_inventory_limit_rows_by_last_id(**filter_parser_args) result = marshal(data, fields_item_inventory, envelope=structure_key_items) return jsonify(result) def post(self): """ Example: curl http://0.0.0.0:5000/bearings/inventories -H "Content-Type: application/json" -X POST -d ' { "inventory": { "product_id": 1, "warehouse_id": 1, "rack_id": "1", "stock_qty": 100, "note": "SNFA" } }' :return: """ request_args = request_parser.parse_args() request_item_args = request_parser_item_post.parse_args(req=request_args) if not request_item_args: raise BadRequest('Bad request.') request_data = request_item_args result = add_inventory(request_data) if result: SUCCESS_MSG['id'] = result return make_response(jsonify(SUCCESS_MSG), 201) raise NotFound class InventoryPaginationResource(Resource): """ InventoryPaginationResource """ decorators = [token_auth.login_required] def get(self): """ Example: curl http://0.0.0.0:5000/bearings/inventories/pagination curl http://0.0.0.0:5000/bearings/inventories/pagination?page=2&per_page=2 :return: """ # 条件参数 filter_parser = reqparse.RequestParser(bundle_errors=True) filter_parser.add_argument('page', type=int, default=1, location='args') filter_parser.add_argument('per_page', type=int, default=20, location='args') filter_parser_args = filter_parser.parse_args() pagination_obj = get_inventory_pagination(**filter_parser_args) result = marshal(pagination_obj.items, fields_item_inventory, envelope=structure_key_items) result['total'] = pagination_obj.total return jsonify(result)
5,527
1,801
#!/usr/bin/env python # from bob.bio.spear.database import AudioBioFile import bob.bio.base import bob.io.base import bob.io.video from bob.extension import rc from .common import SwanVideoFile, SwanAudioFile, SwanVideoDatabase # class SwanAudioBioFile(SwanAudioFile, AudioBioFile): # """SwanAudioBioFile are video files actually""" class SwanVideoBioFile(SwanVideoFile): """SwanVideoBioFile are video files actually""" class Database(bob.bio.base.database.FileListBioDatabase, SwanVideoDatabase): """Wrapper class for the SWAN database for speaker recognition (http://www.idiap.ch/dataset/swan). This class defines a simple protocol for training, dev and and by splitting the audio files of the database in three main parts. """ def __init__(self, original_directory=rc['bob.db.swan.directory'], bio_file_class=SwanVideoBioFile, annotation_directory=rc['bob.db.swan.annotation_dir'], annotation_extension='.json', annotation_type='json', name='swan', **kwargs): # call base class constructor from pkg_resources import resource_filename folder = resource_filename(__name__, 'lists') super(Database, self).__init__( folder, name=name, bio_file_class=bio_file_class, annotation_directory=annotation_directory, annotation_extension=annotation_extension, annotation_type=annotation_type, original_directory=original_directory, training_depends_on_protocol=False, models_depend_on_protocol=True, **kwargs ) def objects(self, groups=None, protocol=None, purposes=None, model_ids=None, classes=None, filter_samples=None, **kwargs): files = super(Database, self).objects( groups=groups, protocol=protocol, purposes=purposes, model_ids=model_ids, classes=classes, **kwargs) files = self.update_files(files) if filter_samples is None: return files files = list(filter(filter_samples, files)) return files
2,141
585
from tokenizer import StringTokenizer from token import tokentype text = open('test.ex', 'r').read() t = StringTokenizer(text=text, tokentype=tokentype) token_generator = t.create_token_generator() print(token_generator)
226
76
#!/usr/bin/smilx '''========================================================================= The Software is copyright (c) Commonwealth Scientific and Industrial Research Organisation (CSIRO) ABN 41 687 119 230. All rights reserved. Licensed under the CSIRO BSD 3-Clause License You may not use this file except in compliance with the License. You may obtain a copy of the License in the file LICENSE.md or at https://stash.csiro.au/projects/SMILI/repos/smili/browse/license.txt Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. =========================================================================''' ''' This script opens a pv map and its corresponding image, thresholds and overlays the two, loads camera view, links windows and disables interpolation. To use it, run smilx, in python console type "execfile('smilx_pv_visualisation.py')" ''' #names of images, set them to empty string to get popup dialogs imageName = "PAVEL_FLAWS_INV2.nii.gz" pvImageName = "PAVEL_BiExp_Combined_PV_GM.nii.gz" tmpOutName = "thresholded_pvimage.nii.gz" MainWindow.loadFile(imageName) # load image image = MainWindow.activeImage() # get pointer to image window MainWindow.loadFile(pvImageName) # load image pvImage = MainWindow.activeImage() # get pointer to image window #process pv map for display belowLevel = 0.01 aboveLevel = 0.25 pvImage.binaryThreshold(255, belowLevel, aboveLevel) milxQtFile.saveImage(tmpOutName, pvImage) # save result #overlay the pv processed result pvImage.viewToSagittal() pvImage.loadView("camera.cam") pvImage.interpolateDisplay() image.overlay(tmpOutName) image.viewToSagittal() image.loadView("camera.cam") image.interpolateDisplay() MainWindow.tileTab() MainWindow.link() #link all the windows
2,008
628
expected_output ={ "lentry_label": { 24: { "aal": { "deagg_vrf_id": 0, "eos0": {"adj_hdl": "0xf9000002", "hw_hdl": "0x7f02737e2ca8"}, "eos1": {"adj_hdl": "0xf9000002", "hw_hdl": "0x7f02737e2a98"}, "id": 1996488716, "lbl": 24, "lspa_handle": "0", }, "adj": { 109: { "adj": "0xdf000026", "ifnum": "0x33", "link_type": "MPLS", "si": "0x7f0273423ab8", }, 139: { "adj": "0x5c000037", "ifnum": "0x36", "link_type": "MPLS", "si": "0x7f02737a2348", }, }, "backwalk_cnt": 0, "label": { 31: { "adj_handle": "0x62000061", "bwalk_cnt": 0, "collapsed_oce": 0, "flags": {"0x1": ["REAL"]}, "label_aal": { 1644167265: { "adj_flags": "0", "di_id": "0x526d", "dmac": "0027.90bf.2ee7", "label_type": 2, "lbl": 0, "link_type": 2, "phdl": "0xab000447", "ref_cnt": 1, "rewrite_type": "PSH1(119)", "ri": "0x7f02737e8a98", "ri_id": "0x4e", "si": "0x7f02737c1b08", "si_id": "0x4034", "smac": "00a7.42d6.c41f", "sub_type": 0, "vlan_id": 0, "vrf_id": 0, } }, "link_type": "MPLS", "local_adj": 0, "local_label": 24, "modify_cnt": 0, "olbl_changed": 0, "outlabel": "(34, 0)", "pdflags": {"0": ["INSTALL_HW_OK"]}, "subwalk_cnt": 0, "unsupported_recursion": 0, }, 32: { "adj_handle": "0x89000062", "bwalk_cnt": 0, "collapsed_oce": 0, "flags": {"0x1": ["REAL"]}, "label_aal": { 2298478690: { "adj_flags": "0", "di_id": "0x5268", "dmac": "00a7.42ce.f69f", "label_type": 2, "lbl": 0, "link_type": 2, "phdl": "0x7c000442", "ref_cnt": 1, "rewrite_type": "PSH1(119)", "ri": "0x7f027379b138", "ri_id": "0x24", "si": "0x7f02737a4d58", "si_id": "0x4035", "smac": "00a7.42d6.c41f", "sub_type": 0, "vlan_id": 0, "vrf_id": 0, } }, "link_type": "MPLS", "local_adj": 0, "local_label": 24, "modify_cnt": 0, "olbl_changed": 0, "outlabel": "(29, 0)", "pdflags": {"0": ["INSTALL_HW_OK"]}, "subwalk_cnt": 0, "unsupported_recursion": 0, }, }, "lb": { 38: { "aal": { "af": 0, "ecr_id": 4177526786, "ecr_type": "0", "ecrh": "0x7f02737e49f8(28:2)", "hwhdl": ":1937656312 " "::0x7f02737e11c8,0x7f02737e2728,0x7f02737e11c8,0x7f02737e2728", "ref": 3, }, "bwalk": {"in_prog": 0, "nested": 0, "req": 0}, "bwalk_cnt": 0, "ecr_map_objid": 0, "ecrh": "0xf9000002", "finish_cnt": 0, "flags": "0", "link_type": "IP", "local_label": 24, "modify_cnt": 0, "mpls_ecr": 1, "num_choices": 2, "old_ecrh": "0", "path_inhw": 2, "subwalk_cnt": 0, } }, "lentry_hdl": "0x7700000c", "lspa_handle": "0", "modify_cnt": 8, "nobj": ["LB", " 38"], "sw_enh_ecr_scale": { 38: { "adjs": 2, "ecr_adj": { 1644167265: { "adj_lentry": "[eos0:0x7f02734123b8 " "eos1:0x7f02737ec5e8]", "di_id": 20499, "is_mpls_adj": 1, "l3adj_flags": "0x100000", "recirc_adj_id": 3120562239, "rih": "0x7f02737e0bf8(74)", "sih": "0x7f02737e11c8(182)", }, 2298478690: { "adj_lentry": "[eos0:0x7f02737e6dd8 " "eos1:0x7f02737b21d8]", "di_id": 20499, "is_mpls_adj": 1, "l3adj_flags": "0x100000", "recirc_adj_id": 1442840640, "rih": "0x7f02737dcbe8(75)", "sih": "0x7f02737e2728(183)", }, 2483028067: { "di_id": 20499, "rih": "0x7f02737eaa18(52)", "sih": "0x7f02737e4c08(170)", }, }, "ecr_hwhdl": "0x7f02737e49f8", "ecrhdl": "0xf9000002", "eos": 1, "llabel": 24, "mixed_adj": "0", "mod_cnt": 0, "pmismatch": 0, "pordermatch": 0, "prev_npath": 0, "reprogram_hw": "0", } }, } } }
6,887
2,421
#!/usr/bin/env python3 import nose.tools as nose from cachesimulator.bin_addr import BinaryAddress from cachesimulator.word_addr import WordAddress def test_get_bin_addr_unpadded(): """get_bin_addr should return unpadded binary address of word address""" nose.assert_equal( BinaryAddress(word_addr=WordAddress(180)), '10110100') def test_get_bin_addr_padded(): """get_bin_addr should return padded binary address of word address""" nose.assert_equal( BinaryAddress(word_addr=WordAddress(44), num_addr_bits=8), '00101100') def test_prettify_bin_addr_16_bit(): """prettify_bin_addr should prettify 8-bit string into groups of 3""" nose.assert_equal( BinaryAddress.prettify('1010101110101011', min_bits_per_group=3), '1010 1011 1010 1011') def test_prettify_bin_addr_8_bit(): """prettify_bin_addr should prettify 8-bit string into groups of 3""" nose.assert_equal( BinaryAddress.prettify('10101011', min_bits_per_group=3), '1010 1011') def test_prettify_bin_addr_7_bit(): """prettify_bin_addr should prettify 7-bit string into groups of 3""" nose.assert_equal( BinaryAddress.prettify('1011010', min_bits_per_group=3), '101 1010') def test_prettify_bin_addr_6_bit(): """prettify_bin_addr should prettify 6-bit string into groups of 3""" nose.assert_equal( BinaryAddress.prettify('101011', min_bits_per_group=3), '101 011') def test_prettify_bin_addr_5_bit(): """prettify_bin_addr should prettify 5-bit string into groups of 3""" nose.assert_equal( BinaryAddress.prettify('10110', min_bits_per_group=3), '10110') def test_get_tag_5_bit(): """get_tag should return correct 5 tag bits for an address""" nose.assert_equal( BinaryAddress('10110100').get_tag(num_tag_bits=5), '10110') def test_get_tag_0_bit(): """get_tag should return None if no bits are allocated to a tag""" nose.assert_is_none( BinaryAddress('10110100').get_tag(num_tag_bits=0)) def test_get_index_2_bit(): """get_index should return correct 2 index bits for an address""" nose.assert_equal( BinaryAddress('11111101').get_index( num_offset_bits=1, num_index_bits=2), '10') def test_get_index_0_bit(): """get_index should return None if no bits are allocated to an index""" nose.assert_is_none( BinaryAddress('11111111').get_index( num_offset_bits=1, num_index_bits=0)) def test_get_offset_2_bit(): """get_offset should return correct 2 offset bits for an address""" nose.assert_equal( BinaryAddress('11111101').get_offset(num_offset_bits=2), '01') def test_get_offset_0_bit(): """get_offset should return None if no bits are allocated to an offset""" nose.assert_is_none( BinaryAddress('10110100').get_offset(num_offset_bits=0)) def test_get_consecutive_words_1_word(): """get_consecutive_words should return same word for 1-word blocks""" nose.assert_equal( WordAddress(23).get_consecutive_words(num_words_per_block=1), [23]) def test_get_consecutive_words_2_word(): """get_consecutive_words should return correct words for 2-word blocks""" nose.assert_equal( WordAddress(22).get_consecutive_words(num_words_per_block=2), [22, 23]) def test_get_consecutive_words_4_word(): """get_consecutive_words should return correct words for 4-word blocks""" nose.assert_equal( WordAddress(21).get_consecutive_words(num_words_per_block=4), [20, 21, 22, 23])
3,610
1,363
class Vector(object,IFormattable): """ Represents a displacement in 2-D space. Vector(x: float,y: float) """ @staticmethod def Add(*__args): """ Add(vector: Vector,point: Point) -> Point Translates the specified point by the specified vector and returns the resulting point. vector: The amount to translate the specified point. point: The point to translate. Returns: The result of translating point by vector. Add(vector1: Vector,vector2: Vector) -> Vector Adds two vectors and returns the result as a System.Windows.Vector structure. vector1: The first vector to add. vector2: The second vector to add. Returns: The sum of vector1 and vector2. """ pass @staticmethod def AngleBetween(vector1,vector2): """ AngleBetween(vector1: Vector,vector2: Vector) -> float Retrieves the angle,expressed in degrees,between the two specified vectors. vector1: The first vector to evaluate. vector2: The second vector to evaluate. Returns: The angle,in degrees,between vector1 and vector2. """ pass @staticmethod def CrossProduct(vector1,vector2): """ CrossProduct(vector1: Vector,vector2: Vector) -> float Calculates the cross product of two vectors. vector1: The first vector to evaluate. vector2: The second vector to evaluate. Returns: The cross product of vector1 and vector2. The following formula is used to calculate the cross product: (Vector1.X * Vector2.Y) - (Vector1.Y * Vector2.X) """ pass @staticmethod def Determinant(vector1,vector2): """ Determinant(vector1: Vector,vector2: Vector) -> float Calculates the determinant of two vectors. vector1: The first vector to evaluate. vector2: The second vector to evaluate. Returns: The determinant of vector1 and vector2. """ pass @staticmethod def Divide(vector,scalar): """ Divide(vector: Vector,scalar: float) -> Vector Divides the specified vector by the specified scalar and returns the result as a System.Windows.Vector. vector: The vector structure to divide. scalar: The amount by which vector is divided. Returns: The result of dividing vector by scalar. """ pass @staticmethod def Equals(*__args): """ Equals(self: Vector,value: Vector) -> bool Compares two vectors for equality. value: The vector to compare with this vector. Returns: true if value has the same System.Windows.Vector.X and System.Windows.Vector.Y values as this vector; otherwise,false. Equals(self: Vector,o: object) -> bool Determines whether the specified System.Object is a System.Windows.Vector structure and,if it is,whether it has the same System.Windows.Vector.X and System.Windows.Vector.Y values as this vector. o: The vector to compare. Returns: true if o is a System.Windows.Vector and has the same System.Windows.Vector.X and System.Windows.Vector.Y values as this vector; otherwise,false. Equals(vector1: Vector,vector2: Vector) -> bool Compares the two specified vectors for equality. vector1: The first vector to compare. vector2: The second vector to compare. Returns: true if t he System.Windows.Vector.X and System.Windows.Vector.Y components of vector1 and vector2 are equal; otherwise,false. """ pass def GetHashCode(self): """ GetHashCode(self: Vector) -> int Returns the hash code for this vector. Returns: The hash code for this instance. """ pass @staticmethod def Multiply(*__args): """ Multiply(vector: Vector,matrix: Matrix) -> Vector Transforms the coordinate space of the specified vector using the specified System.Windows.Media.Matrix. vector: The vector structure to transform. matrix: The transformation to apply to vector. Returns: The result of transforming vector by matrix. Multiply(vector1: Vector,vector2: Vector) -> float Calculates the dot product of the two specified vectors and returns the result as a System.Double. vector1: The first vector to multiply. vector2: The second vector structure to multiply. Returns: A System.Double containing the scalar dot product of vector1 and vector2,which is calculated using the following formula: (vector1.X * vector2.X) + (vector1.Y * vector2.Y) Multiply(vector: Vector,scalar: float) -> Vector Multiplies the specified vector by the specified scalar and returns the resulting System.Windows.Vector. vector: The vector to multiply. scalar: The scalar to multiply. Returns: The result of multiplying vector and scalar. Multiply(scalar: float,vector: Vector) -> Vector Multiplies the specified scalar by the specified vector and returns the resulting System.Windows.Vector. scalar: The scalar to multiply. vector: The vector to multiply. Returns: The result of multiplying scalar and vector. """ pass def Negate(self): """ Negate(self: Vector) Negates this vector. The vector has the same magnitude as before,but its direction is now opposite. """ pass def Normalize(self): """ Normalize(self: Vector) Normalizes this vector. """ pass @staticmethod def Parse(source): """ Parse(source: str) -> Vector Converts a string representation of a vector into the equivalent System.Windows.Vector structure. source: The string representation of the vector. Returns: The equivalent System.Windows.Vector structure. """ pass @staticmethod def Subtract(vector1,vector2): """ Subtract(vector1: Vector,vector2: Vector) -> Vector Subtracts the specified vector from another specified vector. vector1: The vector from which vector2 is subtracted. vector2: The vector to subtract from vector1. Returns: The difference between vector1 and vector2. """ pass def ToString(self,provider=None): """ ToString(self: Vector,provider: IFormatProvider) -> str Returns the string representation of this System.Windows.Vector structure with the specified formatting information. provider: The culture-specific formatting information. Returns: A string that represents the System.Windows.Vector.X and System.Windows.Vector.Y values of this System.Windows.Vector. ToString(self: Vector) -> str Returns the string representation of this System.Windows.Vector structure. Returns: A string that represents the System.Windows.Vector.X and System.Windows.Vector.Y values of this System.Windows.Vector. """ pass def __add__(self,*args): """ x.__add__(y) <==> x+yx.__add__(y) <==> x+y """ pass def __div__(self,*args): """ x.__div__(y) <==> x/y """ pass def __eq__(self,*args): """ x.__eq__(y) <==> x==y """ pass def __format__(self,*args): """ __format__(formattable: IFormattable,format: str) -> str """ pass def __init__(self,*args): """ x.__init__(...) initializes x; see x.__class__.__doc__ for signaturex.__init__(...) initializes x; see x.__class__.__doc__ for signaturex.__init__(...) initializes x; see x.__class__.__doc__ for signature """ pass def __mul__(self,*args): """ x.__mul__(y) <==> x*yx.__mul__(y) <==> x*yx.__mul__(y) <==> x*y """ pass def __neg__(self,*args): """ x.__neg__() <==> -x """ pass @staticmethod def __new__(self,x,y): """ __new__(cls: type,x: float,y: float) __new__[Vector]() -> Vector """ pass def __ne__(self,*args): pass def __radd__(self,*args): """ __radd__(vector1: Vector,vector2: Vector) -> Vector Adds two vectors and returns the result as a vector. vector1: The first vector to add. vector2: The second vector to add. Returns: The sum of vector1 and vector2. """ pass def __reduce_ex__(self,*args): pass def __repr__(self,*args): """ __repr__(self: object) -> str """ pass def __rmul__(self,*args): """ __rmul__(vector1: Vector,vector2: Vector) -> float Calculates the dot product of the two specified vector structures and returns the result as a System.Double. vector1: The first vector to multiply. vector2: The second vector to multiply. Returns: Returns a System.Double containing the scalar dot product of vector1 and vector2,which is calculated using the following formula:vector1.X * vector2.X + vector1.Y * vector2.Y __rmul__(scalar: float,vector: Vector) -> Vector Multiplies the specified scalar by the specified vector and returns the resulting vector. scalar: The scalar to multiply. vector: The vector to multiply. Returns: The result of multiplying scalar and vector. """ pass def __rsub__(self,*args): """ __rsub__(vector1: Vector,vector2: Vector) -> Vector Subtracts one specified vector from another. vector1: The vector from which vector2 is subtracted. vector2: The vector to subtract from vector1. Returns: The difference between vector1 and vector2. """ pass def __str__(self,*args): pass def __sub__(self,*args): """ x.__sub__(y) <==> x-y """ pass Length=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets the length of this vector. Get: Length(self: Vector) -> float """ LengthSquared=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets the square of the length of this vector. Get: LengthSquared(self: Vector) -> float """ X=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets or sets the System.Windows.Vector.X component of this vector. Get: X(self: Vector) -> float Set: X(self: Vector)=value """ Y=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets or sets the System.Windows.Vector.Y component of this vector. Get: Y(self: Vector) -> float Set: Y(self: Vector)=value """
10,347
3,372
import random import math from collections import OrderedDict class BinaryEightQueensEnhancedNum: def __init__(self, populationSize, crossOverMethod=1, selectMethod=1, mutationMethod=1, entireFit=False): ''' inicializa a classe. @params populationSize o tamanho da população. ''' self.populationSize = populationSize self.population = [] self.crossOverMethod = self.crossOver if crossOverMethod == 2: self.crossOverMethod = self.crossOver2 self.selectionMethod = self.selectParents if selectMethod == 2: self.selectionMethod = self.selectRoulette self.mutate = self.mutate1 if mutationMethod == 2: self.mutate = self.mutate2 while len(self.population) < populationSize: gen = list(range(0,8)) random.shuffle(gen) if gen not in self.population: self.population.append(gen) self.history = [self.population] if not entireFit: self.checkSolution = self.checkSolution1 else: self.checkSolution = self.checkSolution2 def buildFenotype(self, gen): ''' Constroi o fenotipo a partir do genoma (posição das rainhas). @params gen string o gene a ser calculado. ''' return gen def mutate1(self, gen): # swap # todo: verificar se vale trocar o fitness mutationProbability = 0.4 for i in range(len(gen)): if random.random() < mutationProbability: a = random.randrange(len(gen)) gen[i] = a output = gen return output def mutate2(self, gen): # swap # todo: verificar se vale trocar o fitness mutationProbability = 0.4 if random.random() < mutationProbability: a = random.randrange(len(gen)) b = random.randrange(len(gen)) gen[a], gen[b] = gen[b], gen[a] output = gen return output def crossOver(self, gen1, gen2): pos = random.randrange(len(gen1)) child1 = gen1[:pos] + gen2[pos:] child2 = gen2[:pos] + gen1[pos:] return child1, child2 def crossOver2(self, gen1, gen2): pos = random.randrange(len(gen1)) child1 = gen1[:pos] + gen2[pos:] + gen2[:pos] child2 = gen2[:pos] + gen1[pos:] + gen1[:pos] # remove repetitions child1 = list(OrderedDict.fromkeys(child1)) child2 = list(OrderedDict.fromkeys(child2)) return child1, child2 def selectRoulette(self, population): fitnesses = [self.fitness(gen) for gen in population] total = sum(fitnesses) roulette = [] prevProb = 0 for fit in fitnesses: roulette.append(prevProb + fit/total) prevProb += fit/total prob1 = random.random() prob2 = random.random() prev = 0 parents = [] # print(prob1, prob2) for i in range(len(roulette)): if prev <= prob1 and prob1 <= roulette[i]: parents.append(population[i]) if prev <= prob2 and prob2 <= roulette[i]: parents.append(population[i]) prev = roulette[i] # print(parents) return parents[:2] def selectParents(self, population): gens = random.sample(population, 5) fenotypes = [(self.buildFenotype(gen), gen) for gen in gens] fitness = [(self.fitness(fen), gen) for fen, gen in fenotypes] fitness.sort(reverse=True) selectedGens = [gen for fitness, gen in fitness] return selectedGens[:2] def checkSolution1(self, population): found = False ans = [] for gen in population: if self.fitness(gen) == 28: found = True ans = gen break return found, ans def checkSolution2(self, population): found = True ans = [] for gen in population: if self.fitness(gen) != 28: found = False ans = gen break return found, ans def fitness(self, gen): maxHits = 8*(8-1)//2 fenotypes = self.buildFenotype(gen) hits = 0 for i in range(len(fenotypes)): for j in range(i): # print(fenotypes[i], fenotypes[j]) if fenotypes[i] == fenotypes[j]: hits += 1 elif abs(fenotypes[i] - fenotypes[j]) == abs(i - j): hits += 1 return maxHits - hits def fit(self): didFinish = False countGenerations = 0 populationFitness = [] convergentNumber = 0 for i in range(10000): found, gen = self.checkSolution(self.population) if found: print('alcançou a solução com ' + str(i) + ' iterações') # print(self.population) values = [(self.fitness(gen), self.buildFenotype(gen)) for gen in self.population if self.fitness(gen) == 28] # print(values) ############################## ## measurement of metrics ############################## countGenerations = i didFinish = True populationFitness = [self.fitness(gen) for gen in self.population] convergentNumber = len(values) ############################## break fitElements = [(self.fitness(gen), gen) for gen in self.population] fitElements.sort(reverse=True) if i%100 == 0: print('passo', i) # print(fitElements[0][0], self.buildFenotype(fitElements[0][1])) limit = math.floor(0.9*len(fitElements)) newPopulation = [] for i in range(limit, -1, -2): parents = self.selectionMethod(self.population) # aaa|bbbbbbb # bbb|aaaaaaa # print(parents) gens = self.crossOverMethod(parents[0], parents[1]) m1, m2 = self.mutate(gens[0]), self.mutate(gens[1]) newPopulation.append(m1) newPopulation.append(m2) population = [gen for fitness,gen in fitElements] population = population[:len(population) - limit] + newPopulation self.population = population return didFinish, countGenerations, populationFitness, convergentNumber # print('finalizou com:', [self.buildFenotype(gen) for gen in self.population])
6,869
1,987
def coding_problem_07(s): """ Given the mapping a = 1, b = 2, ... z = 26, and an encoded message, count the number of ways it can be decoded. Examples: >>> coding_problem_07('111') # possible interpretations: 'aaa', 'ka', 'ak' 3 >>> coding_problem_07('2626') # 'zz', 'zbf', 'bfz', 'bfbf' 4 """ symbols = map(str, range(1, 27)) if not s: return 1 matches = filter(lambda symbol: s.startswith(symbol), symbols) encodings = [coding_problem_07(s[len(m):]) for m in matches] return sum(encodings) if __name__ == '__main__': import doctest doctest.testmod(verbose=True)
639
246
# django imports from django.conf import settings from django.contrib.auth.decorators import permission_required from django.http import HttpResponse # lfs imports from lfs.core.utils import import_symbol @permission_required("core.manage_shop") def add_criterion(request): """ Adds a new criterion form. """ try: default_criterion = settings.LFS_CRITERIA[0] default_criterion = import_symbol(default_criterion[0]) result = default_criterion().render(request, 10) except: result = "" default_criterion = settings.LFS_CRITERIA[0] default_criterion = import_symbol(default_criterion[0]) result = default_criterion().render(request, 10) return HttpResponse(result) @permission_required("core.manage_shop") def change_criterion_form(request): """ Changes the changed criterion form to the given type (via request body) form. This is called via an AJAX request. The result is injected into the right DOM node. """ type = request.POST.get("type", "price") criterion = import_symbol(type) # create dummy criterion result = criterion(pk=1).render(request, 10) return HttpResponse(result)
1,201
372
# Generated by Django 3.1.7 on 2021-05-26 12:56 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('apps', '0054_auto_20210525_1114'), ] operations = [ migrations.AddField( model_name='appinstance', name='access', field=models.CharField(blank=True, default='private', max_length=20, null=True), ), ]
428
153
# COMO TRABALHAMOS MUITO COM ROTINAS, CRIAR UMA FUNÇÃO E CHAMA-LA NO CÓDIGO FACILITA MUITO NÃO TER QUE ESCREVER A MESMA LINHA VÁRIAS VEZES # USAMOS O COMANDO DEF PARA DECLARAR UMA FUNCAO, DEPOIS UM NOME PARA A FUNÇÃO EX: # def nomeDaFuncao(): # # BLOCO COM O CÓDIGO QUE DESEJA QUE SEJÁ EXECULTADO NA CHAMADA DA FUNÇÃO # print('Funcionando') # nomeDaFuncao()# O RESULTADO FUI ("Funcionando") # SEMPRE QUE PRESCISAMOS MOSTRAR UMA LINHA ESCREVEMOS TODA VEZ O MESMO CÓDIGO # PARA FACILITAR PODEMOS CRIAR UMA FUNÇÃO QUE MOSTRE ESSA LINHA SEM PRESCISAR ESCREVER O CÓDIGO PRINT VÁRIAS VEZES def caixa(tipo, msg): print(tipo * (len(msg) * 2)) print(msg) print(tipo * (len(msg) * 2)) caixa('#', 'Python')# O RESULTADO É ''' ############ Python ############''' # MAS PODE VARIAER DE ACORDO COM O TEXTO SOLICITADO def soma(a, b):# DOIS PARAMENTROS a E b QUE SEMPRE QUANDO A FUNÇÃO FOR CHAMADA DEVE INSIRIR ESSES DOIS VALORES s = a + b print(s) soma(1, 4)# RESULTADO É 5 soma(b=3, a=1)# RESULTADO É 4, RECLARAMOS QUE O B=3 E A=1 def contador(*n):# O PARAMETRO AGORA PODE RECEBER VÁRIOS VALORES, QUANDDO COLOCAMOS O * O PYTHON SABE QUE VAI RECEBER VALORES VARIADOS # TODOS OS VALORES VÃO SER GUARDADO DENTRO DE UMA TUPLA print(len(n)) contador(1, 5, 7, 7) def dobrar(list): p = 0 while p < len(list): list[p] *= 2 p += 1 valoresd = list[:] return valoresd valores = [1, 2, 9, 7, 4] print(valores) dobrar(valores) print(valores)
1,512
758
''' Tests for seg2toph5 ''' import os import sys import unittest import logging from mock import patch from testfixtures import OutputCapture, LogCapture from ph5.utilities import seg2toph5, initialize_ph5 from ph5.core.tests.test_base import LogTestCase, TempDirTestCase,\ initialize_ex from ph5.core import ph5api class TestSeg2toPH5_main(TempDirTestCase, LogTestCase): def tearDown(self): self.ph5object.ph5close() super(TestSeg2toPH5_main, self).tearDown() def test_main(self): # check external links created data_nodes = ['Das_g_0000SV01', 'Das_g_0000SV02', 'Das_g_0000SV03', 'Das_g_0000SV04', 'Das_g_0000SV05', 'Das_g_0000SV06', 'Das_g_0000SV07', 'Das_g_0000SV08', 'Das_g_0000SV09', 'Das_g_0000SV10', 'Das_g_0000SV11', 'Das_g_0000SV12', 'Das_g_0000SV13', 'Das_g_0000SV14', 'Das_g_0000SV15', 'Das_g_0000SV16', 'Das_g_0000SV17', 'Das_g_0000SV18', 'Das_g_0000SV19', 'Das_g_0000SV20', 'Das_g_0000SV21', 'Das_g_0000SV22', 'Das_g_0000SV23', 'Das_g_0000SV24', 'Das_g_0000SV25', 'Das_g_0000SV26', 'Das_g_0000SV27', 'Das_g_0000SV28', 'Das_g_0000SV29', 'Das_g_0000SV30', 'Das_g_0000SV31', 'Das_g_0000SV32', 'Das_g_0000SV33', 'Das_g_0000SV34', 'Das_g_0000SV35', 'Das_g_0000SV36', 'Das_g_0000SV37', 'Das_g_0000SV38', 'Das_g_0000SV39', 'Das_g_0000SV40', 'Das_g_0000SV41', 'Das_g_0000SV42', 'Das_g_0000SV43', 'Das_g_0000SV44', 'Das_g_0000SV45', 'Das_g_0000SV46', 'Das_g_0000SV47', 'Das_g_0000SV48', 'Das_g_0000SV49', 'Das_g_0000SV50', 'Das_g_0000SV51', 'Das_g_0000SV52', 'Das_g_0000SV53', 'Das_g_0000SV54', 'Das_g_0000SV55', 'Das_g_0000SV56', 'Das_g_0000SV57', 'Das_g_0000SV58', 'Das_g_0000SV59', 'Das_g_0000SV60'] testargs = ['initialize_ph5', '-n', 'master.ph5'] with patch.object(sys, 'argv', testargs): initialize_ph5.main() # add seg2 to ph5 testargs = ['seg2toph5', '-n', 'master.ph5', '-r', os.path.join(self.home, "ph5/test_data/seg2/15001.dat")] with patch.object(sys, 'argv', testargs): with OutputCapture(): with LogCapture() as log: log.setLevel(logging.ERROR) seg2toph5.main() # before commit caf6978, there would be 1 error log if run this in # environment that uses Obspy 1.2.2 self.assertEqual(len(log.records), 0) self.ph5object = ph5api.PH5(path=self.tmpdir, nickname='master.ph5') target_p1 = 'miniPH5_00001.ph5:/Experiment_g/Maps_g/' targets = [target_p1 + n for n in data_nodes] node = self.ph5object.ph5.get_node("/Experiment_g/Maps_g/") i = 0 ret_targets = [] for n in self.ph5object.ph5.list_nodes(node): if hasattr(n, 'target'): ret_targets.append(n.target) i += 1 self.assertEqual(ret_targets, targets) target_p1 = 'miniPH5_00001.ph5:/Experiment_g/Receivers_g/' targets = [target_p1 + n for n in data_nodes] node = self.ph5object.ph5.get_node("/Experiment_g/Receivers_g/") i = 0 ret_targets = [] for n in self.ph5object.ph5.list_nodes(node): if hasattr(n, 'target'): ret_targets.append(n.target) i += 1 self.assertEqual(ret_targets, targets) def test_update_external_references(self): self.ph5object = seg2toph5.EX = \ initialize_ex('master.ph5', '.', True) keys = ['external_file_name_s', 'hdf5_path_s', 'serial_number_s'] INDEX_T_DAS_rows = \ [{'external_file_name_s': './miniPH5_00001.ph5', 'hdf5_path_s': '/Experiment_g/Receivers_g/Das_g_0000SV01'}] seg2toph5.INDEX_T_DAS = seg2toph5.Rows_Keys(INDEX_T_DAS_rows, keys) INDEX_T_MAP_rows = \ [{'external_file_name_s': './miniPH5_00001.ph5', 'hdf5_path_s': '/Experiment_g/Maps_g/Das_g_0000SV01'}] seg2toph5.INDEX_T_MAP = seg2toph5.Rows_Keys(INDEX_T_MAP_rows, keys) seg2toph5.update_external_references() # check if external links are created node = self.ph5object.ph5.get_node("/Experiment_g/Receivers_g/") target = 'miniPH5_00001.ph5:/Experiment_g/Receivers_g/Das_g_0000SV01' for n in self.ph5object.ph5.list_nodes(node): if hasattr(n, 'target'): self.assertEqual(n.target, target) break node = self.ph5object.ph5.get_node("/Experiment_g/Maps_g/") target = 'miniPH5_00001.ph5:/Experiment_g/Maps_g/Das_g_0000SV01' for n in self.ph5object.ph5.list_nodes(node): if hasattr(n, 'target'): self.assertEqual(n.target, target) break if __name__ == "__main__": unittest.main()
5,160
2,229
from .dreamer import Dreamer
29
10
import time from zcash.node import * node1 = Node(8002, "node 1") node1.start() # wait for string time.sleep(2) node1.print_blockchain() time.sleep(2) node2 = Node(8003, "node 2") node2.start() # wait for string time.sleep(2) node2.print_blockchain() node1.get_balance() node2.get_balance() # node1 mint and pour to send to node2 tx1 = node1.mint_coin(1) node1.broadcast_new_transaction(tx1) # waiting for tx broadcast time.sleep(2) node1.show_coin() tx2 = node1.mint_coin(1) node1.broadcast_new_transaction(tx2) # waiting for tx broadcast time.sleep(2) node1.show_coin() coin_old_1 = list(node1.coin_set)[0] coin_old_2 = list(node1.coin_set)[1] tx3 = node1.pour_coin(coin_old_1, coin_old_2, node1.addr_sk, node1.addr_sk, 1, 1, node2.addr_pk, node2.addr_pk, 0, "") sn_list = [tx3.tx_pour[1], tx3.tx_pour[2]] # tx = Transaction(sender=node1.wallet.address, receiver=node2.wallet.address, amount=0.3) # sig = node1.wallet.sign(str(tx)) # tx.set_sign(node1.wallet.pubkey, sig) node1.broadcast_new_transaction(tx3) # waiting for tx broadcast time.sleep(2) node1.show_coin() node2.show_coin() node2.receive_coin(node2.addr_pk, node2.addr_sk) node2.print_blockchain() node2.get_balance() node1.get_balance() node1.print_blockchain()
1,251
568
from common_python.tellurium import util import pandas as pd import numpy as np import unittest class TestFunctions(unittest.TestCase): def testDfToSer(self): data = range(5) df = pd.DataFrame({'a': data, 'b': [2*d for d in data]}) ser = util.dfToSer(df) assert(len(ser) == len(df.columns)*len(df)) def testDfToSer(self): data = range(5) df = pd.DataFrame({'a': data, 'b': [2*d for d in data]}) ser = util.dfToSer(df) assert(len(ser) == len(df.columns)*len(df)) def testInterpolateTime(self): MAX = 10 SER = pd.Series(range(MAX), index=range(MAX)) self.assertEqual(util.interpolateTime(SER, 0.4), 0.4) self.assertEqual(util.interpolateTime(SER, -1), 0) self.assertEqual(util.interpolateTime(SER, MAX), MAX-1) if __name__ == '__main__': unittest.main()
822
324
# -*- coding: utf-8 -*- # © 2017 Sunflower IT (http://sunflowerweb.nl) # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from openerp import fields, models, api, _ from openerp.exceptions import Warning class AccountInvoice(models.Model): _inherit = 'account.invoice' @api.multi def release_timesheet_lines(self): self.ensure_one() self.env['hr.analytic.timesheet'].search([ ('invoice_id', '=', self.id) ]).with_context(override_check=True).write({'invoice_id': None})
535
194
import configparser import logging import threading from selenium.webdriver.common.by import By from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.support import expected_conditions as EC from selenium import webdriver logger = logging.getLogger(__name__) class SeleniumWorker(threading.Thread): def __init__(self, queue, config): threading.Thread.__init__(self) self.chromedriver_path = config['CHROMEDRIVER_PATH'] self.desk_url = config['DESK_URL'] self.login_required = config['DESK_LOGIN_REQUIRED'] self.queue = queue self.chromedriver = None # Set options for chromedriver self.chromecaps = webdriver.DesiredCapabilities.CHROME.copy() # Accept insecure connections self.chromecaps['acceptInsecureCerts'] = \ config['BROWSER_INSECURE_CERTS'] # Load javascript file # If the filename ends with '.url', read with config-parser bookmarklet_path = config['BOOKMARKLET_PATH'] if bookmarklet_path.endswith('.url'): # parse with config parser parser = configparser.ConfigParser() parser.read(bookmarklet_path) if 'InternetShortcut' in parser: self.js = parser['InternetShortcut']['URL'] else: raise ValueError('Bookmarklet file must be a web link!') else: with open(bookmarklet_path, "r") as f: self.js = f.read() def run(self): """Runs in an endless loop until False was put on the queue. If True is on the queue, opens a browser and runs bookmarklet. If None is on the queue, closes the browser.""" logger.debug('Thread running.') while True: q = self.queue.get() if q: logger.info('Starting browser...') # Instantiate driver (opens browser) if self.chromedriver is None: logger.debug('No browser running. Starting browser...') self.chromedriver = webdriver.Chrome( self.chromedriver_path, desired_capabilities=self.chromecaps) # Open a website logger.debug('Calling url') self.chromedriver.get(self.desk_url) # Log in if needed if self.login_required: self.desk_login() # Execute JavaScript if self.js is not None: logger.info('Executing JavaScript...') # Execute javascript self.chromedriver.execute_script(self.js) else: logger.info('Closing browser...') # Close browser if self.chromedriver is not None: self.chromedriver.quit() self.chromedriver = None if q is False: logger.info('Exiting worker loop...') break def desk_login(self): logger.info('attempting login to desk...') # the user-input fields have weird ids, so we need to select # them by searching for ids containing 'Username' or 'Password' userfields = self.chromedriver.find_elements_by_css_selector( "input[id*='Username']") pwdfields = self.chromedriver.find_elements_by_css_selector( "input[id*='Password']") if (len(userfields) > 0) and (len(pwdfields) > 0): userfields[0].send_keys(self.desk_username) pwdfields[0].send_keys(self.desk_password) loginbtn = self.chromedriver.find_element_by_xpath( "//button[@type='submit']") loginbtn.click() # Wait for the new page to be fully loaded WebDriverWait(self.chromedriver, 10).until( EC.presence_of_element_located((By.CLASS_NAME, "timeline-header")) ) else: logger.info( 'Expected Login page but found no login fields. Ignored')
4,205
1,113
#!/usr/bin/env python # -*- coding: utf-8 -*- # """The setup.py file.""" import os import sys from setuptools import find_packages, setup from setuptools.command.install import install VERSION = "0.3.15" URL = "https://github.com/zxdavb/evohome-async" with open("README.md", "r") as fh: LONG_DESCRIPTION = fh.read() class VerifyVersionCommand(install): """Custom command to verify that the git tag matches our VERSION.""" def run(self): tag = os.getenv("CIRCLE_TAG") if tag != VERSION: info = f"The git tag: '{tag}' does not match the package ver: '{VERSION}'" sys.exit(info) setup( name="evohome-async", description="An async client for connecting to Honeywell's TCC RESTful API.", keywords=["evohome", "total connect comfort", "round thermostat"], author="Andrew Stock & David Bonnes", author_email="zxdavb@gmail.com", url=URL, download_url=f"{URL}/archive/{VERSION}.tar.gz", install_requires=[val.strip() for val in open("requirements.txt")], long_description=LONG_DESCRIPTION, long_description_content_type="text/markdown", packages=find_packages(exclude=["test", "docs"]), version=VERSION, license="Apache 2", python_requires=">=3.7", classifiers=[ "Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.7", "Topic :: Home Automation", ], cmdclass={ "verify": VerifyVersionCommand, }, )
1,595
525
#!/usr/bin/env python # Copyright (c) 2019 - The Procedural Generation for Gazebo authors # For information on the respective copyright owner see the NOTICE file # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import unittest import numpy as np import os import shutil from pcg_gazebo import random from pcg_gazebo.utils import generate_random_string from pcg_gazebo.generators.creators import create_models_from_config, extrude from pcg_gazebo.simulation.properties import Material, Pose from pcg_gazebo.simulation import SimulationModel from pcg_gazebo.path import Path from shapely.geometry import Polygon, MultiPoint, LineString def _get_colors(): color_names = list(Material.get_xkcd_colors_list().keys()) return [None, 'xkcd', 'random'] + \ [color_names[random.choice(range(len(color_names)))] for _ in range(2)] + \ [[random.rand() for _ in range(4)] for _ in range(2)] def _delete_generated_meshes(sdf): for i in range(len(sdf.links)): for j in range(len(sdf.links[i].collisions)): if sdf.links[i].collisions[j].geometry.mesh is not None: uri = Path(sdf.links[i].collisions[j].geometry.mesh.uri.value) if os.path.isfile(uri.absolute_uri): os.remove(uri.absolute_uri) for j in range(len(sdf.links[i].visuals)): if sdf.links[i].visuals[j].geometry.mesh is not None: uri = Path(sdf.links[i].visuals[j].geometry.mesh.uri.value) if os.path.isfile(uri.absolute_uri): os.remove(uri.absolute_uri) class TestModelFactory(unittest.TestCase): def test_box_after_random_seed(self): box_name = generate_random_string(5) seed = random.randint(0, 10000) random.init_random_state(seed) random_box = dict( type='box', args=dict( size="__import__('pcg_gazebo').random.rand(3)", mass="__import__('pcg_gazebo').random.rand()", name=box_name ) ) ref = create_models_from_config([random_box])[0] for _ in range(3): random.init_random_state(seed) model = create_models_from_config([random_box])[0] self.assertEqual(ref.to_sdf(), model.to_sdf()) random.init_random_state(seed) refs = create_models_from_config( [random_box for _ in range(3)]) for _ in range(3): random.init_random_state(seed) models = create_models_from_config( [random_box for _ in range(3)]) for r, m in zip(refs, models): self.assertEqual(r.to_sdf(), m.to_sdf()) def test_sphere_after_random_seed(self): sphere_name = generate_random_string(5) seed = random.randint(0, 10000) random.init_random_state(seed) random_sphere = dict( type='sphere', args=dict( radius="__import__('pcg_gazebo').random.rand()", mass="__import__('pcg_gazebo').random.rand()", name=sphere_name ) ) ref = create_models_from_config([random_sphere])[0] for _ in range(3): random.init_random_state(seed) model = create_models_from_config([random_sphere])[0] self.assertEqual(ref.to_sdf(), model.to_sdf()) random.init_random_state(seed) refs = create_models_from_config( [random_sphere for _ in range(3)]) for _ in range(3): random.init_random_state(seed) models = create_models_from_config( [random_sphere for _ in range(3)]) for r, m in zip(refs, models): self.assertEqual(r.to_sdf(), m.to_sdf()) def test_cylinder_after_random_seed(self): cyl_name = generate_random_string(5) seed = random.randint(0, 10000) random.init_random_state(seed) random_cyl = dict( type='cylinder', args=dict( radius="__import__('pcg_gazebo').random.rand()", length="__import__('pcg_gazebo').random.rand()", mass="__import__('pcg_gazebo').random.rand()", name=cyl_name ) ) ref = create_models_from_config([random_cyl])[0] for _ in range(3): random.init_random_state(seed) model = create_models_from_config([random_cyl])[0] self.assertEqual(ref.to_sdf(), model.to_sdf()) random.init_random_state(seed) refs = create_models_from_config( [random_cyl for _ in range(3)]) for _ in range(3): random.init_random_state(seed) models = create_models_from_config( [random_cyl for _ in range(3)]) for r, m in zip(refs, models): self.assertEqual(r.to_sdf(), m.to_sdf()) def test_static_box_model(self): for color in _get_colors(): name = generate_random_string(3) pose = [random.rand() for _ in range(6)] size = [random.rand() for _ in range(3)] model_config = [ dict( type='box', args=dict( size=size, name=name, pose=pose, color=color, collision_parameters=dict( mu=random.uniform(0, 10), mu2=random.uniform(0, 10), friction=random.uniform(0, 10), friction2=random.uniform(0, 10), slip1=random.uniform(0, 1), slip2=random.uniform(0, 1), rolling_friction=random.uniform(0, 1), fdir1=[0, 0, 0], max_contacts=1, soft_cfm=random.uniform(0, 10), soft_erp=random.uniform(0, 10), kp=random.uniform(0, 100000), kd=random.uniform(0, 10), max_vel=random.uniform(0, 0.1), min_depth=random.uniform(0, 0.1), split_impulse=False, split_impulse_penetration_threshold=-0.01, restitution_coefficient=random.uniform(0, 1), threshold=random.uniform(0, 1) ) )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) self.assertIsInstance(models[0], SimulationModel) # Test pose of the model self.assertEqual(models[0].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[0].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertTrue(models[0].static) self.assertEqual(len(models[0].links), 1) link_name = models[0].link_names[0] # Test visual element self.assertEqual(len(models[0].links[link_name].visuals), 1) geometry = models[0].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'box') self.assertEqual(geometry.get_param('size'), size) # Test collision element self.assertEqual(len(models[0].links[link_name].collisions), 1) collision = models[0].links[link_name].get_collision_by_name( 'collision') tags = ['mu', 'mu2', 'slip1', 'slip2', 'fdir1'] for tag in tags: self.assertEqual( collision.get_ode_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['friction', 'friction2', 'rolling_friction', 'fdir1'] for tag in tags: self.assertEqual( collision.get_bullet_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'max_vel', 'min_depth'] for tag in tags: self.assertEqual( collision.get_ode_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'split_impulse', 'split_impulse_penetration_threshold'] for tag in tags: self.assertEqual( collision.get_bullet_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['restitution_coefficient', 'threshold'] print(collision.sdf) for tag in tags: self.assertEqual( collision.get_bounce_param(tag), model_config[0]['args']['collision_parameters'][tag]) geometry = collision.geometry self.assertEqual(geometry.get_type(), 'box') self.assertEqual(geometry.get_param('size'), size) # Test color exists material = models[0].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_dynamic_box_model(self): for color in _get_colors(): name = generate_random_string(3) pose = [random.rand() for _ in range(6)] size = [random.rand() for _ in range(3)] mass = random.rand() model_config = [ dict( type='box', args=dict( size=size, name=name, pose=pose, mass=mass, color=color, collision_parameters=dict( mu=random.uniform(0, 10), mu2=random.uniform(0, 10), friction=random.uniform(0, 10), friction2=random.uniform(0, 10), slip1=random.uniform(0, 1), slip2=random.uniform(0, 1), rolling_friction=random.uniform(0, 1), fdir1=[0, 0, 0], max_contacts=1, soft_cfm=random.uniform(0, 10), soft_erp=random.uniform(0, 10), kp=random.uniform(0, 100000), kd=random.uniform(0, 10), max_vel=random.uniform(0, 0.1), min_depth=random.uniform(0, 0.1), split_impulse=False, split_impulse_penetration_threshold=-0.01, restitution_coefficient=random.uniform(0, 1), threshold=random.uniform(0, 1) ) )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) self.assertIsInstance(models[0], SimulationModel) # Test pose of the model self.assertEqual(models[0].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[0].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[0].static) self.assertEqual(len(models[0].links), 1) link_name = models[0].link_names[0] # Test link properties link = models[0].links[link_name] self.assertEqual(link.inertial.mass, mass) self.assertAlmostEqual(link.inertial.ixx, 1. / 12 * mass * (size[1]**2 + size[2]**2)) self.assertAlmostEqual(link.inertial.iyy, 1. / 12 * mass * (size[0]**2 + size[2]**2)) self.assertAlmostEqual(link.inertial.izz, 1. / 12 * mass * (size[0]**2 + size[1]**2)) self.assertEqual(link.inertial.ixy, 0) self.assertEqual(link.inertial.ixz, 0) self.assertEqual(link.inertial.iyz, 0) # Test visual element self.assertEqual(len(link.visuals), 1) geometry = link.visuals[0].geometry self.assertEqual(geometry.get_type(), 'box') self.assertEqual(geometry.get_param('size'), size) # Test collision element self.assertEqual(len(link.collisions), 1) collision = link.get_collision_by_name('collision') tags = ['mu', 'mu2', 'slip1', 'slip2', 'fdir1'] for tag in tags: self.assertEqual( collision.get_ode_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['friction', 'friction2', 'rolling_friction', 'fdir1'] for tag in tags: self.assertEqual( collision.get_bullet_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'max_vel', 'min_depth'] for tag in tags: self.assertEqual( collision.get_ode_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'split_impulse', 'split_impulse_penetration_threshold'] for tag in tags: self.assertEqual( collision.get_bullet_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['restitution_coefficient', 'threshold'] for tag in tags: self.assertEqual( collision.get_bounce_param(tag), model_config[0]['args']['collision_parameters'][tag]) geometry = collision.geometry self.assertEqual(geometry.get_type(), 'box') self.assertEqual(geometry.get_param('size'), size) # Test color exists material = models[0].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_box_factory_fixed_args_with_permutation(self): for color in _get_colors(): n_sizes = random.randint(2, 5) n_masses = random.randint(2, 5) name = generate_random_string(3) sizes = [[random.rand() for _ in range(3)] for _ in range(n_sizes)] pose = [random.rand() for _ in range(6)] masses = [random.rand() for _ in range(n_masses)] model_config = [ dict( type='box_factory', args=dict( name=name, size=sizes, mass=masses, pose=pose, use_permutation=True, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_sizes * n_masses) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] self.assertIn(models[i].links[link_name].inertial.mass, masses) # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'box') self.assertIn(geometry.get_param('size'), sizes) # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'box') self.assertIn(geometry.get_param('size'), sizes) # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_box_factory_fixed_args_no_permutation(self): for color in _get_colors(): n_models = random.randint(2, 5) name = generate_random_string(3) sizes = [[random.rand() for _ in range(3)] for _ in range(n_models)] pose = [random.rand() for _ in range(6)] masses = [random.rand() for _ in range(n_models)] model_config = [ dict( type='box_factory', args=dict( name=name, size=sizes, mass=masses, pose=pose, use_permutation=False, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_models) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] self.assertIn(models[i].links[link_name].inertial.mass, masses) # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'box') self.assertIn(geometry.get_param('size'), sizes) # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'box') self.assertIn(geometry.get_param('size'), sizes) # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_box_factory_lambda_args_with_permutation(self): for color in _get_colors(): n_sizes = random.randint(2, 5) n_masses = random.randint(2, 5) name = generate_random_string(3) sizes = "__import__('numpy').random.random(({}, 3))".format( n_sizes) masses = "__import__('numpy').linspace(1, 10, {})".format(n_masses) pose = [random.rand() for _ in range(6)] model_config = [ dict( type='box_factory', args=dict( name=name, size=sizes, mass=masses, pose=pose, use_permutation=True, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_sizes * n_masses) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'box') # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'box') # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def text_box_forced_permutation(self): n_sizes = 2 n_masses = 3 name = generate_random_string(3) # Test with lambda functions sizes = "__import__('numpy').random.random(({}, 3))".format(n_sizes) masses = "__import__('numpy').linspace(1, 10, {})".format(n_masses) pose = [random.rand() for _ in range(6)] model_config = [ dict( type='box_factory', args=dict( name=name, size=sizes, mass=masses, pose=pose, use_permutation=False, color=None )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_sizes * n_masses) # Test with fixed arguments sizes = [[random.rand() for _ in range(3)] for _ in range(n_sizes)] masses = [random.rand() for _ in range(n_masses)] pose = [random.rand() for _ in range(6)] model_config = [ dict( type='box_factory', args=dict( name=name, size=sizes, mass=masses, pose=pose, use_permutation=False, color=None )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_sizes * n_masses) def test_static_cylinder_model(self): for color in _get_colors(): name = generate_random_string(3) pose = [random.rand() for _ in range(6)] radius = random.rand() length = random.rand() model_config = [ dict( type='cylinder', args=dict( radius=radius, length=length, name=name, pose=pose, color=color, collision_parameters=dict( mu=random.uniform(0, 10), mu2=random.uniform(0, 10), friction=random.uniform(0, 10), friction2=random.uniform(0, 10), slip1=random.uniform(0, 1), slip2=random.uniform(0, 1), rolling_friction=random.uniform(0, 1), fdir1=[0, 0, 0], max_contacts=1, soft_cfm=random.uniform(0, 10), soft_erp=random.uniform(0, 10), kp=random.uniform(0, 100000), kd=random.uniform(0, 10), max_vel=random.uniform(0, 0.1), min_depth=random.uniform(0, 0.1), split_impulse=False, split_impulse_penetration_threshold=-0.01, restitution_coefficient=random.uniform(0, 1), threshold=random.uniform(0, 1) ) )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) self.assertIsInstance(models[0], SimulationModel) # Test pose of the model self.assertEqual(models[0].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[0].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertTrue(models[0].static) self.assertEqual(len(models[0].links), 1) link_name = models[0].link_names[0] # Test visual element self.assertEqual(len(models[0].links[link_name].visuals), 1) geometry = models[0].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertEqual(geometry.get_param('radius'), radius) self.assertEqual(geometry.get_param('length'), length) # Test collision element self.assertEqual(len(models[0].links[link_name].collisions), 1) collision = models[0].links[link_name].get_collision_by_name( 'collision') tags = ['mu', 'mu2', 'slip1', 'slip2', 'fdir1'] for tag in tags: self.assertEqual( collision.get_ode_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['friction', 'friction2', 'rolling_friction', 'fdir1'] for tag in tags: self.assertEqual( collision.get_bullet_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'max_vel', 'min_depth'] for tag in tags: self.assertEqual( collision.get_ode_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'split_impulse', 'split_impulse_penetration_threshold'] for tag in tags: self.assertEqual( collision.get_bullet_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['restitution_coefficient', 'threshold'] for tag in tags: self.assertEqual( collision.get_bounce_param(tag), model_config[0]['args']['collision_parameters'][tag]) geometry = models[0].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertEqual(geometry.get_param('radius'), radius) self.assertEqual(geometry.get_param('length'), length) # Test color exists material = models[0].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_dynamic_cylinder_model(self): for color in _get_colors(): name = generate_random_string(3) pose = [random.rand() for _ in range(6)] radius = random.rand() length = random.rand() mass = random.rand() model_config = [ dict( type='cylinder', args=dict( radius=radius, length=length, mass=mass, name=name, pose=pose, color=color, collision_parameters=dict( mu=random.uniform(0, 10), mu2=random.uniform(0, 10), friction=random.uniform(0, 10), friction2=random.uniform(0, 10), slip1=random.uniform(0, 1), slip2=random.uniform(0, 1), rolling_friction=random.uniform(0, 1), fdir1=[0, 0, 0], max_contacts=1, soft_cfm=random.uniform(0, 10), soft_erp=random.uniform(0, 10), kp=random.uniform(0, 100000), kd=random.uniform(0, 10), max_vel=random.uniform(0, 0.1), min_depth=random.uniform(0, 0.1), split_impulse=False, split_impulse_penetration_threshold=-0.01, restitution_coefficient=random.uniform(0, 1), threshold=random.uniform(0, 1) ) )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) self.assertIsInstance(models[0], SimulationModel) # Test pose of the model self.assertEqual(models[0].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[0].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[0].static) self.assertEqual(len(models[0].links), 1) link_name = models[0].link_names[0] # Test link properties link = models[0].links[link_name] self.assertEqual(link.inertial.mass, mass) self.assertAlmostEqual( link.inertial.ixx, 1. / 12 * mass * (3 * radius**2 + length**2)) self.assertAlmostEqual( link.inertial.iyy, 1. / 12 * mass * (3 * radius**2 + length**2)) self.assertAlmostEqual(link.inertial.izz, 0.5 * mass * radius**2) self.assertEqual(link.inertial.ixy, 0) self.assertEqual(link.inertial.ixz, 0) self.assertEqual(link.inertial.iyz, 0) # Test visual element self.assertEqual(len(models[0].links[link_name].visuals), 1) geometry = models[0].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertEqual(geometry.get_param('radius'), radius) self.assertEqual(geometry.get_param('length'), length) # Test collision element self.assertEqual(len(models[0].links[link_name].collisions), 1) collision = models[0].links[link_name].get_collision_by_name( 'collision') tags = ['mu', 'mu2', 'slip1', 'slip2', 'fdir1'] for tag in tags: self.assertEqual( collision.get_ode_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['friction', 'friction2', 'rolling_friction', 'fdir1'] for tag in tags: self.assertEqual( collision.get_bullet_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'max_vel', 'min_depth'] for tag in tags: self.assertEqual( collision.get_ode_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'split_impulse', 'split_impulse_penetration_threshold'] for tag in tags: self.assertEqual( collision.get_bullet_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['restitution_coefficient', 'threshold'] for tag in tags: self.assertEqual( collision.get_bounce_param(tag), model_config[0]['args']['collision_parameters'][tag]) geometry = models[0].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertEqual(geometry.get_param('radius'), radius) self.assertEqual(geometry.get_param('length'), length) # Test color exists material = models[0].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_cylinder_factory_fixed_args_with_permutation(self): for color in _get_colors(): n_radius = random.randint(2, 4) n_length = random.randint(2, 4) n_masses = random.randint(2, 4) name = generate_random_string(3) radius = [random.rand() for _ in range(n_radius)] length = [random.rand() for _ in range(n_length)] pose = [random.rand() for _ in range(6)] masses = [random.rand() for _ in range(n_masses)] model_config = [ dict( type='cylinder_factory', args=dict( name=name, radius=radius, length=length, mass=masses, pose=pose, use_permutation=True, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_length * n_masses) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] self.assertIn(models[i].links[link_name].inertial.mass, masses) # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertIn(geometry.get_param('radius'), radius) self.assertIn(geometry.get_param('length'), length) # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertIn(geometry.get_param('radius'), radius) self.assertIn(geometry.get_param('length'), length) # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_cylinder_factory_fixed_args_no_permutation(self): for color in _get_colors(): n_models = random.randint(2, 4) name = generate_random_string(3) radius = [random.rand() for _ in range(n_models)] length = [random.rand() for _ in range(n_models)] pose = [random.rand() for _ in range(6)] masses = [random.rand() for _ in range(n_models)] model_config = [ dict( type='cylinder_factory', args=dict( name=name, radius=radius, length=length, mass=masses, pose=pose, use_permutation=False, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_models) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] self.assertIn(models[i].links[link_name].inertial.mass, masses) # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertIn(geometry.get_param('radius'), radius) self.assertIn(geometry.get_param('length'), length) # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') self.assertIn(geometry.get_param('radius'), radius) self.assertIn(geometry.get_param('length'), length) # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_cylinder_factory_lambda_args_with_permutation(self): for color in _get_colors(): n_radius = random.randint(2, 4) n_length = random.randint(2, 4) n_masses = random.randint(2, 4) name = generate_random_string(3) radius = "__import__('numpy').random.random({})".format(n_radius) length = "__import__('numpy').linspace(1, 10, {})".format(n_length) masses = "__import__('numpy').linspace(1, 10, {})".format(n_masses) pose = [random.rand() for _ in range(6)] model_config = [ dict( type='cylinder_factory', args=dict( name=name, radius=radius, length=length, mass=masses, pose=pose, use_permutation=True, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_length * n_masses) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'cylinder') # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_cylinder_forced_permutation(self): n_radius = 1 n_masses = 2 n_length = 3 name = generate_random_string(3) # Test with lambda functions radius = "__import__('numpy').random.random({})".format(n_radius) length = "__import__('numpy').random.random({})".format(n_length) masses = "__import__('numpy').linspace(1, 10, {})".format(n_masses) pose = [random.rand() for _ in range(6)] model_config = [ dict( type='cylinder_factory', args=dict( name=name, radius=radius, length=length, mass=masses, pose=pose, use_permutation=False, color=None )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_length * n_masses) # Test with fixed arguments radius = [random.rand() for _ in range(n_radius)] length = [random.rand() for _ in range(n_length)] masses = [random.rand() for _ in range(n_masses)] pose = [random.rand() for _ in range(6)] model_config = [ dict( type='cylinder_factory', args=dict( name=name, radius=radius, length=length, mass=masses, pose=pose, use_permutation=False, color=None )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_length * n_masses) def test_static_sphere_model(self): for color in _get_colors(): name = generate_random_string(3) pose = [random.rand() for _ in range(6)] radius = random.rand() model_config = [ dict( type='sphere', args=dict( radius=radius, name=name, pose=pose, color=color, collision_parameters=dict( mu=random.uniform(0, 10), mu2=random.uniform(0, 10), friction=random.uniform(0, 10), friction2=random.uniform(0, 10), slip1=random.uniform(0, 1), slip2=random.uniform(0, 1), rolling_friction=random.uniform(0, 1), fdir1=[0, 0, 0], max_contacts=1, soft_cfm=random.uniform(0, 10), soft_erp=random.uniform(0, 10), kp=random.uniform(0, 100000), kd=random.uniform(0, 10), max_vel=random.uniform(0, 0.1), min_depth=random.uniform(0, 0.1), split_impulse=False, split_impulse_penetration_threshold=-0.01, restitution_coefficient=random.uniform(0, 1), threshold=random.uniform(0, 1) ) )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) self.assertIsInstance(models[0], SimulationModel) # Test pose of the model self.assertEqual(models[0].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[0].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertTrue(models[0].static) self.assertEqual(len(models[0].links), 1) link_name = models[0].link_names[0] # Test visual element self.assertEqual(len(models[0].links[link_name].visuals), 1) geometry = models[0].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertEqual(geometry.get_param('radius'), radius) # Test collision element self.assertEqual(len(models[0].links[link_name].collisions), 1) collision = models[0].links[link_name].get_collision_by_name( 'collision') tags = ['mu', 'mu2', 'slip1', 'slip2', 'fdir1'] for tag in tags: self.assertEqual( collision.get_ode_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['friction', 'friction2', 'rolling_friction', 'fdir1'] for tag in tags: self.assertEqual( collision.get_bullet_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'max_vel', 'min_depth'] for tag in tags: self.assertEqual( collision.get_ode_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'split_impulse', 'split_impulse_penetration_threshold'] for tag in tags: self.assertEqual( collision.get_bullet_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['restitution_coefficient', 'threshold'] for tag in tags: self.assertEqual( collision.get_bounce_param(tag), model_config[0]['args']['collision_parameters'][tag]) geometry = models[0].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertEqual(geometry.get_param('radius'), radius) # Test color exists material = models[0].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_dynamic_sphere_model(self): for color in _get_colors(): name = generate_random_string(3) pose = [random.rand() for _ in range(6)] radius = random.rand() mass = random.rand() model_config = [ dict( type='sphere', args=dict( radius=radius, mass=mass, name=name, pose=pose, color=color, collision_parameters=dict( mu=random.uniform(0, 10), mu2=random.uniform(0, 10), friction=random.uniform(0, 10), friction2=random.uniform(0, 10), slip1=random.uniform(0, 1), slip2=random.uniform(0, 1), rolling_friction=random.uniform(0, 1), fdir1=[0, 0, 0], max_contacts=1, soft_cfm=random.uniform(0, 10), soft_erp=random.uniform(0, 10), kp=random.uniform(0, 100000), kd=random.uniform(0, 10), max_vel=random.uniform(0, 0.1), min_depth=random.uniform(0, 0.1), split_impulse=True, split_impulse_penetration_threshold=-0.01, restitution_coefficient=random.uniform(0, 1), threshold=random.uniform(0, 1) ) )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) self.assertIsInstance(models[0], SimulationModel) # Test pose of the model self.assertEqual(models[0].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[0].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[0].static) self.assertEqual(len(models[0].links), 1) link_name = models[0].link_names[0] # Test link properties link = models[0].links[link_name] self.assertEqual(link.inertial.mass, mass) inertia = 2. / 5 * mass * radius**2 self.assertAlmostEqual(link.inertial.ixx, inertia) self.assertAlmostEqual(link.inertial.iyy, inertia) self.assertAlmostEqual(link.inertial.izz, inertia) self.assertEqual(link.inertial.ixy, 0) self.assertEqual(link.inertial.ixz, 0) self.assertEqual(link.inertial.iyz, 0) # Test visual element self.assertEqual(len(models[0].links[link_name].visuals), 1) geometry = models[0].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertEqual(geometry.get_param('radius'), radius) # Test collision element self.assertEqual(len(models[0].links[link_name].collisions), 1) collision = models[0].links[link_name].get_collision_by_name( 'collision') tags = ['mu', 'mu2', 'slip1', 'slip2', 'fdir1'] for tag in tags: self.assertEqual( collision.get_ode_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['friction', 'friction2', 'rolling_friction', 'fdir1'] for tag in tags: self.assertEqual( collision.get_bullet_friction_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'max_vel', 'min_depth'] for tag in tags: self.assertEqual( collision.get_ode_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['soft_cfm', 'soft_erp', 'kp', 'kd', 'split_impulse', 'split_impulse_penetration_threshold'] for tag in tags: self.assertEqual( collision.get_bullet_contact_param(tag), model_config[0]['args']['collision_parameters'][tag]) tags = ['restitution_coefficient', 'threshold'] for tag in tags: self.assertEqual( collision.get_bounce_param(tag), model_config[0]['args']['collision_parameters'][tag]) geometry = models[0].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertEqual(geometry.get_param('radius'), radius) # Test color exists material = models[0].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_sphere_factory_fixed_args_with_permutation(self): for color in _get_colors(): n_radius = random.randint(2, 4) n_masses = random.randint(2, 4) name = generate_random_string(3) radius = [random.rand() for _ in range(n_radius)] pose = [random.rand() for _ in range(6)] masses = [random.rand() for _ in range(n_masses)] model_config = [ dict( type='sphere_factory', args=dict( name=name, radius=radius, mass=masses, pose=pose, use_permutation=True, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_masses) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] self.assertIn(models[i].links[link_name].inertial.mass, masses) # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertIn(geometry.get_param('radius'), radius) # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertIn(geometry.get_param('radius'), radius) # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_sphere_factory_fixed_args_no_permutation(self): for color in _get_colors(): n_models = random.randint(2, 4) name = generate_random_string(3) radius = [random.rand() for _ in range(n_models)] pose = [random.rand() for _ in range(6)] masses = [random.rand() for _ in range(n_models)] model_config = [ dict( type='sphere_factory', args=dict( name=name, radius=radius, mass=masses, pose=pose, use_permutation=False, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_models) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] self.assertIn(models[i].links[link_name].inertial.mass, masses) # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertIn(geometry.get_param('radius'), radius) # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'sphere') self.assertIn(geometry.get_param('radius'), radius) # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_sphere_factory_lambda_args_with_permutation(self): for color in _get_colors(): n_radius = random.randint(2, 4) n_masses = random.randint(2, 4) name = generate_random_string(3) radius = "__import__('numpy').random.random({})".format(n_radius) masses = "__import__('numpy').linspace(1, 10, {})".format(n_masses) pose = [random.rand() for _ in range(6)] model_config = [ dict( type='sphere_factory', args=dict( name=name, radius=radius, mass=masses, pose=pose, use_permutation=True, color=color )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_masses) for i in range(len(models)): # Check the name generator with counter self.assertIn(name + '_', models[i].name) self.assertTrue(models[i].name.split('_')[-1].isdigit()) self.assertIsInstance(models[i], SimulationModel) # Test pose of the model self.assertEqual(models[i].pose.position.tolist(), pose[0:3]) q = Pose.rpy2quat(*pose[3::]) diff = Pose.get_transform(models[i].pose.quat, q) # Test model properties self.assertAlmostEqual(np.sum(diff[0:3]), 0) self.assertFalse(models[i].static) self.assertEqual(len(models[i].links), 1) link_name = models[i].link_names[0] # Test visual element self.assertEqual(len(models[i].links[link_name].visuals), 1) geometry = models[i].links[link_name].visuals[0].geometry self.assertEqual(geometry.get_type(), 'sphere') # Test collision element self.assertEqual(len(models[i].links[link_name].collisions), 1) geometry = models[i].links[link_name].collisions[0].geometry self.assertEqual(geometry.get_type(), 'sphere') # Test color exists material = models[i].links[link_name].visuals[0].material self.assertIsNotNone(material) self.assertIsNotNone(material.ambient) self.assertIsNotNone(material.diffuse) if not isinstance( color, str) and isinstance( color, list) and color is not None: self.assertEqual(material.ambient.value, color) self.assertEqual(material.diffuse.value, color) def test_sphere_forced_permutation(self): n_radius = random.randint(2, 3) n_masses = 2 * n_radius name = generate_random_string(3) # Test with lambda functions radius = "__import__('numpy').random.random({})".format(n_radius) masses = "__import__('numpy').linspace(1, 10, {})".format(n_masses) pose = [random.rand() for _ in range(6)] model_config = [ dict( type='sphere_factory', args=dict( name=name, radius=radius, mass=masses, pose=pose, use_permutation=False, color=None )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_masses) # Test with fixed arguments radius = [random.rand() for _ in range(n_radius)] masses = [random.rand() for _ in range(n_masses)] pose = [random.rand() for _ in range(6)] model_config = [ dict( type='sphere_factory', args=dict( name=name, radius=radius, mass=masses, pose=pose, use_permutation=False, color=None )) ] models = create_models_from_config(model_config) self.assertEqual(len(models), n_radius * n_masses) def test_extrude_polygon(self): # Create mesh by extruding a polygon vertices = [(0, 0), (0, 2), (2, 2), (2, 0), (0, 0)] poly = Polygon(vertices) name = generate_random_string(3) pose = [random.rand() for _ in range(6)] mass = random.rand() height = random.rand() model_config = [ dict( type='extrude', args=dict( polygon=poly, name=name, mass=mass, height=height, pose=pose, color=None ) ) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) for model in models: _delete_generated_meshes(model.to_sdf()) # Extrude only the boundaries cap_style = ['round', 'flat', 'square'] join_style = ['round', 'mitre', 'bevel'] for cs in cap_style: for js in join_style: model_config = [ dict( type='extrude', args=dict( polygon=poly, name=name, mass=mass, height=height, pose=pose, color=None, extrude_boundaries=True, thickness=random.rand(), cap_style=cs, join_style=js ) ) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) for model in models: _delete_generated_meshes(model.to_sdf()) # Create a mesh by dilating point vertices = [(random.rand() * 5, random.rand() * 5)] poly = MultiPoint(vertices) name = generate_random_string(3) pose = [random.rand() for _ in range(6)] mass = random.rand() height = random.rand() model_config = [ dict( type='extrude', args=dict( polygon=poly, name=name, mass=mass, height=height, pose=pose, color=None, thickness=random.rand() ) ) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) for model in models: _delete_generated_meshes(model.to_sdf()) # Create a mesh by dilating a line vertices = [(random.rand() * 5, random.rand() * 5) for _ in range(5)] poly = LineString(vertices) name = generate_random_string(3) pose = [random.rand() for _ in range(6)] mass = random.rand() height = random.rand() for cs in cap_style: for js in join_style: model_config = [ dict( type='extrude', args=dict( polygon=poly, name=name, mass=mass, height=height, pose=pose, color=None, cap_style=cs, join_style=js, thickness=random.rand() ) ) ] models = create_models_from_config(model_config) self.assertEqual(len(models), 1) for model in models: _delete_generated_meshes(model.to_sdf()) def test_invalid_polygon_extrude_inputs(self): vertices = [(random.rand() * 5, random.rand() * 5)] model_config = [ dict( type='extrude', args=dict( polygon=MultiPoint(vertices), thickness=0, height=random.rand() ) ) ] with self.assertRaises(AssertionError): create_models_from_config(model_config) def test_export_to_gazebo_model(self): # Create mesh by extruding a polygon vertices = [(0, 0), (0, 2), (2, 2), (2, 0), (0, 0)] poly = Polygon(vertices) name = generate_random_string(3) pose = [random.rand() for _ in range(6)] mass = random.rand() height = 10 * random.rand() model_config = dict( polygon=poly, name=name, mass=mass, height=height, pose=pose, color=None ) model = extrude(**model_config) model_dir = model.to_gazebo_model() self.assertTrue(os.path.isdir(model_dir)) shutil.rmtree(model_dir) if __name__ == '__main__': unittest.main()
74,403
20,640
#!/usr/bin/env python import rospy import numpy as np import pickle import time import socket from rospy.numpy_msg import numpy_msg from radbot_nuke.msg import detector_msg rospy.init_node('detServer', anonymous=False) pub = rospy.Publisher('/detector_data', detector_msg, queue_size=10) def send_ros_msg(ts, adc, det_sn): ''' takes output from client, packages it, and publishes as message ''' msg = detector_msg() now = time.time() rnow = rospy.Time.now() msg.header.stamp = rnow msg.det_sn.data = det_sn msg.detid = assign_det_num(det_sn) msg.ts_sys = now msg.ts_det = ts msg.channel = adc msg.event_rate = len(adc)*1.0 #process rate data if needed pub.publish(msg) print len(ts), len(adc), det_sn, " Published!!" def assign_det_num(ser_no): assignments = {'eRC4129':1, 'eRC4216':3, 'eRC4131':2, 'eRC4214':5, 'eRC4130':4} return assignments[ser_no] def process_rates(rates): ''' If we start to hit the list mode buffer throughput, this parses the "rate data" flag returned with each eMorpho buffer readout. Using the live and dead time values you can assess higher count rate envs https://www.bridgeportinstruments.com/products/mds/mds_doc/read_rates.php ''' # old_acq_tme = 0 # old_ded_tme = 0 # old_dt = 0 # old_trig = 0 # old_event = 0 if rates[3] <= 0: #prevents dt frac from going to inf rates[3] = 1 acq_tme = (rates[0] - old_acq_tme) ded_tme = (rates[3] - old_ded_tme) events = rates[1] - old_event trigs = rates[2] - old_trig active_time = (acq_tme * np.divide(65536, 80000000, dtype=np.float64)) # 16bit encoding, 80MHz, must use np.divide to return float64 msg.event_rate = events/active_time #event rate in seconds msg.trig_rate = trigs/active_time #trigger rate in seconds msg.true_rate = np.divide(trigs, active_time, dtype=np.float64) *(1 - np.divide(ded_tme, acq_tme, dtype=np.float64)) #true incident pulse rate i n seconds old_acq_tme = rates[0] old_ded_tme = rates[3] old_event = rates[1] old_trig = rates[2] def start_server(port): ''' Starts a server to constantly monior on some port for incoming data from one of the multiprocessing subprocesses. The clinet will send a pickeled list consiting of the detectors serial number, an array of event time stamps and an array of event ACD (channel) values This data is then published on the /detector_data topic using the detector_msg message ''' try: serv = socket.socket(socket.AF_INET, socket.SOCK_STREAM) serv.bind(('0.0.0.0', port)) serv.listen(5) buf = None print "Waiting to receive connection..." while True: conn, addr = serv.accept() while True: data = conn.recv(4096) if buf is None: buf = data else: buf+=data if '<-o->' in buf: message = buf[:-len('<-o->')] # PROCESS MESSAGE HERE incoming_data = pickle.loads(message) # print incoming_data[0], len(incoming_data[1]), len(incoming_data[2]) send_ros_msg(incoming_data[1], incoming_data[2], incoming_data[0]) #ts, acd, sn buf = None conn.close() break # print 'client disconnected' except socket.error, exc: print "Caught exception socket.error : %s" % exc def main(): start_server(port=8082) if __name__ == '__main__': main()
3,646
1,281
import re from itertools import chain from collections import defaultdict, namedtuple import pandas as pd from tqdm import tqdm Address = namedtuple('Address', ['to', 'kai', 'ku', 'mune', 'chome', 'ban', 'go', 'postal', 'endgo', 'tokens']) class Postman: """ Converts Japanese address in raw string to structured object. """ to_regex = r'(({})都)'.format('|'.join(['東京'])) ku_regex = r'#TO# (.*区(佃)?)' postal_regex = r'((〒?[0-90-9]+([ー-‐-])){1,}([0-90-9]+))' chome_ban_go = r'([0-90-9一⼀⼆三四五六七⼋九⼗ー]+([{}]+))' # Generic pattern for chome, ban and go. chome_regex = chome_ban_go.format('丁目') ban_regex = chome_ban_go.format('番街目') go_regex = chome_ban_go.format('号棟室館') endgo_regex = r'(([0-90-9一⼀⼆三四五六七⼋九⼗]+$))' # Number at the end. ban_go_regex = ban_regex + go_regex + '?' # Not useful? mune_regex = r'([0-90-9A-Z一⼀⼆三四五六七⼋九⼗]+([号]?[棟]))' kai_regex = r'(([0-90-9ー]+|[一⼀⼆三四五六七⼋九⼗]+)[階F])' JAPANESE_ADDRESS_REGEXES = [ (to_regex, ' #TO# '), (kai_regex, ' #KAI# '), (ku_regex, ' #KU# '), (mune_regex, ' #MUNE# '), (chome_regex, ' #CHOME# '), # Catch chome / ban / go individually. (ban_regex, ' #BAN# '), (go_regex, ' #GO# '), (postal_regex, ' #POSTAL# '), (endgo_regex, ' #ENDGO# ') ] def __init__(self): pass def normalize(self, pattern, substitution, text): matches = [match[0] for match in re.findall(pattern, text)] for match in matches: text = text.replace(match, substitution) return matches, ' '.join(text.split()).strip() def tokenize(self, addr, use_char=False, return_str=False): """ Returns an address object. """ # Go through the regex and keep the group matches. regex_matches = defaultdict(list) for pattern, substitute in self.JAPANESE_ADDRESS_REGEXES: matches, addr = self.normalize(pattern, substitute, addr) regex_matches[substitute.strip()] = matches # Options to return different tokens. if use_char: # Split the non-captured groups by characters. tokens = list(chain(*[[token] if token.strip().startswith('#') and token.strip().endswith('#') else list(token) for token in addr.split()])) else: # Simply return the tokens split by spaces. tokens = addr.split() # Replace the placeholders back. for k in regex_matches: for v in regex_matches[k]: tokens[tokens.index(k)] = v address = Address(to=regex_matches['#TO#'], kai=regex_matches['#KAI#'], ku=regex_matches['#KU#'], mune=regex_matches['#MUNE#'], chome=regex_matches['#CHOME#'], ban=regex_matches['#BAN#'], go=regex_matches['#GO#'], postal=regex_matches['#POSTAL#'], endgo=regex_matches['#ENDGO#'], tokens=tokens) return ' '.join(address.tokens) if return_str else address
3,166
1,161
from setuptools import setup, find_packages with open('requirements.txt') as f: install_requires = f.read().strip().split('\n') # get version from __version__ variable in vitalpbx/__init__.py from vitalpbx import __version__ as version setup( name='vitalpbx', version=version, description='Something', author='Someone', author_email='someone@somewhere.com', packages=find_packages(), zip_safe=False, include_package_data=True, install_requires=install_requires )
476
160
for g in range(int(input())): n, k = [int(x) for x in input().split()] v = [1 for x in range(1, n+1)] m = 0 i = p = 1 while (m < n-1): if v[i] == 1: p += 1 if p == k: v[i] = 0 m += 1 p = 0 i += 1 if i == n: i = 0 i = 0 while v[i] == 0: i += 1 print('Case {}: {}'.format(g+1, i+1))
381
172
#!/usr/bin/env python3 # Copyright (c) 2020 Jarret Dyrbye # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php import time import sys import json import argparse from twisted.internet import reactor from twisted.internet.task import LoopingCall from autobahn.twisted.websocket import WebSocketServerProtocol from autobahn.twisted.websocket import WebSocketServerFactory from txzmq import ZmqEndpoint, ZmqEndpointType from txzmq import ZmqFactory from txzmq import ZmqSubConnection from bolt.util import h2b from onionstudio.art_db import ArtDb from onionstudio.compressor import compressor from onionstudio.extension import Extension, PIXEL_TLV_TYPE UNPAID_PRUNE_CHECK = 60 UNPAID_PRUNE_SECONDS = 120 ############################################################################### class AppClient(WebSocketServerProtocol): def onConnect(self, request): print("Client connecting: {0}".format(request.peer)) def onOpen(self): print("WebSocket client connection open.") self.server.clients.append(self) art_bin = self.server.app.art_db.to_bin() compressed_bin = compressor(art_bin) print("compressed art bin len: %d" % len(compressed_bin)) self.sendMessage(compressed_bin, isBinary=True) def onMessage(self, payload, isBinary): print("got message?") def onClose(self, wasClean, code, reason): print("WebSocket connection closed: {0}".format(reason)) ############################################################################### class AppServer(WebSocketServerFactory): def __init__(self, port, app): ws_url = u"ws://0.0.0.0:%d" % port super().__init__() self.setProtocolOptions(openHandshakeTimeout=15, autoPingInterval=30, autoPingTimeout=5) self.protocol = AppClient self.protocol.server = self self.clients = [] print("listening on websocket %s" % ws_url) reactor.listenTCP(port, self) self.app = app def echo_to_clients(self, pixels): pixels = [{'x': p.x, 'y': p.y, 'rgb': p.rgb} for p in pixels] message = {'pixels': pixels} message = json.dumps(message) print("echoing to clients: %s" % message) for c in self.clients: c.sendMessage(message.encode("utf8")) ############################################################################### HTLC_ACCEPTED_TAG = "htlc_accepted".encode("utf8") FORWARD_EVENT_TAG = "forward_event".encode("utf8") class App(object): def __init__(self, endpoint, mock_endpoint, port, art_db_dir): self.endpoint = endpoint self.mock_endpoint = mock_endpoint self.port = port self.art_db_dir = art_db_dir self.unpaid_htlcs = {} self.prune_loop = LoopingCall(self.prune_unpaid) self.prune_loop.start(interval=UNPAID_PRUNE_CHECK, now=False) ########################################################################### def setup_art_db(self): self.art_db = ArtDb(self.art_db_dir) ########################################################################### def setup_websocket(self): self.ws_server = AppServer(self.port, self) ########################################################################### def setup_zmq(self): zmq_factory = ZmqFactory() print("subscribing on: %s" % self.endpoint) sub_endpoint = ZmqEndpoint(ZmqEndpointType.connect, self.endpoint) sub_connection = ZmqSubConnection(zmq_factory, sub_endpoint) sub_connection.gotMessage = self.zmq_message sub_connection.subscribe(FORWARD_EVENT_TAG) sub_connection.subscribe(HTLC_ACCEPTED_TAG) print("subscribing on: %s" % self.mock_endpoint) sub_mock_endpoint = ZmqEndpoint(ZmqEndpointType.connect, self.mock_endpoint) sub_mock_connection = ZmqSubConnection(zmq_factory, sub_mock_endpoint) sub_mock_connection.gotMessage = self.zmq_message sub_mock_connection.subscribe(FORWARD_EVENT_TAG) sub_mock_connection.subscribe(HTLC_ACCEPTED_TAG) def forward_event_message(self, message): d = json.loads(message.decode('utf8'))['forward_event'] print("got %s" % json.dumps(d, indent=1)) if d['status'] != 'settled': print("invoice is not settled") return if d['payment_hash'] in self.unpaid_htlcs.keys(): self.finish_htlc(d['payment_hash']) def htlc_accepted_message(self, message): d = json.loads(message.decode('utf8')) payment_hash = d['htlc']['payment_hash'] amount = int(d['htlc']['amount'][:-4]) forward_amount = int(d['onion']['forward_amount'][:-4]) payload_hex = d['onion']['payload'] payload = h2b(payload_hex) paid = amount - forward_amount parsed_payload, err = Extension.parse(payload) if err: print("could not parse payload: %s" % err) return print("parsed payload %s" % parsed_payload) pixels = parsed_payload['tlvs'][PIXEL_TLV_TYPE]['pixels'] if (amount - forward_amount) < len(pixels) * 1000: print("forward fee not enough") return self.unpaid_htlcs[payment_hash] = {'payload_hex': payload_hex, 'recv_time': time.time()} def finish_htlc(self, payment_hash): payload_hex = self.unpaid_htlcs.pop(payment_hash)['payload_hex'] payload = h2b(payload_hex) parsed_payload, err = Extension.parse(payload) assert err is None, "could not parse the second time?" pixels = parsed_payload['tlvs'][PIXEL_TLV_TYPE]['pixels'] self.art_db.record_pixels(payload_hex, pixels) self.ws_server.echo_to_clients( pixels) def zmq_message(self, message, tag): if tag == FORWARD_EVENT_TAG: self.forward_event_message(message) elif tag == HTLC_ACCEPTED_TAG: self.htlc_accepted_message(message) else: sys.exit("unknown tag: %s" % tag) ########################################################################### def prune_unpaid(self): now = time.time() new = {k: v for k, v in self.unpaid_htlcs.items() if (now - v['recv_time']) < UNPAID_PRUNE_SECONDS} #print("pruning: %d to %d" % (len(self.unpaid_htlcs), len(new))) self.unpaid_htlcs = new ########################################################################### def run(self): self.setup_websocket() self.setup_zmq() self.setup_art_db() def stop(self): self.art_db.unmap_art_bin() ############################################################################### DEFAULT_WEBSOCKET_PORT = 9000 DEFAULT_ZMQ_SUBSCRIBE_ENDPOINT = "tcp://127.0.0.1:6666" DEFAULT_MOCK_ZMQ_SUBSCRIBE_ENDPOINT = "tcp://127.0.0.1:5557" DEFAULT_ART_DB_DIR = "/tmp/onionstudio/" parser = argparse.ArgumentParser(prog="onionstudio.py") parser.add_argument("-e", "--endpoint", type=str, default=DEFAULT_ZMQ_SUBSCRIBE_ENDPOINT, help="endpoint to subscribe to for zmq notifications from " "c-lightning via cl-zmq.py plugin") parser.add_argument("-m", "--mock-endpoint", type=str, default=DEFAULT_MOCK_ZMQ_SUBSCRIBE_ENDPOINT, help="endpoint to subscribe to zmq notifcations from a " "test script such as mock-png.py") parser.add_argument("-w", "--websocket-port", type=int, default=DEFAULT_WEBSOCKET_PORT, help="port to listen for incoming websocket connections") parser.add_argument("-a", "--art-db-dir", type=str, default=DEFAULT_ART_DB_DIR, help="directory to save the image state and logs") settings = parser.parse_args() a = App(settings.endpoint, settings.mock_endpoint, settings.websocket_port, settings.art_db_dir) a.run() reactor.addSystemEventTrigger("before", "shutdown", a.stop) reactor.run()
8,215
2,554
#!/usr/bin/env python3 from __future__ import print_function from rosflight_holodeck_interface import ROSflightHolodeck import numpy as np import matplotlib.pyplot as plt from tqdm import tqdm import time import holodeck from holodeck import agents from holodeck.environments import * from holodeck import sensors from IPython.core.debugger import Pdb import os import cv2 print( '\nHOLODECK PATH: {}\n'.format( os.path.dirname( holodeck.__file__ ) ) ) np.set_printoptions(precision=3, suppress=True, sign=' ', floatmode='fixed') if __name__ == '__main__': env = holodeck.make("Ocean") RF = ROSflightHolodeck() env.reset() env.tick() # wave intensity: 1-13(int), wave size: 1-8(int), wave direction: 0-360 degreese (float) env.set_ocean_state(3, 3, 90) env.set_aruco_code(False) x0 = np.array([0, 0, 0, # position [0-2] 1, 0, 0, 0, # attitude [3-6] 0, 0, 0, # velocity [7-9] 0, 0, 0, # omega [10-12] 0, 0, 0], dtype=np.float64) # acc [13-15] h0 = 41 # initial altitude [m] x0[2] = -h0 # x0[0] = -34.5 RF.init() RF.setState(x0) RF.setTime(10) uav_cmd = np.array([0, 0, 0, 0]) boat_cmd = 0 env.act("uav0", uav_cmd) env.act("boat0", boat_cmd) pos0 = x0[:3] * [100, 100, -100] # convert to cm, negate z env.teleport("uav0", location=pos0, rotation=[0,0,0]) env.teleport("boat0", location=[-2000,0,0], rotation=[0,0,0]) frate = 30 # frame rate of camera/rendering [Hz] simrate = 800 # rate of simulated dynamics [Hz] n = simrate//frate # Number of sim iterations between frames dt = 1.0/simrate #'F' order because eigen matrices are column-major while numpy are row-major # x_arr = np.zeros((16, n), order='F') x = np.zeros(16) t = np.zeros(n) rf_outputs = np.zeros(4) state = np.array([]) pos = np.zeros(3) att = np.zeros(3) ground = -0.1 collision = False count = 0 while 1: # Main loop: 1 iteration = 1 rendered frame if not collision: for i in range(n): # Loop between frames (dynamics/control) RF.run(dt) time.sleep(dt) RF.getState(x) # Make sure mav doesn't fall through ground if x[2] > ground and x[9] > 0: # at ground level and not gaining altitude x_ground = np.copy(x0) x_ground[:3] = [x[0], x[1], ground] att_eul = Quaternion2Euler(x[3:7]) ground_eul = np.array([0, 0, att_eul[2]]) x_ground[3:7] = Euler2Quaternion(ground_eul) pos = x_ground[:3] * [100,100,-100] att = ground_eul * 180/np.pi RF.setState(x_ground) state = env.set_state("uav0", pos, att, [0,0,0], [0,0,0])["uav0"] else: # format state for holodeck pos = x[:3] * [100,100,-100] # cm, negate z to convert to LH frame att = Quaternion2Euler(x[3:7]) * 180/np.pi vel = x[7:10] * [100,100,-100] angvel = np.copy(x[10:13]) state = env.set_state("uav0", pos, att, vel, angvel)["uav0"] collision = state['CollisionSensor'] elif collision: # Use position given by holodeck state = env.tick()["uav0"] x = np.copy(x0) x[:3] = state['LocationSensor'] * [1,1,-1] x[7:10] = state['VelocitySensor'] * [1,1,-1] RF.setState(x) for k in range(10): RF.run(dt) time.sleep(dt*(n/10)) RF.getState(x) RF.getOutputs(rf_outputs) if x[9] < 0: # gaining altitude, switch back to RF dynamics collision = False # Show UAV's camera feed # frame = state['RGBCamera'] # cv2.imshow('Camera', frame) # cv2.waitKey(1) # For debugging RF.getState(x)
4,078
1,539
# -*- coding: utf-8 -*- # <nbformat>3.0</nbformat> # <codecell> from IPython.display import HTML # <markdowncell> # #Graphing Calculator - inspired by AP Calculus # Graphing calculators are permitted for use during "calculator active" problems on AP exams. Here we attempt to provide open access to those functionalities required for the AP exam. # # https://apstudent.collegeboard.org/apcourse/ap-calculus-ab/calculator-policy # # * Plot the graph of a function within an arbitrary viewing window # * Find the zeros of functions (solve equations numerically) # * Numerically calculate the derivative of a function # * Numerically calculate the value of a definite integral # <codecell> %%HTML <!DOCTYPE html> <html> <head> <style> </style> </head> <body> <div id="menu" style="width: 270px; float: left;"> <ul> <input class="txtbox" type="text" id="inputF" placeholder="Type a function f(x)" value="x^3+x^2-6*x" font-size: "10px"> <input class="btn" type="button" value="Plot" onClick="plotter()"> <!-- <input class="btn" type="button" value="add tangent" onClick="addTangent()"> --> <input class="btn" type="button" value="Add Derivative" onClick="plotDerivative()"> <input class="btn" type="button" value="Clear All" onClick="clearAll()"> </ul> <br></br> <ul> <input class="txtbox" type="text" id="inputZstart" placeholder="Start value for zero search" font-size: "6"> <input class="btn" type="button" value="find Zero" onClick="findZeroes()"> </ul> </div> <div id='jxgbox' class='jxgbox' style='width:450px; height:350px; margin-left: 180px; border: solid #1f628d 1px;'></div> <script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script> <script type="text/javascript" src="https://cdnjs.cloudflare.com/ajax/libs/jsxgraph/0.98/jsxgraphcore.js"></script> <script type='text/javascript'> var initBoundingBox = [-11,11,11,-11]; var board = JXG.JSXGraph.initBoard('jxgbox', {boundingbox:initBoundingBox, axis:true, showCopyright:false}); var f, curve; // global objects function plotter() { var txtraw = document.getElementById('inputF').value; f = board.jc.snippet(txtraw, true, 'x', true); curve = board.create('functiongraph',[f, function(){ var c = new JXG.Coords(JXG.COORDS_BY_SCREEN,[0,0],board); return c.usrCoords[1]; }, function(){ var c = new JXG.Coords(JXG.COORDS_BY_SCREEN,[board.canvasWidth,0],board); return c.usrCoords[1]; } ], {name:txtraw, withLabel:false}); var q = board.create('glider', [2, 3, curve], {withLabel:false}); var t = board.create('text', [ function(){ return q.X()+0.2; }, function(){ return q.Y()+0.1; }, //function(){ return "The slope of the function f(x)=" + txtraw + "<br>at x=" + q.X().toFixed(2) + " is equal to " + (JXG.Math.Numerics.D(f))(q.X()).toFixed(2); } //function(){ return "f(x)=" + txtraw + "<br>f(x=" + q.X().toFixed(2) + ") = " + f(q.X()).toFixed(2)} function(){ return "f(x=" + q.X().toFixed(2) + ") = " + f(q.X()).toFixed(2)} ], {fontSize:15}); } function clearAll() { JXG.JSXGraph.freeBoard(board); board = JXG.JSXGraph.initBoard('jxgbox', {boundingbox:initBoundingBox, axis:true, showCopyright:false}); f = null; curve = null; } function addTangent() { if (JXG.isFunction(f)) { board.suspendUpdate(); var p = board.create('glider',[1,0,curve], {name:'drag me'}); board.create('tangent',[p], {name:'drag me'}); board.unsuspendUpdate(); } } function plotDerivative() { if (JXG.isFunction(f)) { board.create('functiongraph',[JXG.Math.Numerics.D(f), function(){ var c = new JXG.Coords(JXG.COORDS_BY_SCREEN,[0,0],board); return c.usrCoords[1]; }, function(){ var c = new JXG.Coords(JXG.COORDS_BY_SCREEN,[board.canvasWidth,0],board); return c.usrCoords[1]; }], {dash:2}); } } function isNumeric(num){ return !isNaN(num) } function findZeroes() { var zeroraw = document.getElementById('inputZstart').value; if (JXG.isFunction(f) && isNumeric(zeroraw)) { board.suspendUpdate(); var zero = JXG.Math.Numerics.fzero(f,parseFloat(zeroraw)); var f_zero = f(zero); var p = board.create('point',[zero,f_zero], {name:'f(x='+zero.toFixed(2)+')=0.0', strokeColor:'gray', face:'<>',fixed:true}); //board.create('tangent',[p], {name:'drag me'}); board.unsuspendUpdate(); } } function findNumDerivative() { var zeroraw = document.getElementById('inputNumDerivative').value; if (JXG.isFunction(f) && isNumeric(zeroraw)) { board.suspendUpdate(); var zero = JXG.Math.Numerics.fzero(f,parseFloat(zeroraw)); var f_zero = f(zero); var p = board.create('point',[zero,f_zero], {name:'a zero of the function', strokeColor:'gray', face:'<>'}); //board.create('tangent',[p], {name:'drag me'}); board.unsuspendUpdate(); } } </script> </body> </html> # <markdowncell> # ### No Grading setup at this time. # <codecell> # <codecell>
6,762
2,010
# The MIT License (MIT) # Copyright (c) 2017 Massachusetts Institute of Technology # # Authors: Cody Rude # This software is part of the NSF DIBBS Project "An Infrastructure for # Computer Aided Discovery in Geoscience" (PI: V. Pankratius) and # NASA AIST Project "Computer-Aided Discovery of Earth Surface # Deformation Phenomena" (PI: V. Pankratius) # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN # THE SOFTWARE. # Standard library imports from collections import OrderedDict # scikit discovery imports from skdiscovery.data_structure.framework.base import PipelineItem # Pyinsar imports from pyinsar.processing.utilities.generic import coherence # Scikit data access imports from skdaccess.utilities.support import progress_bar class Coherence(PipelineItem): ''' Calculate coherence between single-look complex SAR images ''' def __init__(self, str_description, window, pairing='neighbor', use_progress_bar = False): ''' Initialize coherence pipeline item @param str_description: Short string describing item @param window: Tuple indicating the y and x window size @param pairing: How to pair slc images. "neighbor" computes coherence between neighboring images @param use_progress_bar: Display progress using a progress bar ''' self.window = window self.pairing = pairing self.use_progress_bar = use_progress_bar super(Coherence, self).__init__(str_description,[]) def process(self, obj_data): ''' Compute the coherency between two @param obj_data: Data wrapper ''' results_dict = OrderedDict() if self.pairing == 'neighbor': first_image_it = obj_data.getIterator() second_image_it = obj_data.getIterator() next(second_image_it) for (label1, image1), (label2, image2) in progress_bar(zip(first_image_it, second_image_it), total = len(obj_data)-1, enabled = self.use_progress_bar): results_dict[label2] = coherence(image1, image2, self.window) obj_data.addResult(self.str_description, results_dict)
3,294
963
# -*- coding: utf-8 -*- # # on_rtd is whether we are on readthedocs.org import os import sys sys.path.insert(0, os.path.abspath('.')) sys.path.append(os.path.abspath('extensions')) extensions = [ 'sphinx.ext.graphviz', 'sphinx.ext.todo', 'wikipedia', 'examplecode' ] on_rtd = os.environ.get('READTHEDOCS', None) == 'True' if not on_rtd: # only import and set the theme if we're building docs locally html_theme_path = ['doc/themes'] html_theme = 'sphinx_rtd_theme' # otherwise, readthedocs.org uses their theme by default, so no need to specify it # Using downloaded sphinx_rtd_theme # import sphinx_rtd_theme # html_theme = "sphinx_rtd_theme" # html_theme_path = [sphinx_rtd_theme.get_html_theme_path()] project = 'FIWARE-Stream-Oriented-GE' master_doc = 'index' html_context = { 'css_files': [ '_static/css/theme.css', 'https://fiware.org/style/fiware_readthedocs.css', 'https://fiware.org/style/fiware_readthedocs_media_streams.css' ] }
997
392
test_input = [4, 8] puzzle_input = [4, 9] def part1(): position = puzzle_input score = [0,0] die_roll = 0 num_rolls = 0 cur_player = 0 while True: die_total = 0 for _ in range(3): num_rolls += 1 die_roll += 1 if die_roll > 100: die_roll = 1 die_total += die_roll for _ in range(die_total): position[cur_player] += 1 if position[cur_player] > 10: position[cur_player] = 1 score[cur_player] += position[cur_player] if score[cur_player] >= 1000: print("Game over, p1 score", score[0], "p2 score", score[1], "magic number", num_rolls * min(score[0],score[1])) break if cur_player == 0: cur_player = 1 else: cur_player = 0 #part1() import collections def part2(): position = puzzle_input roll_dist = collections.Counter(i + j + k for i in range(1,4) for j in range(1,4) for k in range(1,4)) p1_wins = 0 p2_wins = 0 game_states = collections.defaultdict(int) game_states[(position[0], 0, position[1], 0)] = 1 while game_states: updated_game_states = collections.defaultdict(int) for game_state in game_states: for roll in roll_dist: (pos1, score1, pos2, score2) = game_state pos1 = (pos1 + roll - 1) % 10 + 1 score1 += pos1 if score1 >= 21: p1_wins += game_states[game_state] * roll_dist[roll] else: updated_game_states[(pos1, score1, pos2, score2)] += game_states[game_state] * roll_dist[roll] game_states = updated_game_states updated_game_states = collections.defaultdict(int) for game_state in game_states: for roll in roll_dist: (pos1, score1, pos2, score2) = game_state pos2 = (pos2 + roll - 1) % 10 + 1 score2 += pos2 if score2 >= 21: p2_wins += game_states[game_state] * roll_dist[roll] else: updated_game_states[(pos1, score1, pos2, score2)] += game_states[game_state] * roll_dist[roll] game_states = updated_game_states print("p1 wins", p1_wins, "p2 wins", p2_wins) part2()
2,372
798
from typing import Optional import numpy as np import matplotlib.pyplot as plt from matplotlib.axis import Axis from matplotlib.patches import Rectangle from superresolution import SIM_3D_Data def make_axis_if_none(ax: Optional[Axis]) -> Axis: if ax is None: _, ax = plt.subplots(1, 1, figsize = (10, 10)) return ax def plot_overlay(SIM_image: SIM_3D_Data, plt_axis: Optional[Axis] = None, projection_dim: int = 0, **imshow_kwargs) -> Axis: ax = make_axis_if_none(plt_axis) overlay = np.mean(SIM_image.data, axis = projection_dim) ax.imshow(overlay, **imshow_kwargs) ax.set_xlabel('x (pixels)') ax.set_ylabel('y (pixels)') fig = ax.get_figure() fig.tight_layout() return ax def plot_selection_box(x_range, y_range, ax, **rectangle_kwargs): origin = (x_range[0], y_range[0]) widths = (x_range[1]-x_range[0], y_range[1]-y_range[0]) rect = Rectangle(origin, *widths, **rectangle_kwargs) ax.add_patch(rect) return ax # TODO: add support for plotting slices along x and y axes as well. # Will need to use transpose to swap that dimension with zero and proceed with the rest of the logic def plot_slices(SIM_image: SIM_3D_Data, ncols: int, nrows: int, slice_dim: int = 0, **imshow_kwargs): fig, axes = plt.subplots(nrows, ncols, figsize = (10, 10)) num_slices = SIM_image.shape[slice_dim] plot_inds = np.linspace(0, num_slices, ncols*nrows, endpoint = False) plot_inds = np.int_(np.floor(plot_inds)) for i, ax in zip(plot_inds, axes.flat): ax.imshow(SIM_image.data[i], **imshow_kwargs) ax.set_title(f'Slice #{i}/{num_slices}') return fig, ax
1,666
628
#!/usr/bin/python2 import pyghmi.util.webclient as webclient import json import os import sys missingargs = False if 'XCCUSER' not in os.environ: print('Must set XCCUSER environment variable') missingargs = True if 'XCCPASS' not in os.environ: print('Must set XCCPASS environment variable') missingargs = True if missingargs: sys.exit(1) w = webclient.SecureHTTPConnection(sys.argv[1], 443, verifycallback=lambda x: True) w.connect() adata = json.dumps({'username': os.environ['XCCUSER'], 'password': os.environ['XCCPASS']}) headers = {'Connection': 'keep-alive', 'Content-Type': 'application/json'} w.request('POST', '/api/login', adata, headers) rsp = w.getresponse() if rsp.status == 200: rspdata = json.loads(rsp.read()) w.set_header('Content-Type', 'application/json') w.set_header('Authorization', 'Bearer ' + rspdata['access_token']) if '_csrf_token' in w.cookies: w.set_header('X-XSRF-TOKEN', w.cookies['_csrf_token']) print(repr(w.grab_json_response('/api/dataset', { 'USER_GlobalPassComplexRequired': '0', })))
1,097
382
from github import Github from my_token import token g = Github(token) repo = g.get_repo("adafruit/Adafruit_CircuitPython_Display_Text") repo_topics = repo.get_topics() print(repo_topics)
193
74
import hashlib import os import tempfile import numpy as np import fcswrite def test_write_fcs(): """test that fcm can read the data files""" fname = tempfile.mktemp(suffix=".fcs", prefix="write_test") data = 1.0*np.arange(400).reshape((100, 4)) chn_names = ['cha', 'chb', 'ch3', 'ch4'] # monkey-patch fcswrite version to have reproducible result oldver = fcswrite.__version__ fcswrite.fcswrite.version = "0.5.0" fcswrite.write_fcs(filename=fname, chn_names=chn_names, data=data ) # write back correct version fcswrite.fcswrite.version = oldver with open(fname, "rb") as fd: data = fd.read() data = np.frombuffer(data, dtype=np.uint8) # remove empty lines data = data[data != 8224] data = data.tostring() hasher = hashlib.md5() hasher.update(data) hexval = hasher.hexdigest() assert hexval == "2b4fdb7012b0693285c31aa91c606216" os.remove(fname) if __name__ == "__main__": # Run all tests loc = locals() for key in list(loc.keys()): if key.startswith("test_") and hasattr(loc[key], "__call__"): loc[key]()
1,200
449
from moviepy.editor import VideoFileClip from datetime import datetime import numpy as np import time import os def manage_time(timestamp): """ Given the string representation of a the time using the "minutes:seconds[:miliseconds]" representation, returns the number of seconds using double precision """ time_strip = timestamp.split(":") seconds = int(time_strip[0]) * 60 + int(time_strip[1]) # Add miliseconds if len(time_strip) == 3: seconds += int(time_strip[2]) / 60 return seconds def preprocess_video(filename, start, end, target_name, audio, codec=None): """ Preprocess an input video by cutting it given start time to end time, optionally removing the audio and changing video encoding """ # Load the video file clip = VideoFileClip(filename) # Calculate start and end points in seconds starting_point = manage_time(start) end_point = manage_time(end) # Resize the video and save the file subclip = clip.subclip(starting_point, end_point) subclip.write_videofile(target_name, audio=audio, codec=codec) def concatenate_videos(base="", ext="mp4"): stringa = 'ffmpeg -i "concat:' elenco_video = glob.glob(base + "*." + ext) elenco_file_temp = [] for f in elenco_video: file = "temp" + str(elenco_video.index(f) + 1) + ".ts" os.system( "ffmpeg -i " + f + " -c copy -bsf:v h264_mp4toannexb -f mpegts " + file ) elenco_file_temp.append(file) print(elenco_file_temp) for f in elenco_file_temp: stringa += f if elenco_file_temp.index(f) != len(elenco_file_temp) - 1: stringa += "|" else: stringa += '" -c copy -bsf:a aac_adtstoasc output.mp4' print(stringa) os.system(stringa)
1,810
604
#!/usr/bin/env python def get_localization(args): # TODO: Implement the localization fitting the centers pass
119
36
import snap G = snap.GenFull(snap.PNEANet, 100) # get a new random generator, provide the seed value Rnd = snap.TRnd(42) # randomize the generator, every execution will produce a different sequence. # Comment out the line to get the same sequence on every execution. Rnd.Randomize() for i in range(0,10): # provide the random generator as a parameter to the function NId = G.GetRndNId(Rnd) print(NId) # result is not well formed, the following statement fails #print(NI.GetId())
505
162
print("Thank you Jesus") # Read a value from standard input a value # input("Thank you") # Evaluate expression x = 1 print(x) x += 3 print(x) # loops if x > 1: print("great than 1") else: print("less than 1") n = 3 while n > 1: print(n) n -= 1 # Arithmetic operator print({100 % 3}, {100 / 3}) z = 1 print(z, type(z)) z = "dsf" print(z, type(z)) print(17 // 3) # floor division discards the fractional part print(5 ** 2) # 5 squared # use raw strings by adding an r before the first quote print(r'C:\some\name') # String literals can span multiple lines. One way is using triple-quotes: """...""" or '''...''' print("""\ dsfdsfds ddsf dfdf dfdfds """) w = 'thanks' print(w[:3] + w[3:]) # All slice operations return a new list containing the requested elements. # This means that the following slice returns a new (shallow) copy of the list: squares = [1, 4, 9, 16, 25] for n in squares: print(n, end='-') if n == 16: squares.insert(n, 100) print('\n') print(squares[:]) squares.append(10) print(squares[:]) squares = [] print(squares[:]) # Fibonacci a, b = 0, 1 while b < 25: print(b) a, b = b, a + b # Fibonacci a, b = 0, 1 while b < 1000: print(b, end=',') a, b = b, a + b print('\n') # Loops for i in range(3, 15, 4): print(i) a = ['Mary', 'had', 'a', 'little', 'lamb'] for i in range(len(a)): print(i, a[i]) print(list(range(5))) def f(ab, l=[]): l.append(ab) return l print(f(1)) print(f(2)) print(f(3)) # if __name__ == "__main__": # import sys # # print(int(sys.argv[1])) import sys print(dir(sys)) print('12'.zfill(5)) print('We are the {} who say "{}!"'.format('knights', 'Ni')) print('{0} and {1}'.format('spam', 'eggs')) # Formatting: https://docs.python.org/3.6/tutorial/inputoutput.html quit(1)
1,846
793