text
string
size
int64
token_count
int64
""" Lacework Run Reports API wrapper. """ import logging logger = logging.getLogger(__name__) class RunReportsAPI(object): """ Lacework RunReports API. """ def __init__(self, session): """ Initializes the RunReportsAPI object. :param session: An instance of the HttpSession class :return RunReportsAPI object. """ super(RunReportsAPI, self).__init__() self._session = session def run_report(self, type, account_id): """ A method to initiate a compliance assessment. :param type: A string representing the type of compliance assessment to initiate. :param account_id: A string representing the account identifier for which to initiate a compliance assessment. :return response json """ logger.info(f"Initiating '{type}' compliance assessment in Lacework...") # Build the Run Report request URI api_uri = f"/api/v1/external/runReport/{type}/{account_id}" response = self._session.post(api_uri) return response.json() def aws(self, aws_account_id): """ A method to initiate a compliance assessment for an AWS account. :param aws_account_id: A string representing which AWS account to assess. :return response json """ return self.run_report("aws", aws_account_id) def azure(self, azure_tenant_id): """ A method to initiate a compliance assessment for an Azure tenant. :param azure_tenant_id: A string representing which Azure tenant to assess. :return response json """ return self.run_report("azure", azure_tenant_id) def gcp(self, gcp_project_id): """ A method to initiate a compliance assessment for a GCP project. :param gcp_project_id: A string representing which GCP project to assess. :return response json """ return self.run_report("gcp", gcp_project_id) def integration(self, integration_guid): """ A method to run a compliance report based on a Lacework integration GUID. :param integration_guid: A string representing the Lacework integration ID to query. :return response json """ return self.run_report("integration", integration_guid)
2,447
665
admin_username = "richard" admin_password = "richard666"
56
21
import json class ConfigObject(dict): def __init__(self, **kwargs): super().__init__(**kwargs) self.load_attributes(kwargs) def load_attributes(self, d): for k, v in d.items(): if isinstance(v, dict): self[k] = self.__class__(**v) elif isinstance(v, list): vs = [] for i in v: if isinstance(i, dict): vs.append(self.__class__(**i)) else: vs.append(i) self[k] = vs def __getattr__(self, name): try: value = self[name] return value except Exception as _: return None def __setattr__(self, key, value): self[key] = value def __str__(self): return json.dumps(self, indent=4, ensure_ascii=False)
888
254
import discord from discord.ext import commands import os from discord.ext import tasks from datetime import datetime bot = commands.Bot(command_prefix='&') token = os.environ['DISCORD_BOT_TOKEN'] guild = bot.get_guild(417245684656373766) #ステータス @bot.event async def on_ready(): activity = discord.Activity(name='&help | Pornhub',type=discord.ActivityType.streaming) channel = bot.get_channel(417245684656373768) await bot.change_presence(activity=activity) await channel.send('おはようございます!') @bot.command() async def ping(ctx): await ctx.send('pong') #話すう @bot.command() async def rrr(ctx, arg1, arg2): if ctx.author.id != 540536805099831299: await ctx.send('君には使えないよ!') return channel = bot.get_channel(int(arg1)) await channel.send(arg2) #アウト @bot.command() async def out(ctx, arg:discord.Member): if ctx.author.id != 540536805099831299: await ctx.send('君には使えないよ!') return role = guild.get_role(714777128079720478) await arg.add_roles(role) await ctx.send(arg.mention+'を出禁にしました!') #ステータス変更 @bot.command() async def stats(ctx, arg): if ctx.author.id == 275574408372944897: arg1 = arg.replace('影','禿').replace('か','は').replace('カ','ハ').replace('k','h').replace('K','H').replace('K','H').replace('k','h') activity = discord.Activity(name='&help | '+arg1,type=discord.ActivityType.streaming) await bot.change_presence(activity=activity) await ctx.send('ステータスを '+arg1+'を配信中 にしました!') activity = discord.Activity(name='&help | '+arg,type=discord.ActivityType.streaming) await bot.change_presence(activity=activity) await ctx.send('ステータスを '+arg+'を配信中 にしました!') #上級VC変更 @bot.command() @commands.has_role('⚜️上級ロメダ民') async def lcname(ctx, arg): arg = arg.replace('影','禿').replace('か','は').replace('カ','ハ').replace('k','h').replace('K','H').replace('K','H').replace('k','h') channel = bot.get_channel(801398685828382751) await channel.edit(name='📍'+arg) await ctx.send('固定チャンネル名を📍'+arg+'にしました!') #鯖名変更 @bot.command() @commands.has_role('⚜️上級ロメダ民') async def sname(ctx, arg): guild = guild await guild.edit(name='&ROMEDA-'+arg) await ctx.send('サーバー名を&ROMEDA-'+arg+'にしました!') #ニックネーム変更 @bot.command() @commands.has_role('👑KING OF ROMEDA') async def nick(ctx, arg1:discord.Member, arg2): if arg1.id == 714776261410553907: await ctx.send('やめてください!') return arg = arg1.name + arg2 await arg1.edit(nick=arg) bot.run(token)
2,523
1,152
#=========================================================================== # # Copyright (c) 2014, California Institute of Technology. # U.S. Government Sponsorship under NASA Contract NAS7-03001 is # acknowledged. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are # met: # # 1. Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # 3. Neither the name of the copyright holder nor the names of its # contributors may be used to endorse or promote products derived from # this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT # HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. # #=========================================================================== """A library of top-level style functions. """ __version__ = "$Revision: #1 $" #=========================================================================== import os import os.path as path from .SubStyle import SubStyle #=========================================================================== __all__ = [ 'cleanupFilename', 'mergeDicts', 'resolveDefaults', 'stylePath', ] #=========================================================================== def cleanupFilename( fname ): """: Make the filename usable. = INPUT VARIABLES - fname Given a filename, clean-it up to make sure we can use it with the file system. = RETURN VALUE - Returns a cleaned up form of the input file name. """ fname = fname.replace( ' ', '_' ) fname = fname.replace( '/', '_' ) fname = fname.replace( '\\', '_' ) fname = fname.replace( '!', '_' ) fname = fname.replace( '*', '_' ) fname = fname.replace( '`', '_' ) fname = fname.replace( "'", "_" ) fname = fname.replace( '"', "_" ) fname = fname.replace( '{', '(' ) fname = fname.replace( '}', ')' ) fname = fname.replace( '&', '_and_' ) return fname #=========================================================================== # For internal use only. def mergeDicts( d1, d2 ): """: Recursively merge two dictionary data structures. This essentially performs a union of nested dictionary data """ r = {} r.update( d1 ) for key in d2: value = d2[ key ] if key in d1: if isinstance( value, SubStyle ): value = value.kwargs() if isinstance( value, dict ) and ( key in d1 ): other = d1[ key ] if isinstance( other, SubStyle ): other = other.kwargs() value = mergeDicts( other, value ) r[ key ] = value return r #=========================================================================== # For internal use only. def resolveDefaults( defaults, subNames = [], **kwargs ): """: Resolve a new set of defaults. What this funtion will do is: 1) Make a duplicate of the default dictionary to be modified and returned. 2) For each keyword-value parameter that is not set to None, that value will be set in the dictionary to be returned. If the value is itself a dictionary, then it will be "merged" into the return dictionary. 3) For each of the names specified by subNames that exists in the default dictionary, its values will be set in the dictionary to be returned. If the value is itself a dictionary, then it will be "merged" into the return dictionary. It is important to note that the order of the names specified in 'subNames' is important as that is the order in which they are resolved. 4) Returns the return dictionary. When a dictionary 'A' is "merged" into another dictionary 'B', this is much like the built-in dictionary 'update' method ( 'B.update( A )' ). The difference is that any value in 'A' that is set to None is not 'updated' in 'B' and for any values that are themselves dictionaries, then they will be "merged". = INPUT VARIABLES - defaults The current set of default values to resolve with. - subNames A list of names of sub-properties to resolve (in the order to resolve them in). - kwargs Optional keyword arguments to also resolve. = RETURN VALUE - Return a new dictionary of default values. """ # First duplicate the given defaults subDefaults = {} subDefaults.update( defaults ) # Next add in any keyword arguments for key in kwargs: value = kwargs[ key ] # If the kw value is not set, then ignore if value is None: continue # We have a kw value and nothing has been set yet. if isinstance( value, SubStyle ): value = value.kwargs() if isinstance( value, dict ) and ( key in subDefaults ): other = subDefaults[ key ] if isinstance( other, SubStyle ): other = other.kwargs() value = mergeDicts( other, value ) # Store the value subDefaults[ key ] = value for name in subNames: if name in defaults: tmp = defaults[ name ] if tmp is None: continue if isinstance( tmp, SubStyle ): tmp = tmp.kwargs() if isinstance( tmp, dict ): subDefaults = mergeDicts( subDefaults, tmp ) else: subDefaults[ name ] = tmp return subDefaults #=========================================================================== def stylePath( envvar = 'STYLEPATH' ): """: Get the value of the STYLEPATH environment variable = INPUT VARIABLE - envvar The name of the environment variable to use for the style path. = RETURN VALUE - Return a list of paths as defined by the STYLEPATH environment variable. """ result = [] if envvar.startswith( '$' ): envvar = envvar[ 1: ] stylepath = os.getenv( envvar, "" ) stylepath = stylepath.split( ':' ) for directory in stylepath: if len( directory.strip() ) == 0: continue p = path.normpath( path.expanduser( path.expandvars( directory ) ) ) result.append( p ) return result #===========================================================================
7,225
2,097
import math import pylab from app import StandardRobot, LeastDistanceRobot, RandomWalkRobot, runSimulation def timeNumberPlot(title, x_label, y_label, dim_length): """ Plots the relation between the number of robots and the average time taken by different robots to clean a portion of the room. """ num_robot_range = range(1, 11) times1, times2, times3 = ([] for i in range(3)) time_Robot_list = [times1, times2, times3] Robots = [StandardRobot, LeastDistanceRobot, RandomWalkRobot] for i in range(len(Robots)): for num_robots in num_robot_range: result = runSimulation( num_robots, 1.0, dim_length, dim_length, 1, 100, Robots[i]) time_Robot_list[i].append(result[0]) for time in time_Robot_list: pylab.plot(num_robot_range, time) pylab.title(title+f"\n for room size of {dim_length}x{dim_length}") pylab.legend(('StandardRobot', 'LeastDistanceRobot', 'RandomWalkRobot')) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def timeAspectRatioPlot(title, x_label, y_label, area, num_robots): """ Plots the relation between the area of a square room and the average and the average time taken by diffrent robots to clean a portion of th room. """ aspect_ratios = [] times1, times2, times3 = ([] for i in range(3)) time_Robot_list = [times1, times2, times3] Robots = [StandardRobot, LeastDistanceRobot, RandomWalkRobot] start = math.sqrt(area) aspect_dim_list = [] for dim in range(1, 11): aspect_dim_list.append(start*dim) for width in aspect_dim_list: height = area / width aspect_ratios.append("1 : {0}".format(int(width/height))) for i in range(len(Robots)): result = runSimulation( num_robots, 1.0, int(width), int(height), 1, 100, Robots[i]) time_Robot_list[i].append(result[0]) for time in time_Robot_list: pylab.plot(aspect_ratios, time) pylab.title( title+f"\n for {num_robots} Robots & Area of {area}") pylab.legend(('StandardRobot', 'LeastDistanceRobot', "RandomWalkRobot")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def wasteAspectRatioPlot(title, x_label, y_label, num_robots, area): """ Plots the relation between room's aspect ratio and the waste percentage for each robot. """ aspect_ratios = [] times1, times2, times3 = ([] for i in range(3)) waste_Robot_list = [times1, times2, times3] Robots = [StandardRobot, LeastDistanceRobot, RandomWalkRobot] start = math.sqrt(area) aspect_dim_list = [] for dim in range(1, 11): aspect_dim_list.append(start*dim) for width in aspect_dim_list: height = area / width aspect_ratios.append("1 : {0}".format(int(width/height))) for i in range(len(Robots)): result = runSimulation( num_robots, 1.0, int(width), int(height), 1, 100, Robots[i]) waste_Robot_list[i].append(result[1]/result[0]) for time in waste_Robot_list: pylab.plot(aspect_ratios, time) pylab.title( title+"\n for {0} Robots & Area of {1}".format(num_robots, area)) pylab.legend(('StandardRobot', 'LeastDistanceRobot', "RandomWalkRobot")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def timeAreaPlot(title, x_label, y_label, num_robots): """ Plots the relation between the area of a room and the average time taken by different robots to clean a certain portion of that room. """ dim_length_range = range(5, 31, 5) times1, times2, times3 = ([] for i in range(3)) time_Robot_list = [times1, times2, times3] Robots = [StandardRobot, LeastDistanceRobot, RandomWalkRobot] for i in range(len(Robots)): for dim_length in dim_length_range: result = runSimulation( num_robots, 1.0, dim_length, dim_length, 1, 100, Robots[i]) time_Robot_list[i].append(result[0]) for time in time_Robot_list: pylab.plot(dim_length_range, time) pylab.title(title+"\n for {0} Robots".format(num_robots)) pylab.legend(('StandardRobot', 'LeastDistanceRobot', "RandomWalkRobot")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def consistencyPlot(title, x_label, y_label, dim_length, num_robots): """ Performs the same exact experiment multiple of times for a robot or a number of robots and plots the outcomes of these experiments in terms of time taken for each individual experiment to measure the consistency of performance for various robots. """ try_num_range = range(16) times1, times2, times3 = ([] for i in range(3)) time_Robot_list = [times1, times2, times3] Robots = [StandardRobot, LeastDistanceRobot, RandomWalkRobot] for i in range(len(Robots)): for try_num in try_num_range: result = runSimulation( num_robots, 1.0, dim_length, dim_length, 1, 1, Robots[i]) time_Robot_list[i].append(result[0]) for time in time_Robot_list: pylab.plot(try_num_range, time) pylab.title( title+"\n for size of {0}x{0} & {1} Robots".format(dim_length, num_robots)) pylab.legend(('StandardRobot', 'LeastDistanceRobot', "RandomWalkRobot")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def timeAreaPortionPlot(title, x_label, y_label, num_robots, robot_type): """ Plots the relation between the area of square room and the average time taken by a robot to clean a specific portion of the room, (Different portions are plotted) """ dim_length_range = range(5, 31, 5) coverage_percent_range = range(70, 105, 5) coverage_percent_range = [i/100 for i in coverage_percent_range] alist, blist, clist, dlist, elist, flist, glist = ([] for i in range(7)) coverage_percent_list = [alist, blist, clist, dlist, elist, flist, glist] for dim_length in dim_length_range: for i in range(len(coverage_percent_range)): result = runSimulation( num_robots, 1.0, dim_length, dim_length, coverage_percent_range[i], 100, robot_type) coverage_percent_list[i].append(result[0]) for percentlist in coverage_percent_list: pylab.plot(dim_length_range, percentlist) pylab.title( title+f"\n for {num_robots} bots of {robot_type.__name__} type") pylab.legend(("0.7", "0.75", "0.8", "0.85", "0.9", "0.95", "1.0")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() # Not sure about it def CoverageWasteRatio(title, x_label, y_label, num_robots, robot_type): """ efficiency """ dim_length_range = range(5, 31, 5) coverage_percent_range = range(70, 105, 5) coverage_percent_range = [i/100 for i in coverage_percent_range] alist, blist, clist, dlist, elist, flist, glist = ([] for i in range(7)) coverage_percent_list = [alist, blist, clist, dlist, elist, flist, glist] for dim_length in dim_length_range: for i in range(len(coverage_percent_range)): result = runSimulation( num_robots, 1.0, dim_length, dim_length, coverage_percent_range[i], 100, robot_type) coverage_percent_list[i].append(result[1]/result[0]) for percentlist in coverage_percent_list: pylab.plot(dim_length_range, percentlist) pylab.title( title+f"\n for {num_robots} bots of {robot_type.__name__} type") pylab.legend(("0.7", "0.75", "0.8", "0.85", "0.9", "0.95", "1.0")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def timeNumberPortionPlot(title, x_label, y_label, dim_length, robot_type): """ Plots the relation between the number of robots and the average time taken to clean a certain portion of the room, (each portion is plotted) """ num_robot_range = range(1, 11) coverage_percent_range = range(70, 105, 5) coverage_percent_range = [i/100 for i in coverage_percent_range] alist, blist, clist, dlist, elist, flist, glist = ([] for i in range(7)) coverage_percent_list = [alist, blist, clist, dlist, elist, flist, glist] for num_robots in num_robot_range: for i in range(len(coverage_percent_range)): result = runSimulation( num_robots, 1.0, dim_length, dim_length, coverage_percent_range[i], 100, robot_type) coverage_percent_list[i].append(result[0]) for percentlist in coverage_percent_list: pylab.plot(num_robot_range, percentlist) pylab.title( title+f"\n for {num_robots} bots of {robot_type.__name__} type") pylab.legend(("0.7", "0.75", "0.8", "0.85", "0.9", "0.95", "1.0")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() def wasteAreaNumberPlot(title, x_label, y_label): """ Plots the relation between the waste percentage and the area of the room for a different number of robots, (each plotted individually) """ alist, blist, clist, dlist, elist = ([] for i in range(5)) num_robots_list = [alist, blist, clist, dlist, elist] t1list, t2list, t3list, t4list, t5list = ([] for i in range(5)) time_robots_list = [t1list, t2list, t3list, t4list, t5list] dim_length_range = range(10, 51, 5) num_robots_range = range(5, 26, 5) for dim_length in dim_length_range: for i in range(len(num_robots_list)): results = runSimulation( num_robots_range[i], 1.0, dim_length, dim_length, 1, 100, LeastDistanceRobot) num_robots_list[i].append(results[1]/results[0]) time_robots_list[i].append(results[0]) for i in range(len(num_robots_range)): pylab.plot(dim_length_range, num_robots_list[i]) pylab.title(title) pylab.legend(('5', '10', "15", "20", "25")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() # Not sure about it def NumOfBotSizeWasteRatio(title, x_label, y_label, dim_length): """ What information does the plot produced by this function tell you? waste on average with num on bots to find the sweet spot for this room another function must be madfe with """ alist, blist, clist, dlist, elist = ([] for i in range(5)) num_robots_list = [alist, blist, clist, dlist, elist] t1list, t2list, t3list, t4list, t5list = ([] for i in range(5)) time_robots_list = [t1list, t2list, t3list, t4list, t5list] num_robots_range = range(5, 26, 5) for i in range(len(num_robots_list)): results = runSimulation( num_robots_range[i], 1.0, dim_length, dim_length, 1, 50, LeastDistanceRobot) num_robots_list[i].append(results[1]/results[0]) time_robots_list[i].append(results[0]) for i in range(len(num_robots_list)): pylab.scatter(time_robots_list[i], num_robots_list[i], s=100) pylab.title(title) pylab.legend(('5', '10', "15", "20", "25")) pylab.xlabel(x_label) pylab.ylabel(y_label) pylab.show() # Plots # timeNumberPlot('Number of robots & Time relation', # 'Number of robots', 'Time (Tick)', 20) # timeAspectRatioPlot('Aspect Ratio & Time relation', # 'Room Aspect Ratio', 'Time', 100, 1) # wasteAspectRatioPlot("AspectRatio & WasteRatio relation", # "Room Aspect Ratio", "Waste Percentage", 1, 100) # timeAreaPlot("Time & Room Area relation", # "Length (squared)", "Time", 1) # consistencyPlot("Consistency", "Try number", "Time", 20, 1) # timeAreaPortionPlot('Room Portion & Time relation', # 'Length (squared)', 'Time', 1, RandomWalkRobot) # # CoverageWasteRatio('Coverage Percent & Size Relation','Length (squared)', # # 'Waste Percentage', 5, LeastDistanceRobot) # timeNumberPortionPlot("CostQualityTime", "Number of robots", # "Time", 10, LeastDistanceRobot) # wasteAreaNumberPlot('Waste & Size Relation', 'Length (squared)', # 'Waste Percentage') # # NumOfBotSizeWasteRatio('Waste & Size & Num of bots Relation\n LeastDistanceRobot', # # 'Time', ' Waste Percentage', 20)
12,285
4,477
""" Ingest Package Config """ # The list of entities that will be loaded into the target service. These # should be class_name values of your target API config's target entity # classes. target_service_entities = [ "family", "participant", "diagnosis", "phenotype", "outcome", "biospecimen", "read_group", "sequencing_experiment", "genomic_file", "biospecimen_genomic_file", "sequencing_experiment_genomic_file", "read_group_genomic_file", ] # All paths are relative to the directory this file is in extract_config_dir = "extract_configs" transform_function_path = "transform_module.py" # TODO - Replace this with your own unique identifier for the project. This # will become CONCEPT.PROJECT.ID during the Load stage. project = "SD_ME0WME0W"
795
249
#!/usr/bin/env python """ Written by Anthony Nguyen 2015/06/20 """ import sqlite3 import sys #import smap_analytics as smap_analytics open_this_file = None if len(sys.argv) == 2: open_this_file = sys.argv[1] else: open_this_file = "config/smap_2015.csv" with sqlite3.connect('dashboard.db') as conn: c = conn.cursor() # Remove any existing house_layout table command = '''DROP TABLE IF EXISTS house_layout''' print command c.execute(command) # Create a new house_layout table command = '''CREATE TABLE house_layout (path VARCHAR(100) PRIMARY KEY, uuid VARCHAR(36) NOT NULL UNIQUE, heat_map_enable BOOLEAN, x_coord INT NOT NULL, y_coord INT NOT NULL, room VARCHAR(16), description VARCHAR(255), tab_type VARCHAR(16), channel_units VARCHAR(16))''' print command c.execute(command) #Parse the file of UUIDs and add each as a row with open(open_this_file) as f: for line in f.readlines(): row = line.split(",") #print row uuid = row[0] path = row[1] heat_map_enable = row[2] x_coord = int(row[3]) y_coord = int(row[4]) room = row[5] description = row[6] tab_type = row[7] channel_units = row[8] # Insert a row of data command = '''INSERT INTO house_layout VALUES ('%s','%s','%s','%d','%d','%s','%s','%s', '%s')'''%(path, uuid, heat_map_enable, x_coord, y_coord, room, description, tab_type, channel_units) print command c.execute(command) # Print the whole table at the end to make sure it works c.execute("SELECT * FROM house_layout") print c.fetchall() # Save (commit) the changes conn.commit()
1,877
608
""" Script to perform RC sorting Wren Saylor July 5 2017 Copyright 2017 Harvard University, Wu Lab Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import argparse import pandas as pd from FangsLibrary import run_sliding_window_for_each_nucleotide_string from ElementLibrary import get_bedtools_features from MethylationLibrary import collect_methylation_data_by_element import GlobalVariables # Methylation RCsorting def sort_methylation_by_directionality(negStr,posStr): posMeth = collect_methylation_data_by_element(posStr) negMeth = collect_methylation_data_by_element(negStr) # Zip reversed range to make a dictionary for replacing the location of the neg methylation originalRange = range(0,GlobalVariables.num) reverseRange = originalRange[::-1] rangeDict = dict(zip(originalRange,reverseRange)) # Zip reverse complement sequence for replacing the nucleotides for neg methylation seqDict = {'A':'T','T':'A','C':'G','G':'C','N':'N'} # Convert neg Meth df negMeth['methLocNew'] = negMeth.methLoc.map(rangeDict) negMeth['CytosineNew'] = negMeth.Cytosine.map(seqDict) negMeth['ContextNew'] = negMeth.Context.map(seqDict) negMethNew = negMeth[['id','methLocNew','methPer','methCov','methFreq','CytosineNew','ContextNew','tissue']] negMethNew.columns = ['id','methLoc','methPer','methCov','methFreq','Cytosine','Context','tissue'] # Concat pos and revised neg meth dfs frames = [posMeth,negMethNew] catMerge = pd.concat(frames) # Update Frequencey count column catMerge['methFreqNew'] = catMerge.groupby(['methLoc','tissue','Cytosine'])['methLoc'].transform('count') outMerge = catMerge[['id','methLoc','methPer','methCov','methFreqNew','Cytosine','Context','tissue']] outMerge.columns = ['id','methLoc','methPer','methCov','methFreq','Cytosine','Context','tissue'] return outMerge # Sliding window RCsorting def sort_sliding_window_by_directionality(negStr,posStr): negDF, negNames = run_sliding_window_for_each_nucleotide_string(negStr['reverseComplement'],negStr['id']) posDF, posNames = run_sliding_window_for_each_nucleotide_string(posStr['combineString'],posStr['id']) compWindow = [] for x, y in zip(negDF, posDF): tempCat = pd.concat([x,y],axis=1) tempGroup = tempCat.groupby(tempCat.columns,axis=1).sum() compWindow.append(tempGroup) return compWindow, negNames # Separate on plus and minus orientation, RCsort and return methylation and sliding window computations def sort_elements_by_directionality(directionFeatures): negStr = (directionFeatures[(directionFeatures['compareBoundaries'] == '-')]) posStr = (directionFeatures[(directionFeatures['compareBoundaries'] == '+')]) compWindow, compNames = sort_sliding_window_by_directionality(negStr,posStr) if any(x in GlobalVariables.graphs for x in ['methylation','cluster','methextend']): groupMeth = sort_methylation_by_directionality(negStr,posStr) else: groupMeth = None return groupMeth,compWindow,compNames def main(directionFeatures): groupMeth,compWindow,compNames = sort_elements_by_directionality(directionFeatures) print 'Completed reverse complement sorting for {0} items, with {1} bin sorting'.format(len(directionFeatures.index),GlobalVariables.binDir) return groupMeth,compWindow,compNames if __name__ == "__main__": main()
3,746
1,278
import sys import math # Auto-generated code below aims at helping you parse # the standard input according to the problem statement. n = int(input()) # the number of temperatures to analyse temps = input() # the n temperatures expressed as integers ranging from -273 to 5526 temps = temps.split() if n == 0: print(0) else: temp_ref = int(temps[0]); for temp in temps: temp = int(temp) if abs(temp) < abs(temp_ref): temp_ref = temp elif abs(temp) == abs(temp_ref) and temp > 0: temp_ref = temp; print(temp_ref)
580
185
import numpy as np import cv2 import matplotlib.pyplot as plt from openvino.inference_engine import IECore import openvino_model_experiment_package as omep # Load an IR model model = 'intel/human-pose-estimation-0001/FP16/human-pose-estimation-0001' ie, net, exenet, inblobs, outblobs, inshapes, outshapes = omep.load_IR_model(model) # Load an image and run inference img_orig = cv2.imread('people.jpg') res = omep.infer_ocv_image(exenet, img_orig, inblobs[0]) img = cv2.cvtColor(img_orig, cv2.COLOR_BGR2RGB) plt.imshow(img) plt.show() omep.display_heatmap(res['Mconv7_stage2_L2'], overlay_img=img_orig, statistics=False) omep.display_heatmap(res['Mconv7_stage2_L1'], statistics=False)
694
285
import large_image import urllib import pytest @pytest.mark.parametrize("item, output", [ ('590346ff8d777f16d01e054c', '/tmp/Huron.Image2_JPEG2K.tif') ]) def test_tiff_tile_source(item, output): """Check whether large_image can return a tile with tiff sources.""" test_url = 'https://data.kitware.com/api/v1/item/{}/download'.format(item) urllib.urlretrieve(test_url, output) image = large_image.getTileSource(output) # Make sure it is the tiff tile source assert isinstance(image, large_image.tilesource.TiffFileTileSource) # Make sure we can get a tile without an exception assert type(image.getTile(0, 0, 0)) == str
657
245
import os import sys import pickle import traceback from XSTAF.core.logger import LOGGER class ToolManager(object): def __init__(self): self.settings = {"ToolsLocation" : r"tools", "ToolsConfigureFile" : "config.pickle"} self.tool_name_list = [] self.abs_tools_location = "" self.pickle_config_file = "" def apply_settings(self, **kwargs): for arg in kwargs.items(): if arg[0] in self.settings: self.settings[arg[0]] = arg[1] def config(self): tools_location = self.settings["ToolsLocation"] if not os.path.isabs(tools_location): #check if tools location exist in XSTAF path XSTAF_path = os.path.dirname(os.path.abspath(__file__)) self.abs_tools_location = os.path.join(XSTAF_path, "..", tools_location) else: self.abs_tools_location = tools_location #try get tools name list from pickle file self.pickle_config_file = os.path.join(self.abs_tools_location, self.settings["ToolsConfigureFile"]) self.load_config() def get_tool(self, tool_module_name): if not os.path.isdir(self.abs_tools_location): LOGGER.warning("Can not find tools location: %s", self.abs_tools_location) return None #append abs tools path to python lib path #so we can dynamic import them sys.path.append(self.abs_tools_location) try: tool_module = __import__(tool_module_name) #want to reload the tool if tool is updated tool_module = reload(tool_module) tool = tool_module.Tool except (ImportError, AttributeError) as e: LOGGER.info("Can not import tool: %s" % tool_module_name) LOGGER.debug(traceback.format_exc()) return None else: return tool def load_config(self): if not os.path.isfile(self.pickle_config_file): LOGGER.warning("Can not find config file: %s", self.pickle_config_file) return #we load tool names from pickle file if os.path.isfile(self.pickle_config_file): with open(self.pickle_config_file, 'r') as f: self.tool_name_list = pickle.load(f) def save_config(self): if not os.path.isdir(self.abs_tools_location): LOGGER.warning("Can not find tools location: %s", self.abs_tools_location) return None #we save current tool names to pickle file with open(self.pickle_config_file, 'w') as f: pickle.dump(self.tool_name_list, f) @property def available_tool_name_list(self): if not os.path.isdir(self.abs_tools_location): LOGGER.warning("Can not find tools location: %s", self.abs_tools_location) while False: yield None else: #check all packages under abs_tools_location for name in os.listdir(self.abs_tools_location): #only check dirs abs_name = os.path.join(self.abs_tools_location, name) if os.path.isdir(abs_name): if not(name in self.tool_name_list) and not(self.get_tool(name) is None): yield name
3,359
971
from django import forms from .models import Image, Profile, Comments class NewsLetterForm(forms.Form): your_name = forms.CharField(label='First Name', max_length=30) email = forms.EmailField(label='Email') class NewImageForm(forms.ModelForm): class Meta: model = Image exclude = ['profile', 'pub_date'] widgets = { 'tags':forms.CheckboxSelectMultiple(), } class ProfileForm(forms.ModelForm): class Meta: model = Profile exclude = ['user'] class ImageUpload(forms.ModelForm): class Meta: model = Image exclude = ['pub_date', 'profile'] class CommentForm(forms.ModelForm): class Meta: model = Comments exclude = ['pub_date', 'image', 'profile'] fields = ['comments'] widgets = { 'comments':forms.TextInput(attrs={ 'class': u'comments-input form-control', 'placeholder': u'Insert Comment'}) } class profileEdit(forms.Form): name = forms.CharField(max_length=20) username = forms.CharField(max_length=20) Bio = forms.Textarea() Email = forms.EmailField() phone_number = forms.CharField(max_length=12)
1,200
347
string_input = "amazing" vowels = "aeiou" answer = [char for char in string_input if char not in vowels] print(answer)
120
47
""" Test serializer """ import unittest import glob import json try: from cStringIO import StringIO except ImportError: from io import StringIO from pyingest.parsers.iop import IOPJATSParser from pyingest.serializers.classic import Tagged from pyingest.serializers.refwriter import * class TestClassic(unittest.TestCase): def setUp(self): stubdata_dir = os.path.join(os.path.dirname(__file__), 'data/stubdata') self.inputdocs = glob.glob(os.path.join(stubdata_dir, 'parsed/*.json')) self.outputdir = os.path.join(stubdata_dir, 'serialized') # sys.stderr.write("test cases are: {}\n".format(self.inputdocs)) def test_classic_tagged(self): serializer = Tagged() for file in self.inputdocs: # this will raise exceptions if something is wrong document = '' with open(file, 'r') as fp: document = json.load(fp) self.assertIsNotNone(document, "%s: error reading doc" % file) outputfp = StringIO() serializer.write(document, outputfp) output = outputfp.getvalue() outputfp.close() self.assertNotEqual(output, '') basefile, _ = os.path.splitext(os.path.basename(file)) target = os.path.join(self.outputdir, basefile + '.tag') # save temporary copy target_saved = target + '.parsed' with open(target_saved, 'w') as fp: fp.write(output) ok = False # Python 3 orders the properties dictionary differently if sys.version_info > (3,) and os.path.exists(os.path.join(self.outputdir, 'python3', basefile + '.tag')): target = os.path.join(self.outputdir, 'python3', basefile + '.tag') if os.path.exists(target): with open(target, 'r') as fp: shouldbe = fp.read() self.assertEqual(shouldbe, output, "results differ from %s" % target) ok = True else: sys.stderr.write("could not find shouldbe file %s\n" % target) if ok: os.remove(target_saved) else: sys.stderr.write("parsed output saved to %s\n" % target_saved) class TestReferenceWriter(unittest.TestCase): def setup(self): pass def test_write_refhandler_data(self): paperdata = IOPJATSParser() inputdoc = 'pyingest/tests/data/stubdata/input/iop_apj.xml' with open(inputdoc, 'r') as fm: pdat = paperdata.parse(fm) if 'refhandler_list' in pdat: refwriter = ReferenceWriter() refwriter.topdir = 'pyingest/tests/data/output/' refwriter.refsource = '.jats.iopft.xml' refwriter.writeref(pdat) self.assertEqual('1', '1') else: self.assertEqual('a', 'b') # I changed these to let go with a pass instead of raising an exception # # def test_no_refdata(self): # refwriter = ReferenceWriter() # with self.assertRaises(NoReferencesException): # refwriter.writeref({}) # def test_no_metadata(self): # refwriter = ReferenceWriter() # bogus_data = {'refhandler_list': ['fnord']} # with self.assertRaises(WriteErrorException): # refwriter.writeref(bogus_data)
3,404
1,020
#!/usr/bin/env python # Code Listing #4 """ Maximum subarray problem - final version """ from contextlib import contextmanager import random import time # 復習 @contextmanager def timer(): """ Measure real-time execution of a block of code """ try: start = time.time() yield finally: end = (time.time() - start)*1000 print('time taken=> %.2f ms' % end) def num_array(size): """ Return a list of numbers in a fixed random range of given size """ nums = [] for i in range(size): nums.append(random.randrange(-25, 30)) return nums def max_subarray1(sequence): """ Find sub-sequence in sequence having maximum sum """ # this is the version before the final version for testing purposes max_sum, max_sub = 0, [] for i in range(len(sequence)): for j in range(i+1, len(sequence)): sub_seq = sequence[i:j+1] sum_s = sum(sub_seq) if sum_s > max_sum: max_sum, max_sub = sum_s, sub_seq return max_sum, max_sub def max_subarray(sequence): """ Maximum subarray - optimized version """ max_ending_here = max_so_far = 0 for x in sequence: max_ending_here = max(0, max_ending_here + x) max_so_far = max(max_so_far, max_ending_here) return max_so_far if __name__ == "__main__": #with timer(): # max_subarray1(num_array(10000)) print(max_subarray([-5, 20, -10, 30, 15])) # 55
1,473
513
""" One-time migration script from sqlalchemy models and sqlite database to custom ORM & PostgreSQL. Not designed to work as part of the regular alembic system, merely placed here for archive purposes. Should never need to run this again. 2021-05-03 """ from datetime import datetime, timedelta import sqlite3 from data.post_data import PostData, PostModel from data.snapshot_data import SnapshotData, SnapshotModel, SnapshotFrontpageModel from data.user_data import UserData from services import post_service from utils.logger import logger from utils.reddit import base36decode _post_data = PostData() _snapshot_data = SnapshotData() _user_data = UserData() DB_FILE = "src/database.db" def migrate_posts(offset=0): """Grabs posts in batches of 1000 at a time and migrates them to the new database. Returns number of processed rows. If less than 1000, at end of the table.""" conn = sqlite3.connect(DB_FILE) conn.row_factory = sqlite3.Row rows = conn.execute("SELECT * FROM posts LIMIT 1000 OFFSET ?;", (offset,)).fetchall() conn.close() row = None for row in rows: # If the post already exists in the database we don't need to do anything. post_id36 = row["id"] post = post_service.get_post_by_id(post_id36) if post: continue # OH RIGHT NO USER DATA IS SAVED IN THE OLD DATABASE. # username = row["name"] # if not user_service.get_user(username): # user = UserModel() # user.username = username # _user_data.insert(user, error_on_conflict=False) post = PostModel() post.set_id(post_id36) # post.author = username post.title = row["title"] post.created_time = row["created_time"] post.flair_text = row["flair"] # will add flair id in later mass update/backfill.. and user info _post_data.insert(post, error_on_conflict=False) if not row: logger.warning("No rows processed!") else: logger.info(f"Most recent migrated row: psk={row['psk']}, id={row['id']}") return len(rows) def migrate_snapshots(date, hour): conn = sqlite3.connect(DB_FILE) conn.row_factory = sqlite3.Row row = conn.execute("SELECT * FROM snapshots WHERE date=? and hour=?;", (date, hour)).fetchone() # No data, past the last recorded snapshot? if not row: return old_snapshot_psk = row["psk"] snapshot = SnapshotModel() snapshot.created_time = row["datetime"] snapshot.date = date snapshot.hour = hour snapshot.subscribers = row["subscribers"] new_snapshot = _snapshot_data.insert(snapshot) rows = conn.execute( "SELECT sf.*, p.id FROM snapshot_frontpage sf JOIN posts p on sf.post_psk = p.psk WHERE snapshot_psk=?;", (old_snapshot_psk,), ).fetchall() conn.close() for row in rows: sfp_model = SnapshotFrontpageModel() sfp_model.post_id = base36decode(row["id"]) sfp_model.snapshot_id = new_snapshot.id sfp_model.rank = row["rank"] sfp_model.score = row["score"] _snapshot_data.insert(sfp_model) def main(): current_offset = 0 while True: processed_posts = migrate_posts(current_offset) current_offset += processed_posts if processed_posts < 1000: break if current_offset % 1000 == 0: logger.info(f"Migrated {current_offset} posts total") current_datetime = datetime.fromisoformat("2020-05-12 04:00:00.000") now = datetime.utcnow() while current_datetime <= now: try: migrate_snapshots(current_datetime.date(), current_datetime.hour) except Exception: logger.exception(f"Failed to migrate {current_datetime.date()} - {current_datetime.hour}") current_datetime += timedelta(hours=1) if current_datetime.hour == 0: logger.info(f"Finished migrating {current_datetime.date()}") if __name__ == "__main__": main()
4,012
1,292
# fname = raw_input("Enter file name: ") fname = "romeo.txt" fh = open(fname) lst = list() def add_to_list(words): for word in words: if word not in lst: lst.append(word) for line in fh: words = line.split() add_to_list(words) lst.sort() print lst
291
110
from datasets.dataset_setup import DatasetSetup from models import read_data_text class YahooSetup(DatasetSetup): def __init__(self): super().__init__() self.num_classes = 10 self.size_bottom_out = 10 def set_datasets_for_ssl(self, file_path, n_labeled, party_num): train_labeled_dataset, train_unlabeled_dataset, val_dataset, test_dataset, n_labels = \ read_data_text.get_data(file_path, int(n_labeled / 10)) train_complete_labeled_dataset, _, _, _, _ = \ read_data_text.get_data(file_path, 5000) print("#Labeled:", len(train_labeled_dataset), "#Unlabeled:", len(train_unlabeled_dataset)) return train_labeled_dataset, train_unlabeled_dataset, test_dataset, train_complete_labeled_dataset def get_transformed_dataset(self, file_path, party_num=None, train=True): if train: train_complete_labeled_dataset, _, _, _, _ = \ read_data_text.get_data(file_path, 5000) return train_complete_labeled_dataset else: _, _, _, test_dataset, _ = \ read_data_text.get_data(file_path, 10) return test_dataset if __name__ == '__main__': dataset_setup = YahooSetup() train_dataset = dataset_setup.get_transformed_dataset(file_path='D:/Datasets/yahoo_answers_csv/',train=True) print("s")
1,376
453
from mfl_playoff_leagues import MFL def run(request): """Responds to any HTTP request. Args: request (flask.Request): HTTP request object. Returns: The response text or any set of values that can be turned into a Response object using `make_response <http://flask.pocoo.org/docs/1.0/api/#flask.Flask.make_response>`. """ # get the query string parsed as a dict request_json = request.get_json(silent=True) args = request.args m = MFL(year=args['year'], league=args['league']) if request.args and 'live_scoring' in request.args: return m.live_scoring_html() if request.args and 'league' in request.args: return m.league_html() return {}
744
239
# Copyright 2014 PDFium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. import os path = os.path.abspath(os.path.split(__file__)[0]) execfile(os.path.join(path, 'gyp_pdfium'))
262
96
from pyObjective import Variable, Model import numpy as np """This example script is written to demonstrate the use of classes, and how more complicated models can be built, and still passed to the solver. As a rudimentary example, it has two cubes and a sphere, and we are trying to find the dimensions such that the cube1 - cube2 + sphere volume is minimized, subject to the bounds. """ # define a new class class Cube: def __init__(self, model): self.x = Variable('x', 1, (0.5, 2), "cube length x") self.y = Variable('y', 1, (0.5, 2)) self.z = Variable('z', 1, (0.5, 2)) model.add_var(self.x) model.add_var(self.y) model.add_var(self.z) def volume(self): return self.x() * self.y() * self.z() # define a sphere, but keep the variable definition on the outside. For fun class Sphere: def __init__(self, radius): self.r = radius def volume(self): return (4 / 3) * np.pi * self.r() ** 3 # unfortunate brackets needed in here, and not before :( # define simulation model m = Model() # create cube c1 = Cube(m) c2 = Cube(m) # define the sphere radius r = Variable("r", 1, (0.5, 2), "sphere radius") m.add_var(r) # try commenting this line, and you will see that it was removed from the optimization s = Sphere(r) # define objective function (to be minimized) def cost(): return c1.volume() - c2.volume() + s.volume() m.objective = cost # solve m.solve() # display results m.display_results()
1,508
507
#!/usr/bin/env python # # A library that provides a Python interface to the Telegram Bot API # Copyright (C) 2015-2018 # Leandro Toledo de Souza <devs@python-telegram-bot.org> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser Public License for more details. # # You should have received a copy of the GNU Lesser Public License # along with this program. If not, see [http://www.gnu.org/licenses/]. import pytest from telegram import (InlineKeyboardButton, InlineQueryResultGame, InlineQueryResultVoice, InlineKeyboardMarkup) @pytest.fixture(scope='class') def inline_query_result_game(): return InlineQueryResultGame(TestInlineQueryResultGame.id, TestInlineQueryResultGame.game_short_name, reply_markup=TestInlineQueryResultGame.reply_markup) class TestInlineQueryResultGame(object): id = 'id' type = 'game' game_short_name = 'game short name' reply_markup = InlineKeyboardMarkup([[InlineKeyboardButton('reply_markup')]]) def test_expected_values(self, inline_query_result_game): assert inline_query_result_game.type == self.type assert inline_query_result_game.id == self.id assert inline_query_result_game.game_short_name == self.game_short_name assert (inline_query_result_game.reply_markup.to_dict() == self.reply_markup.to_dict()) def test_to_dict(self, inline_query_result_game): inline_query_result_game_dict = inline_query_result_game.to_dict() assert isinstance(inline_query_result_game_dict, dict) assert inline_query_result_game_dict['type'] == inline_query_result_game.type assert inline_query_result_game_dict['id'] == inline_query_result_game.id assert (inline_query_result_game_dict['game_short_name'] == inline_query_result_game.game_short_name) assert (inline_query_result_game_dict['reply_markup'] == inline_query_result_game.reply_markup.to_dict()) def test_equality(self): a = InlineQueryResultGame(self.id, self.game_short_name) b = InlineQueryResultGame(self.id, self.game_short_name) c = InlineQueryResultGame(self.id, '') d = InlineQueryResultGame('', self.game_short_name) e = InlineQueryResultVoice(self.id, '', '') assert a == b assert hash(a) == hash(b) assert a is not b assert a == c assert hash(a) == hash(c) assert a != d assert hash(a) != hash(d) assert a != e assert hash(a) != hash(e)
3,005
934
import pygame as pg class Sheet: """ Represents tool for extracting sprites from spritesheet. """ def __init__(self, sheet_path): """ Constructor for the sheet tool. Loading the spritesheet. """ self.spritesheet = pg.image.load(sheet_path).convert_alpha() def get_image(self, x, y, width, height, alpha=False): """ Extracts sprite of given point (x, y) (left, top) and width and height. alpha boolean keyword argument for converting the sprite in alpha or non-alpha. """ image = pg.Surface((width, height)) image.blit(self.spritesheet, (0, 0), (x, y, width, height)) image.set_colorkey((0, 0, 0)) image.set_alpha(255) if alpha: return image.convert_alpha() return image.convert()
835
253
# Analytical solutions for problems that can be solved with the two sine basis. ### # Problem one is # # -u`` = f in [0, pi] with u(0) = u(pi) = 0 for f which is # g on [0, pi/2) and h on [pi/2, pi] # ### # Problem two is # # u```` = f in [0, pi] with u(0) = u(pi) = 0, u`(0) = u`(pi) = 0 for f which is # g on [0, pi/2) and h on [pi/2, pi] from __future__ import division from sympy import Symbol, integrate import numpy as np from math import pi def solve_poisson(g, h): '''Solve the Poisson problem with f defined by g, h''' x = Symbol('x') # Primitive functions of g G = integrate(-g, x) GG = integrate(G, x) # Primitive functions of h H = integrate(-h, x) HH = integrate(H, x) # The solution is GG + a0*x + b0 on [-1, 0] and HH + a1*x + b1 on [0, 1] # Build the lin sys for the coefficients. The system reflects bcs and # continuity of u and u` in 0 A = np.array([[0, 1., 0., 0.], [0., 0., pi, 1.], [pi/2., 1., -pi/2, -1.], [1., 0., -1., 0.]]) b = np.array([-GG.subs(x, 0), -HH.subs(x, pi), HH.subs(x, pi/2) - GG.subs(x, pi/2), H.subs(x, pi/2) - G.subs(x, pi/2)]) [a0, b0, a1, b1] = np.linalg.solve(A, b) u0 = GG + a0*x + b0 u1 = HH + a1*x + b1 # Let's the the checks # Boundary conditions bcl = u0.subs(x, 0) bcr = u1.subs(x, pi) # Continuity of solution and the derivative u_cont = u0.subs(x, pi/2) - u1.subs(x, pi/2) du_cont = u0.diff(x, 1).subs(x, pi/2) - u1.diff(x, 1).subs(x, pi/2) # That it in fact solves the laplacian u0_lap = integrate((u0.diff(x, 2) + g)**2, (x, 0, pi/2)) u1_lap = integrate((u1.diff(x, 2) + h)**2, (x, pi/2, pi)) conds = [bcl, bcr, u_cont, du_cont, u0_lap, u1_lap] for i, c in enumerate(conds): print i, c, abs(c) < 1E-13 return u0, u1 def solve_biharmonic(g, h): '''Solve the biharmonic problem with f defined by g, h''' x = Symbol('x') # Primitive functions of g G = integrate(g, x) GG = integrate(G, x) GGG = integrate(GG, x) GGGG = integrate(GGG, x) # Primitive functions of h H = integrate(h, x) HH = integrate(H, x) HHH = integrate(HH, x) HHHH = integrate(HHH, x) # The solution now needs to match bcs and continuity. A = np.array([[-1./6, 1./2, -1., 1., 0., 0., 0., 0.], [0, 0, 0, 0, 1/6., 1/2., 1., 1.], [-1., 1, 0, 0, 0, 0, 0, 0.], [0, 0, 0, 0, 1., 1., 0, 0], [0, 0, 0, 1, 0, 0, 0, -1], [0, 0, 1, 0, 0, 0, -1, 0], [0, 1, 0, 0, 0, -1, 0, 0], [1, 0, 0, 0, -1, 0, 0, 0]]) b = np.array([-GGGG.subs(x, -1), -HHHH.subs(x, 1), -GG.subs(x, -1), -HH.subs(x, 1), HHHH.subs(x, 0) - GGGG.subs(x, 0), HHH.subs(x, 0) - GGG.subs(x, 0), HH.subs(x, 0) - GG.subs(x, 0), H.subs(x, 0) - G.subs(x, 0)]) [a0, a1, a2, a3, b0, b1, b2, b3] = np.linalg.solve(A, b) u0 = GGGG + a0*x**3/6 + a1*x**2/2 + a2*x + a3 u1 = HHHH + b0*x**3/6 + b1*x**2/2 + b2*x + b3 # Let's the the checks checks = [] # Boundary conditions checks.append(u0.subs(x, -1)) checks.append(u1.subs(x, 1)) checks.append(u0.diff(x, 2).subs(x, -1)) checks.append(u1.diff(x, 2).subs(x, 1)) # Continuity of solution and the derivatives checks.append(u0.subs(x, 0) - u1.subs(x, 0)) checks.append(u0.diff(x, 1).subs(x, 0) - u1.diff(x, 1).subs(x, 0)) checks.append(u0.diff(x, 2).subs(x, 0) - u1.diff(x, 2).subs(x, 0)) checks.append(u0.diff(x, 3).subs(x, 0) - u1.diff(x, 3).subs(x, 0)) # That it in fact solves the biharmonic equation checks.append(integrate((u0.diff(x, 4) - g)**2, (x, -1, 0))) checks.append(integrate((u1.diff(x, 4) - h)**2, (x, 0, 1))) assert all(map(lambda v: abs(v) < 1E-13, checks)) return u0, u1 # ----------------------------------------------------------------------------- if __name__ == '__main__': from sympy import S from sympy.plotting import plot x = Symbol('x') g, h = S(1), x problem = 'biharmonic' if problem == 'poisson': u0, u1 = solve_poisson(g, h) p0 = plot(u0, (x, 0, pi/2), show=False) p1 = plot(u1, (x, pi/2, pi), show=False) p2 = plot(g, (x, 0, pi/2), show=False) p3 = plot(h, (x, pi/2, pi), show=False) p4 = plot(u0.diff(x, 1), (x, 0, pi/2), show=False) p5 = plot(u1.diff(x, 1), (x, pi/2, pi), show=False) p0[0].line_color='red' p1[0].line_color='red' # p2[0].line_color='blue' # p3[0].line_color='blue' # p4[0].line_color='green' # p5[0].line_color='green' p0.append(p1[0]) # p0.append(p2[0]) # p0.append(p3[0]) # p0.append(p4[0]) # p0.append(p5[0]) if problem == 'biharmonic': u0, u1 = solve_biharmonic(g, h) u0.subs(x, 2/pi*x-1), u1.subs(x, 2/pi*x-1) # Sol k = 3 p0 = plot(u0.diff(x, k), (x, 0, pi/2), show=False) p1 = plot(u1.diff(x, k), (x, pi/2, pi), show=False) p0[0].line_color='red' p1[0].line_color='red' p0.append(p1[0]) p0.show()
5,398
2,517
# Copyright (c) 2020-2022 The PyUnity Team # This file is licensed under the MIT License. # See https://docs.pyunity.x10.bz/en/latest/license.html """Class to represent a rotation in 3D space.""" __all__ = ["Quaternion", "QuaternionDiff"] from . import Mathf from .vector import Vector3, conv from .other import LockedLiteral class Quaternion(LockedLiteral): """ Class to represent a unit quaternion, also known as a versor. Parameters ---------- w : float Real value of Quaternion x : float x coordinate of Quaternion y : float y coordinate of Quaternion z : float z coordinate of Quaternion """ def __init__(self, w, x, y, z): self.w = w self.x = x self.y = y self.z = z self._lock() def __repr__(self): return f"Quaternion({', '.join(map(conv, self))})" def __str__(self): return f"Quaternion({', '.join(map(conv, self))})" def __getitem__(self, i): if i == 0: return self.w elif i == 1: return self.x elif i == 2: return self.y elif i == 3: return self.z raise IndexError() def __iter__(self): yield self.w yield self.x yield self.y yield self.z def __list__(self): return [self.w, self.x, self.y, self.z] def __len__(self): return 4 def __hash__(self): return hash((self.w, self.x, self.y, self.z)) def __eq__(self, other): if hasattr(other, "__getitem__") and len(other) == 4: return self.w == other[0] and self.x == other[1] and self.y == other[2] and self.z == other[3] else: return False def __ne__(self, other): if hasattr(other, "__getitem__") and len(other) == 4: return self.w != other[0] or self.x != other[1] or self.y != other[2] or self.z != other[3] else: return True def __mul__(self, other): if isinstance(other, Quaternion): w = self.w * other.w - self.x * other.x - self.y * other.y - self.z * other.z x = self.w * other.x + self.x * other.w + self.y * other.z - self.z * other.y y = self.w * other.y - self.x * other.z + self.y * other.w + self.z * other.x z = self.w * other.z + self.x * other.y - self.y * other.x + self.z * other.w return Quaternion(w, x, y, z) elif isinstance(other, (int, float)): angle, axis = self.angleAxisPair return Quaternion.FromAxis((angle * other) % 360, axis) return NotImplemented def __truediv__(self, other): if isinstance(other, (int, float)): angle, axis = self.angleAxisPair return Quaternion.FromAxis((angle / other) % 360, axis) return NotImplemented def __sub__(self, other): if isinstance(other, Quaternion): diff = (self * other.conjugate).normalized() return QuaternionDiff(*diff) def absDiff(self, other): return abs(other - self) def copy(self): """ Deep copy of the Quaternion. Returns ------- Quaternion A deep copy """ return Quaternion(self.w, self.x, self.y, self.z) def normalized(self): """ A normalized Quaternion, for rotations. If the length is 0, then the identity quaternion is returned. Returns ------- Quaternion A unit quaternion """ length = Mathf.Sqrt(self.w ** 2 + self.x ** 2 + self.y ** 2 + self.z ** 2) if length: return Quaternion(self.w / length, self.x / length, self.y / length, self.z / length) else: return Quaternion.identity() @property def conjugate(self): """The conjugate of a unit quaternion""" return Quaternion(self.w, -self.x, -self.y, -self.z) def RotateVector(self, vector): """Rotate a vector by the quaternion""" other = Quaternion(0, *vector) return Vector3(self * other * self.conjugate) @staticmethod def FromAxis(angle, a): """ Create a quaternion from an angle and an axis. Parameters ---------- angle : float Angle to rotate a : Vector3 Axis to rotate about """ axis = a.normalized() cos = Mathf.Cos(angle / 2 * Mathf.DEG_TO_RAD) sin = Mathf.Sin(angle / 2 * Mathf.DEG_TO_RAD) return Quaternion(cos, axis.x * sin, axis.y * sin, axis.z * sin) @staticmethod def Between(v1, v2): a = v1.cross(v2) if a.dot(a) == 0: if v1 == v2 or v1.dot(v1) == 0 or v2.dot(v2) == 0: return Quaternion.identity() else: return Quaternion.FromAxis(180, Vector3.up()) angle = Mathf.Acos(v1.dot(v2) / (Mathf.Sqrt(v1.length * v2.length))) q = Quaternion.FromAxis(angle * Mathf.DEG_TO_RAD, a) return q.normalized() @staticmethod def FromDir(v): a = Quaternion.FromAxis( Mathf.Atan2(v.x, v.z) * Mathf.RAD_TO_DEG, Vector3.up()) b = Quaternion.FromAxis( Mathf.Atan2(-v.y, Mathf.Sqrt(v.z ** 2 + v.x ** 2)) * Mathf.RAD_TO_DEG, Vector3.right()) return a * b @property def angleAxisPair(self): """ Gets the angle and axis pair. Tuple of form (angle, axis). """ angle = 2 * Mathf.Acos(self.w) * Mathf.RAD_TO_DEG if angle == 0: return (0, Vector3.up()) return (angle, Vector3(self).normalized()) @staticmethod def Euler(vector): """ Create a quaternion using Euler rotations. Parameters ---------- vector : Vector3 Euler rotations Returns ------- Quaternion Generated quaternion """ a = Quaternion.FromAxis(vector.x, Vector3.right()) b = Quaternion.FromAxis(vector.y, Vector3.up()) c = Quaternion.FromAxis(vector.z, Vector3.forward()) return b * a * c @property def eulerAngles(self): """Gets the Euler angles of the quaternion""" s = self.w ** 2 + self.x ** 2 + self.y ** 2 + self.z ** 2 r23 = 2 * (self.w * self.x - self.y * self.z) if r23 > 0.999999 * s: x = Mathf.PI / 2 y = 2 * Mathf.Atan2(self.y, self.x) z = 0 elif r23 < -0.999999 * s: x = -Mathf.PI / 2 y = -2 * Mathf.Atan2(self.y, self.x) z = 0 else: x = Mathf.Asin(r23) r13 = 2 * (self.w * self.y + self.z * self.x) / s r33 = 1 - 2 * (self.x ** 2 + self.y ** 2) / s r21 = 2 * (self.w * self.z + self.x * self.y) / s r22 = 1 - 2 * (self.x ** 2 + self.z ** 2) / s y = Mathf.Atan2(r13, r33) z = Mathf.Atan2(r21, r22) euler = [x, y, z] for i in range(3): euler[i] = euler[i] * Mathf.RAD_TO_DEG % 360 if euler[i] > 180: euler[i] -= 360 return Vector3(euler) @staticmethod def identity(): """Identity quaternion representing no rotation""" return Quaternion(1, 0, 0, 0) class QuaternionDiff: def __init__(self, w, x, y, z): self.w = w self.x = x self.y = y self.z = z def __abs__(self): return abs(2 * Mathf.Acos(self.w) * Mathf.DEG_TO_RAD)
7,642
2,674
from config import secret from utils import getLogger DEBUG = getLogger() class MysqlConnection: def __init__(self, database, server_name): self.host = secret[server_name]['mysql']['host'] self.port = secret[server_name]['mysql']['port'] self.username = secret[server_name]['mysql']['username'] self.password = secret[server_name]['mysql']['password'] self.database = database
402
112
#from selenium.webdriver.remote import webdriver from selenium import webdriver #from selenium.webdriver.chrome import options from page_objects import EntryPage, LoginPage, RedactionPage, ContactTracePage, AddNewRecordPage, AddDataToRecordPage, StageForPublishingPage, PublishDataPage, SettingsPage, Tools import unittest import os class TestRedaction(unittest.TestCase): def setUp(self): #Change this to TRUE if you don't want to use a dockerised stack self.local_mode = False #setup environment based on environment variables if 'HOME_DIR' in os.environ.copy(): self.home_dir = os.environ['HOME_DIR'] else: self.home_dir = os.getcwd() if 'DATA_DIR' in os.environ.copy(): self.data_dir = os.environ['DATA_DIR'] else: self.data_dir = 'tests/data/' if 'TMP_DIR' in os.environ.copy(): self.tmp_dir = os.environ['TMP_DIR'] else: self.tmp_dir = '/tmp/' if 'BASE_TEST_URL' in os.environ.copy(): self.base_url = os.environ['BASE_TEST_URL'] else: self.base_url = 'https://react.safeplaces.extremesolution.com/' if 'SELENIUM_URL' in os.environ.copy(): self.sel_url = os.environ['SELENIUM_URL'] else: self.sel_url = 'http://172.17.0.2:4444/wd/hub' chrome_options = webdriver.ChromeOptions() prefs = {'download.default_directory': '/tmp'} chrome_options.add_experimental_option('prefs', prefs) if self.local_mode: self.driver = webdriver.Chrome(chrome_options=chrome_options) else: self.driver = webdriver.Remote(command_executor=self.sel_url, options=chrome_options) def app_loads(self): tools = Tools() entry_page = EntryPage(self.driver,base_url=self.base_url) entry_page.open_page() def contact_trace(self): tools = Tools() entry_page = EntryPage(self.driver,base_url=self.base_url) entry_page.open_page() login_page = LoginPage(self.driver) login_page.login_if_required() entry_page.open_trace() contact_trace_page = ContactTracePage(self.driver) contact_trace_page.add_new_record() add_record_page = AddNewRecordPage(self.driver) add_record_page.create_manually() contact_trace_page.more() contact_trace_page.add_data_point() # start to add a point and cancel editing the point # if the test works this far, we can expand it later point_editor_page = AddDataToRecordPage(self.driver) point_editor_page.enter_location('-122.19732036472264, 37.718665250290684') point_editor_page.enter_date('06/08/2020 07:00') point_editor_page.close() entry_page.open_settings() settings_page = SettingsPage(self.driver) settings_page.logout # leaving test_ out of the method name until the SUT works def settings(self): login_page = LoginPage(self.driver) login_page.login_if_required() entry_page = EntryPage(self.driver,base_url=self.base_url) entry_page.open_page() entry_page.open_settings() settings_page = SettingsPage(self.driver) settings_page.set_health_authority_name('Test Health Authority') settings_page.set_information_website_URL('https://cdc.gov') settings_page.set_reference_website_URL('https://cdc.gov') settings_page.set_api_endpoint('https://s3.aws.com/bucket_name/safepaths.json') settings_page.set_privacy_policy_URL('https://www.cdc.gov/other/privacy.html') # set retention policy slider to 50% of the way across, which would be 15 days # commented out until we find how to get ActionChains working # settings_page.set_retention_policy('50') settings_page.reset_gps_coordinates settings_page.save_and_continue #def test_redaction(self): <--- removed test_ from the method name until the SUT works! def redaction(self): tools = Tools() entry_page = EntryPage(self.driver,base_url=self.base_url) entry_page.open_page() entry_page.setup_case() entry_page.open_redactor() login_page = LoginPage(self.driver) login_page.login_if_required() redaction_page = RedactionPage(self.driver) redaction_page.load_file(self.data_dir +'/privkit31A-synthetic-REDACTED.json') redaction_page.check_start_date_is('1-Mar-2020 1:00pm GMT') redaction_page.check_end_date_is('19-Mar-2020 10:00pm GMT') redaction_page.check_duration_is('18 days 9 hrs') redaction_page.save_file() #TODO: this next step fails because it was designed for backend=OFF. To test this, we need to load the publisher screen and see what data is there when we hit load #tools.compare_files(self.tmp_dir + '/privkit31A-synthetic-REDACTED-REDACTED.json', self.home_dir + '/' + self.data_dir + '/expected_results/privkit31A-synthetic-REDACTED-REDACTED.json') def tearDown(self): self.driver.close()
5,147
1,652
""" A small, very basic OutputCollector-derived class used to test handling of multiple OutputCollectors in inference_runner scenario """ from kenning.core.outputcollector import OutputCollector from kenning.datasets.open_images_dataset import DectObject from typing import Any, Union import numpy as np from pathlib import Path class NamePrinter(OutputCollector): def __init__(self, print_type: str = "detector", file_path: Path = None): self.frame_counter = 0 self.print_type = print_type self.classnames = [] if file_path: with open(file_path, 'r') as f: for line in f: self.classnames.append(line.strip()) super().__init__() @classmethod def form_argparse(cls): parser, group = super().form_argparse() group.add_argument( '--print-type', help='What is the type of model that will input data to the NamePrinter', # noqa: E501 choices=['detector', 'classificator'], default='detector' ) group.add_argument( '--classification-class-names', help='File with class names used to identify the output from classification models', # noqa: E501 type=Path ) return parser, group @classmethod def from_argparse(cls, args): return cls(args.print_type, args.classification_class_names) def detach_from_output(self): pass def should_close(self): return False def process_output(self, i: Any, o: Union[DectObject, np.array]): print("Frame", self.frame_counter, end=": ") o = o[0] if self.print_type == 'detector': for x in o: print(x.clsname, end=" ") print() elif self.print_type == 'classificator': tuples = [] if self.classnames: for i, j in zip(o, self.classnames): tuples.append((i, j)) else: it = 0 for i in o: tuples.append((i, "object {}".format(it))) it += 1 tuples.sort(key=lambda x: x[0], reverse=True) for i in range(min(5, len(tuples))): print( "{}: {:.2f}".format( tuples[i][1], float(tuples[i][0]) ), end=', ' ) print() self.frame_counter += 1
2,532
718
from typing import Any from .. import util FORMAT = '''\ 🤔 {username} 发布了……一些东西 https://t.bilibili.com/{id} 目前机器人还不能理解这个qwq''' def handle(content: Any) -> str: return FORMAT.format( username=content["desc"]["user_profile"]["info"]["uname"], id=content["desc"]["dynamic_id_str"])
291
122
from collections import OrderedDict from geotrek.authent.tests.factories import StructureFactory from georiviere.tests import CommonRiverTest from georiviere.observations.models import Station, ParameterTracking from .factories import ( StationFactory, StationProfileFactory, ParameterFactory ) class StationViewTestCase(CommonRiverTest): model = Station modelfactory = StationFactory def get_expected_json_attrs(self): return { 'id': self.obj.pk, 'code': self.obj.code, 'date_insert': '2020-03-17T00:00:00Z', 'date_update': '2020-03-17T00:00:00Z', 'description': self.obj.description, 'geom': self.obj.geom.ewkt, 'hardness': self.obj.hardness, 'in_service': self.obj.in_service, 'label': self.obj.label, 'local_influence': self.obj.local_influence, 'purpose_code': self.obj.purpose_code, 'site_code': self.obj.site_code, 'station_profiles': [], 'structure': self.obj.structure.pk, 'station_uri': self.obj.station_uri, } def get_bad_data(self): return OrderedDict([ ('label', ''), ('parametertracking_set-TOTAL_FORMS', '0'), ('parametertracking_set-INITIAL_FORMS', '1'), ('parametertracking_set-MAX_NUM_FORMS', '0'), ]), 'This field is required.' def get_good_data(self): structure = StructureFactory.create() station_profile = StationProfileFactory.create() temp_data = self.modelfactory.build( structure=structure, station_profiles=[station_profile] ) return { 'structure': structure.pk, 'geom': '{"geom": "%s", "snap": [%s]}' % (temp_data.geom.transform(4326, clone=True).ewkt, ','.join(['null'])), 'code': '1234', 'label': 'test', 'station_profiles': [station_profile.pk], 'local_influence': Station.LocalInfluenceChoices.UNKNOWN, 'parametertracking_set-TOTAL_FORMS': '2', 'parametertracking_set-INITIAL_FORMS': '0', 'parametertracking_set-MAX_NUM_FORMS': '', 'parametertracking_set-0-label': 'Paramètre suivi Ted', 'parametertracking_set-0-parameter': ParameterFactory.create().pk, 'parametertracking_set-0-measure_frequency': 'par jour', 'parametertracking_set-0-transmission_frequency': 'par semaine', 'parametertracking_set-0-data_availability': ParameterTracking.DataAvailabilityChoice.ONDEMAND, 'parametertracking_set-0-measure_start_date': '2019-06-06', 'parametertracking_set-0-measure_end_date': '2020-06-06', 'parametertracking_set-0-id': '', 'parametertracking_set-0-DELETE': '', 'parametertracking_set-1-label': 'Paramètre suivi Bob', 'parametertracking_set-1-parameter': ParameterFactory.create().pk, 'parametertracking_set-1-measure_frequency': '', 'parametertracking_set-1-transmission_frequency': '', 'parametertracking_set-1-data_availability': ParameterTracking.DataAvailabilityChoice.ONDEMAND, 'parametertracking_set-1-measure_start_date': '', 'parametertracking_set-1-measure_end_date': '', 'parametertracking_set-1-id': '', 'parametertracking_set-1-DELETE': '', } def test_listing_number_queries(self): """Test number queries when get list object""" self.login() self.modelfactory.create_batch(100) with self.assertNumQueries(7): self.client.get(self.model.get_jsonlist_url()) with self.assertNumQueries(6): self.client.get(self.model.get_format_list_url()) def test_detail_number_queries(self): """Test number queries when get detail object""" self.login() station = self.modelfactory.create() with self.assertNumQueries(49): self.client.get(station.get_detail_url())
4,159
1,322
import requests from requests.auth import HTTPBasicAuth import json import pprint class Cloudflare_Enduser_API: def __init__(self, cf_token, cf_email): self.cf_token = cf_token self.cf_email = cf_email self.headers = { 'Content-Type': 'application/json', 'X-Auth-Key': self.cf_token, 'X-Auth-Email': self.cf_email } def cfQuery(self, domain_id='', endpoint='', data={}): """ Main GET query function :param domain_id: ID of domain to query :param endpoint: user, zone, etc... :return: result cloudflare response """ if not domain_id and not endpoint: #print('domain id and endpoint not set') return requests.get(api_endpoint + '/zones' + domain_id + endpoint, headers=self.headers) elif domain_id and endpoint: return requests.get(api_endpoint + '/zones/' + domain_id + '/' + endpoint, headers=self.headers) elif endpoint: return requests.get(api_endpoint + '/' + endpoint, headers=self.headers) def cfQuery_patch(self, data, endpoint): """ Main PATCH query function :param domain_id: ID of domain to query :param endpoint: user, zone, etc... :param data: data to be patched :return: result cloudflare response """ auth = HTTPBasicAuth(self.cf_email, self.cf_token) return requests.patch(api_endpoint + endpoint, data=json.dumps(data), auth=auth, headers=self.headers) def cfQuery_post(self, data, endpoint): """ Main POST query function :param data: data to be posted :return: cloudflare response """ return requests.post(api_endpoint + endpoint, data=json.dumps(data), headers=self.headers) def cfQuery_delete(self, id, endpoint): return requests.delete(api_endpoint + endpoint + id, headers=self.headers) def cfQuery_put(self, data, endpoint): """ Main POST query function :param data: data to be posted :return: cloudflare response """ return requests.put(api_endpoint + endpoint, data=json.dumps(data), headers=self.headers) def pp_print_of(self, some_input): """ Preaty print the output :param some_input: anything it can be a function too :return: function output with pprint """ pp = pprint.PrettyPrinter(indent=4) return pp.pprint(some_input) def list_all(self): """ Get all data associated with cloudflare user :return: all data for the user account """ r = self.cfQuery().json() pp = pprint.PrettyPrinter(indent=4) return pp.pprint(r) def list_domain_ids(self): """ lists domain ID/'s :return: get list of all domain's and id's associated with cloudflare account """ r = self.cfQuery().json() records = [] for domain in r.get('result'): domain_name = domain.get('name') domain_id = domain.get('id') records.append((domain_name, domain_id)) return dict(records) def get_record(self, id): """ get dns records for given domain id :param id: you need to know the domain id that can be found through list_domain_ids function :return: return list of all dns records associated with domain """ domains = self.list_domain_ids() endpoint = 'dns_records' # endpoint of the query for CloudFlare API v4 if str(id) in domains.values(): r = self.cfQuery(domain_id=id, endpoint=endpoint) return r.json().get('result')[0].get('content') elif str(id) not in domains.values(): return ''.join('There is no domain with id: ' + str(id) + "\nlist od available domain id's is:\n" + json.dumps(domains)) def get_user(self): """ Get user info :return: return user info """ return self.cfQuery(endpoint='user').json() def patch_user(self, data): """ Update your user profile :param data: dictionary of data to be changed, :return: PATCH request to update user profile """ return self.cfQuery_patch(endpoint='/user', data=data) def get_user_billing_profile(self, x): """ User Billing Profile :param x: cloudflare app endpoint (profile, history) :return: current billing profile """ return self.cfQuery(endpoint='user/billing/' + x).json() def get_app_subscriptions(self, x=None): """ List all of your app subscriptions :param x: optional parameter to show info for app :return: list """ if x: return self.cfQuery(endpoint='user/billing/subscriptions/apps' + '/' + x).json() return self.cfQuery(endpoint='user/billing/subscriptions/apps').json() ######## FOR NOW WILL LEAVE OUT REST OF PAYED API CALLES def get_firewall_rules(self): """ List firewall access rules :param x: optional parameter to show info for app :return: list """ return self.cfQuery(endpoint='/user/firewall/access_rules/rules').json() def set_access_rule(self, data): return self.cfQuery_post(data, endpoint='/user/firewall/access_rules/rules').json() def update_access_rule(self, data): return self.cfQuery_patch(data, endpoint='/user/firewall/access_rules/rules').json() def del_access_rule(self, id): return self.cfQuery_delete(id, endpoint='/user/firewall/access_rules/rules/').json() ################################################################ ########## Domain related def add_new_zone(self, data): #return self.cfQuery_post(data, endpoint='/zones') pass def get_zone_details(self, id): return self.cfQuery(domain_id=id).json() ################################################################ ############## DNS Record def create_dns_record(self, id, data): return self.cfQuery_post(data, endpoint='/zones/' + id + '/dns_records').json() def get_dns_records(self, id): endpoint = '/dns_records' return self.cfQuery(domain_id=id, endpoint=endpoint).json() def get_dns_record_details(self, id, record_id): endpoint = '/dns_records/' + record_id return self.cfQuery(domain_id=id, endpoint=endpoint).json() def update_dns_record(self, data): endpoint = '/zones/' + data.get('zone_id') + '/dns_records/' + data.get('id') return self.cfQuery_put(data, endpoint=endpoint).json() def del_dns_record(self, zone_id, domain_id): endpoint = '/zones/' + zone_id + '/dns_records/' return self.cfQuery_delete(id=domain_id, endpoint=endpoint).json() ################################################################# ################## Analytics def get_zone_analytics(self, zone_id): endpoint = 'analytics/dashboard' return self.cfQuery(domain_id=zone_id, endpoint=endpoint).json() ################################################################## ################## Zone Firewall rules def get_zone_firewall_rules(self, zone_id): endpoint = 'firewall/access_rules/rules' return self.cfQuery(domain_id=zone_id, endpoint=endpoint).json() def create_zone_firewall_rule(self, zone_id, data): endpoint = '/zones/' + zone_id + '/firewall/access_rules/rules' return self.cfQuery_post(data=data, endpoint=endpoint).json() def update_zone_firewall_rule(self, zone_id, data): endpoint = '/zones/' + zone_id + '/firewall/access_rules/rules' return self.cfQuery_patch(data=data, endpoint=endpoint) def del_zone_firewall_rule(self, zone_id, data): endpoint = '/zones/' + zone_id + '/firewall/access_rules/rules' #return self.cfQuery_delete(id=data.get(id), endpoint=endpoint) pass
8,145
2,364
import unittest import os from subprocess import call import z5py import vigra from test_class import McLuigiTestCase class TestDataTasks(McLuigiTestCase): @classmethod def setUpClass(cls): super(TestDataTasks, cls).setUpClass() @classmethod def tearDownClass(cls): super(TestDataTasks, cls).tearDownClass() def check_features(self, feature_path): rag_path = './cache/StackedRegionAdjacencyGraph_sampleA_watershed.h5' self.assertTrue(os.path.exists(rag_path)) n_edges = vigra.readHDF5(rag_path, 'numberOfEdges') self.assertTrue(os.path.exists(feature_path)) features = z5py.File(feature_path, use_zarr_format=False)['data'][:] self.assertEqual(n_edges, len(features)) for feat_id in range(features.shape[1]): self.assertFalse((features[:, feat_id] == 0).all()) def test_region_features(self): call(['python', './executables/features.py', 'region']) feat_path = '' self.check_features(feat_path) if __name__ == '__main__': unittest.main()
1,085
367
# Generated by Django 3.0.8 on 2020-07-13 19:43 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ("images", "0003_image_creation_date"), ] operations = [ migrations.AddField( model_name="image", name="feed", field=models.BooleanField(default=False, verbose_name="In feed"), ), migrations.AlterField( model_name="image", name="creation_date", field=models.DateTimeField(auto_now_add=True, verbose_name="Creation date"), ), migrations.AlterField( model_name="image", name="files", field=models.ManyToManyField(to="images.ImageFile", verbose_name="Files"), ), migrations.AlterField( model_name="image", name="thumbnail", field=models.ForeignKey( blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name="container", to="images.ImageFile", verbose_name="Thumbnail", ), ), migrations.AlterField( model_name="image", name="title", field=models.CharField(max_length=200, verbose_name="Title"), ), migrations.AlterField( model_name="imagefile", name="height", field=models.IntegerField(verbose_name="Height"), ), migrations.AlterField( model_name="imagefile", name="image_file", field=models.FileField(upload_to="", verbose_name="Image file"), ), migrations.AlterField( model_name="imagefile", name="width", field=models.IntegerField(verbose_name="Width"), ), ]
1,912
533
import numba import numpy as np from scipy.sparse import csr_matrix from .base import BasePointer, GraphBlasContainer from .context import handle_panic, return_error from .exceptions import GrB_Info class MatrixPtr(BasePointer): def set_matrix(self, matrix): self.instance = matrix class Matrix(GraphBlasContainer): def __init__(self, matrix): assert isinstance(matrix, csr_matrix) self.matrix = matrix @classmethod def new_from_dtype(cls, dtype, nrows, ncols): matrix = csr_matrix((nrows, ncols), dtype=dtype) return cls(matrix) @classmethod def new_from_existing(cls, other): matrix = csr_matrix(other) return cls(matrix) @classmethod def get_pointer(cls): return MatrixPtr() @handle_panic def Matrix_new(A: MatrixPtr, dtype: type, nrows: int, ncols: int): if nrows <= 0: return_error(GrB_Info.GrB_INVALID_VALUE, "nrows must be > 0") if ncols <= 0: return_error(GrB_Info.GrB_INVALID_VALUE, "ncols must be > 0") matrix = Matrix.new_from_dtype(dtype, nrows, ncols) A.set_matrix(matrix) return GrB_Info.GrB_SUCCESS @handle_panic def Matrix_dup(C: MatrixPtr, A: Matrix): matrix = Matrix.new_from_existing(A) C.set_matrix(matrix) return GrB_Info.GrB_SUCCESS @handle_panic def Matrix_resize(C: Matrix, nrows: int, ncols: int): if nrows <= 0: return_error(GrB_Info.GrB_INVALID_VALUE, "nrows must be > 0") if ncols <= 0: return_error(GrB_Info.GrB_INVALID_VALUE, "ncols must be > 0") C.matrix.resize((nrows, ncols)) return GrB_Info.GrB_SUCCESS # TODO: this is just the essential code; it needs to handle descriptors, masks, accumulators, etc @handle_panic def mxm(C, A, B, semiring): cr, cc = C.shape ar, ac = A.shape br, bc = B.shape if cr != ar: return_error(GrB_Info.GrB_DIMENSION_MISMATCH, "C.nrows != A.nrows") if cc != bc: return_error(GrB_Info.GrB_DIMENSION_MISMATCH, "C.ncols != B.ncols") if ac != br: return_error(GrB_Info.GrB_DIMENSION_MISMATCH, "A.nrows != B.ncols") b = B.tocsc() d, i, ip = _sparse_matmul( A.data, A.indices, A.indptr, b.data, b.indices, b.indptr, semiring.plus.op, semiring.times, semiring.plus.identity, C.dtype, ) C.data = d C.indices = i C.indptr = ip return GrB_Info.GrB_SUCCESS @numba.njit def _sparse_matmul( a_data, a_indices, a_indptr, b_data, b_indices, b_indptr, plus, times, identity, dtype, ): # Final array size is unknown, so we give ourselves room and then adjust on the fly tmp_output_size = a_data.size * 2 data = np.empty((tmp_output_size,), dtype=dtype) indices = np.empty((tmp_output_size,), dtype=a_indices.dtype) indptr = np.empty((a_indptr.size,), dtype=a_indptr.dtype) output_counter = 0 for iptr in range(a_indptr.size - 1): indptr[iptr] = output_counter for jptr in range(b_indptr.size - 1): a_counter = a_indptr[iptr] a_stop = a_indptr[iptr + 1] b_counter = b_indptr[jptr] b_stop = b_indptr[jptr + 1] val = identity nonempty = False while a_counter < a_stop and b_counter < b_stop: a_k = a_indices[a_counter] b_k = b_indices[b_counter] if a_k == b_k: val = plus(val, times(a_data[a_counter], b_data[b_counter])) nonempty = True a_counter += 1 b_counter += 1 elif a_k < b_k: a_counter += 1 else: b_counter += 1 if nonempty: if output_counter >= tmp_output_size: # We filled up the allocated space; copy existing data to a larger array tmp_output_size *= 2 new_data = np.empty((tmp_output_size,), dtype=data.dtype) new_indices = np.empty((tmp_output_size,), dtype=indices.dtype) new_data[:output_counter] = data[:output_counter] new_indices[:output_counter] = indices[:output_counter] data = new_data indices = new_indices data[output_counter] = val indices[output_counter] = jptr output_counter += 1 # Add final entry to indptr (should indicate nnz in the output) nnz = output_counter indptr[iptr + 1] = nnz # Trim output arrays data = data[:nnz] indices = indices[:nnz] return (data, indices, indptr)
4,733
1,645
from django.contrib import admin from imagekit.admin import AdminThumbnail from .models import Icon @admin.register(Icon) class IconAdmin(admin.ModelAdmin): list_display = ('name', 'icon_thumbnail') icon_thumbnail = AdminThumbnail(image_field='thumbnail')
266
77
def getSKLearnLogisticRegression(self, regParam, dim=1): from DLplatform.learning.batch.sklearnClassifiers import LogisticRegression learner = LogisticRegression(regParam = regParam, dim = dim) return learner
245
69
import json import os import shutil import tempfile from datetime import timedelta from unittest import mock from unittest import TestCase from pytuber.storage import Registry class RegistryTests(TestCase): def tearDown(self): Registry.clear() Registry._obj = {} def test_singleton(self): a = Registry() b = Registry() self.assertIs(a, b) a[1] = 2 self.assertEqual({1: 2}, b) def test_set(self): Registry.set(1, 2, 3, 4, 5) self.assertEqual({1: {2: {3: {4: 5}}}}, Registry()) Registry.set(1, 3, 5) self.assertEqual({1: {2: {3: {4: 5}}, 3: 5}}, Registry()) def test_get(self): Registry.set(1, 2, 3, 4, 5) self.assertEqual({4: 5}, Registry.get(1, 2, 3)) with self.assertRaises(KeyError): Registry.get(2) def test_clear(self): Registry.set(1, 2, 3, 4, 5) self.assertEqual({4: 5}, Registry.get(1, 2, 3)) Registry.clear() self.assertEqual({}, Registry()) def test_from_file(self): try: tmp = tempfile.mkdtemp() file_path = os.path.join(tmp, "foo.json") with open(file_path, "w") as fp: json.dump({"a": True}, fp) Registry.from_file(file_path) self.assertEqual({"a": True}, Registry()) Registry.set("a", False) self.assertFalse(Registry.get("a")) Registry.from_file(file_path) self.assertFalse(Registry.get("a")) finally: shutil.rmtree(tmp) def test_persist(self): try: Registry.set(1, 2, 3, 4) tmp = tempfile.mkdtemp() file_path = os.path.join(tmp, "foo.json") Registry.persist(file_path) Registry.set(1, 2, 3, 5) Registry._obj = {} Registry.from_file(file_path) self.assertEqual({"1": {"2": {"3": 4}}}, Registry()) finally: shutil.rmtree(tmp) @mock.patch("pytuber.storage.time.time") def test_cache(self, time): time.side_effect = [10, 20.1, 20.1, 20.5, 20.8] def callme(ttl, value, refresh=False): return Registry.cache( key="foo", ttl=timedelta(seconds=ttl), func=lambda: value, refresh=refresh, ) self.assertEqual("first", callme(10, "first")) self.assertEqual(("first", 20.0), Registry.get("foo")) self.assertEqual("second", callme(1, "second")) self.assertEqual(("second", 21.1), Registry.get("foo")) self.assertEqual("second", callme(1, "third")) self.assertEqual(("second", 21.1), Registry.get("foo")) self.assertEqual("third", callme(100, "third", refresh=True)) self.assertEqual(("third", 120.8), Registry.get("foo")) self.assertEqual(5, time.call_count)
2,933
1,039
""" Problem: You are given given a list of rectangles represented by min and max x- and y-coordinates. Compute whether or not a pair of rectangles overlap each other. If one rectangle completely covers another, it is considered overlapping. For example, given the following rectangles: { "top_left": (1, 4), "dimensions": (3, 3) # width, height }, { "top_left": (-1, 3), "dimensions": (2, 1) }, { "top_left": (0, 5), "dimensions": (4, 3) } return true as the first and third rectangle overlap each other. """ from typing import Dict, List, Tuple Rectangle = Dict[str, Tuple[int, int]] def get_intersection_area(rect1: List[Rectangle], rect2: List[Rectangle]) -> int: if rect1["top_left"][0] < rect2["top_left"][0]: left = rect1 right = rect2 else: left = rect2 right = rect1 if rect1["top_left"][1] > rect2["top_left"][1]: top = rect1 bottom = rect2 else: top = rect2 bottom = rect1 if (left["top_left"][0] + left["dimensions"][0]) < right["top_left"][0]: return 0 else: span_x = (left["top_left"][0] + left["dimensions"][0]) - right["top_left"][0] if (top["top_left"][1] - top["dimensions"][1]) > bottom["top_left"][1]: return 0 else: span_y = bottom["top_left"][1] - (top["top_left"][1] - top["dimensions"][1]) return span_x * span_y def get_covered_area(rect: Rectangle) -> int: width, height = rect["dimensions"] return width * height def check_rectangles_intersection(rectangles: List[Rectangle]) -> bool: length = len(rectangles) # checking for intersection for each pair of rectangles for i in range(length - 1): for j in range(i + 1, length): intersection_area = get_intersection_area(rectangles[i], rectangles[j]) rect1_area = get_covered_area(rectangles[i]) rect2_area = get_covered_area(rectangles[j]) if intersection_area in (rect1_area, rect2_area): return True return False if __name__ == "__main__": # NOTE: THE QUESTION STATEMENT IS WRONG THE RECTANGLES 1 & 3 DOES NOT OVERLAP BUT # ONLY INTERSECT (SMALL MODIFICATION DONE TO MAKE THEM OVERLAP) rectangles = [ {"top_left": (1, 4), "dimensions": (3, 3)}, {"top_left": (-1, 3), "dimensions": (2, 1)}, {"top_left": (0, 5), "dimensions": (4, 4)}, # MODIFICATION ] print(check_rectangles_intersection(rectangles)) rectangles.pop() print(check_rectangles_intersection(rectangles)) """ SPECS: TIME COMPLEXITY: O(n ^ 2) SPACE COMPLEXITY: O(1) """
2,636
945
import asyncio from asynckafka import Producer, Consumer import config import utils loop = asyncio.get_event_loop() async def fill_topic_with_messages(): producer = Producer( brokers=config.KAFKA_URL, rdk_producer_config=config.RDK_PRODUCER_CONFIG, rdk_topic_config=config.RDK_TOPIC_CONFIG, ) producer.start() messages_consumed = 0 print(f"Preparing benchmark. Filling topic {config.TOPIC} with " f"{config.MESSAGE_NUMBER} messages of {config.MESSAGE_BYTES} bytes " f"each one.") await asyncio.sleep(0.1) with utils.Timer() as timer: for _ in range(config.MESSAGE_NUMBER): messages_consumed += 1 await producer.produce(config.TOPIC, config.MESSAGE) producer.stop() print(f"The producer time to send the messages is {timer.interval} " f"seconds.") utils.print_statistics(timer.interval) async def consume_the_messages_stream_consumer(): stream_consumer = Consumer( brokers=config.KAFKA_URL, topics=[config.TOPIC], rdk_consumer_config=config.RDK_CONSUMER_CONFIG, rdk_topic_config=config.RDK_TOPIC_CONFIG ) stream_consumer.start() messages_consumed = 0 print("Starting to consume the messages.") with utils.Timer() as timer: async for message in stream_consumer: messages_consumed += 1 if messages_consumed == config.MESSAGE_NUMBER: stream_consumer.stop() print(f"The time used to consume the messages is {timer.interval} " f"seconds.") utils.print_statistics(timer.interval) async def main_coro(): await fill_topic_with_messages() await consume_the_messages_stream_consumer() if __name__ == "__main__": loop.run_until_complete(main_coro())
1,819
596
import RPi.GPIO as GPIO GPIO.setmode(GPIO.BCM) GPIO.setup(23, GPIO.IN, pull_up_down=GPIO.PUD_UP) GPIO.setup(18, GPIO.IN, pull_up_down=GPIO.PUD_UP) GPIO.setup(24, GPIO.IN, pull_up_down=GPIO.PUD_DOWN) def my_callback(channel): print "falling edge detected on 18" def my_callback2(channel): print "falling edge detected on 23" raw_input("Press Enter when ready\n>") GPIO.add_event_detect(18, GPIO.FALLING, callback=my_callback, bouncetime=300) GPIO.add_event_detect(23, GPIO.FALLING, callback=my_callback2, bouncetime=300) try: print "Waiting for rising edge on port 24" GPIO.wait_for_edge(24, GPIO.RISING) print "Rising edge detected on port 24. Here endeth the third lesson." except KeyboardInterrupt: GPIO.cleanup() GPIO.cleanup()
744
349
# -*- coding: utf-8 -*- """ file: module_topology_distmat_test.py Unit tests for distance matrix computations """ import os from pandas import DataFrame from interact.md_system import System from tests.module.unittest_baseclass import UnittestPythonCompatibility class DistanceMatrixTests(UnittestPythonCompatibility): currpath = os.path.dirname(__file__) pdb_file = os.path.abspath(os.path.join(currpath, '../files/1acj.pdb')) mol_file = os.path.abspath(os.path.join(currpath, '../files/1acj.mol2')) def setUp(self): """ Prepare TopologyDataFrame once for every test """ self.top = System(self.pdb_file, mol2file=self.mol_file).topology def test_distmat_overflow_exception(self): """ Test OverflowError exception for (too) large distance matrix """ # Unable to compute distance matrix > max_distmat_size self.assertRaises(OverflowError, self.top.distances, max_distmat_size=10000) def test_distmat_attribute_exception(self): """ Test AttributeError on missing or incomplete coordinates """ # No coordinates self.top._coordinates = None self.assertRaises(AttributeError, self.top.distances) def test_distmat_square(self): """ Test computation of default square matrix """ distmat = self.top.distances() self.assertIsInstance(distmat, DataFrame) self.assertEqual(distmat.shape[0], distmat.shape[1]) self.assertEqual(list(distmat.columns), list(distmat.index)) def test_distmat_target(self): """ Test computation of matrix with custom source and target selection """ source = self.top[self.top['resSeq'] == 999] target = self.top[self.top['resName'] == 'HIS'] distmat = source.distances(target=target) self.assertIsInstance(distmat, DataFrame) self.assertEqual(distmat.shape, (17, 138)) self.assertEqual(len(distmat.columns), len(target)) self.assertEqual(len(distmat.index), len(source)) def test_distmat_empty_selection(self): """ Test compuation of matrix when (one of) the input selections is empty """ source = self.top[self.top['resSeq'] == 9999] target = self.top[self.top['resName'] == 'HIS'] self.assertTrue(source.distances().empty) self.assertTrue(source.distances(target=target).empty)
2,465
757
import typing import bs4 import requests class ScrapeAllComicIds(): def __call__( self, ) -> typing.List[int]: self.__find() return self.__ids def __find( self, ) -> typing.NoReturn: self.__ids = [] for q in self.__query: self.__find_per_page(q) def __find_per_page( self, query: str, ) -> typing.NoReturn: response = requests.get( f'{self.__base_url}' f'{query}', ) soup = bs4.BeautifulSoup( response.content, 'html.parser', ) elms = soup.find( id='all_title', ).find('ul').find_all('li') for elm in elms: url = elm.find( 'a', ).get('href') self.__ids.append( int(url.split('=')[-1]) ) def __init__( self, ) -> typing.NoReturn: self.__base_url = ( 'http://ruijianime.com/' 'comic/title/all_title' '.php?q=' ) self.__query = ( 'a', 'ka', 'sa', 'ta', 'na', 'ha', 'ma', 'ya', 'ra', 'wa', )
1,010
392
""" Path Converter. pymdownx.pathconverter An extension for Python Markdown. An extension to covert tag paths to relative or absolute: Given an absolute base and a target relative path, this extension searches for file references that are relative and converts them to a path relative to the base path. -or- Given an absolute base path, this extension searches for file references that are relative and converts them to absolute paths. MIT license. Copyright (c) 2014 - 2017 Isaac Muse <isaacmuse@gmail.com> Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ from __future__ import unicode_literals from markdown import Extension from markdown.postprocessors import Postprocessor from . import util import os import re from urllib.parse import urlunparse RE_TAG_HTML = r'''(?xus) (?: (?P<comments>(\r?\n?\s*)<!--[\s\S]*?-->(\s*)(?=\r?\n)|<!--[\s\S]*?-->)| (?P<open><(?P<tag>(?:%s))) (?P<attr>(?:\s+[\w\-:]+(?:\s*=\s*(?:"[^"]*"|'[^']*'))?)*) (?P<close>\s*(?:\/?)>) ) ''' RE_TAG_LINK_ATTR = re.compile( r'''(?xus) (?P<attr> (?: (?P<name>\s+(?:href|src)\s*=\s*) (?P<path>"[^"]*"|'[^']*') ) ) ''' ) def repl_relative(m, base_path, relative_path): """Replace path with relative path.""" link = m.group(0) try: scheme, netloc, path, params, query, fragment, is_url, is_absolute = util.parse_url(m.group('path')[1:-1]) if not is_url: # Get the absolute path of the file or return # if we can't resolve the path path = util.url2path(path) if (not is_absolute): # Convert current relative path to absolute path = os.path.relpath( os.path.normpath(os.path.join(base_path, path)), os.path.normpath(relative_path) ) # Convert the path, URL encode it, and format it as a link path = util.path2url(path) link = '%s"%s"' % ( m.group('name'), urlunparse((scheme, netloc, path, params, query, fragment)) ) except Exception: # pragma: no cover # Parsing crashed and burned; no need to continue. pass return link def repl_absolute(m, base_path): """Replace path with absolute path.""" link = m.group(0) try: scheme, netloc, path, params, query, fragment, is_url, is_absolute = util.parse_url(m.group('path')[1:-1]) if (not is_absolute and not is_url): path = util.url2path(path) path = os.path.normpath(os.path.join(base_path, path)) path = util.path2url(path) start = '/' if not path.startswith('/') else '' link = '%s"%s%s"' % ( m.group('name'), start, urlunparse((scheme, netloc, path, params, query, fragment)) ) except Exception: # pragma: no cover # Parsing crashed and burned; no need to continue. pass return link def repl(m, base_path, rel_path=None): """Replace.""" if m.group('comments'): tag = m.group('comments') else: tag = m.group('open') if rel_path is None: tag += RE_TAG_LINK_ATTR.sub(lambda m2: repl_absolute(m2, base_path), m.group('attr')) else: tag += RE_TAG_LINK_ATTR.sub(lambda m2: repl_relative(m2, base_path, rel_path), m.group('attr')) tag += m.group('close') return tag class PathConverterPostprocessor(Postprocessor): """Post process to find tag lings to convert.""" def run(self, text): """Find and convert paths.""" basepath = self.config['base_path'] relativepath = self.config['relative_path'] absolute = bool(self.config['absolute']) tags = re.compile(RE_TAG_HTML % '|'.join(self.config['tags'].split())) if not absolute and basepath and relativepath: text = tags.sub(lambda m: repl(m, basepath, relativepath), text) elif absolute and basepath: text = tags.sub(lambda m: repl(m, basepath), text) return text class PathConverterExtension(Extension): """PathConverter extension.""" def __init__(self, *args, **kwargs): """Initialize.""" self.config = { 'base_path': ["", "Base path used to find files - Default: \"\""], 'relative_path': ["", "Path that files will be relative to (not needed if using absolute) - Default: \"\""], 'absolute': [False, "Paths are absolute by default; disable for relative - Default: False"], 'tags': ["img script a link", "tags to convert src and/or href in - Default: 'img scripts a link'"] } super(PathConverterExtension, self).__init__(*args, **kwargs) def extendMarkdown(self, md): """Add post processor to Markdown instance.""" rel_path = PathConverterPostprocessor(md) rel_path.config = self.getConfigs() md.postprocessors.register(rel_path, "path-converter", 2) md.registerExtension(self) def makeExtension(*args, **kwargs): """Return extension.""" return PathConverterExtension(*args, **kwargs)
6,241
1,919
import os, string oldWin = '''span { font-family: Verdana; background: #e0e0d0; font-size: 10pt; } </style> </head> <body bgcolor="#e0e0d0"> ''' oldLinux = '''span { font-family: Verdana; background: #e0e0d0; font-size: 13pt; } </style> </head> <body bgcolor="#e0e0d0"> ''' new = '''span { font-family: Verdana; } </style> </head> <body bgcolor="#f0f0f8"> ''' def fixhtmlfile(file): if os.path.isfile(file) and file[-5:] == '.html': print file fp = open(file, 'rt') cont = fp.read() fp.close() cont = string.replace(cont, '\r\n', '\n') cont = string.replace(cont, oldWin, new) cont = string.replace(cont, oldLinux, new) fp = open(file, 'wt') fp.write(cont) fp.close() def fixhtmlfiles(dir): files = os.listdir(dir) for file in files: fixhtmlfile(dir + os.sep + file)
879
356
from typing import TYPE_CHECKING if TYPE_CHECKING: from .webapp import WebApp class StaticHeader: """Contains all relevant methods to a static header.""" id: int url: str name: str value: dict def __init__(self, resp: dict, webapp: 'WebApp') -> None: """ Initialize the class variables. :param dict resp: json dictionary :param webapp: WebApp class (see pyaww.webapp) """ self._webapp = webapp vars(self).update(resp) def delete(self) -> None: """Delete a static header.""" self._webapp.userclass.request( 'DELETE', f'/api/v0/user/{self._webapp.user}/webapps/{self._webapp.domain_name}/static_headers/{self.id}/', ) def update(self, **kwargs) -> None: """ Update a static header. Sample usage -> StaticHeader.update(...) :param kwargs: takes url, name, value """ self._webapp.userclass.request( 'PATCH', f'/api/v0/user/{self._webapp.user}/webapps/{self._webapp.domain_name}/static_headers/{self.id}/', data=kwargs ) vars(self).update(kwargs)
1,191
365
class Nodo: elemento = None Siguiente = None def __init__(self, elemento, siguiente): self.elemento = elemento self.Siguiente = siguiente class Pila: tamano = 0 top = None def apilar(self, elemento): """ Agrega un elemento al tope de la pila :param elemento: Cualquier elemento :return: None """ nuevo_nodo = Nodo(elemento, self.top) self.top = nuevo_nodo self.tamano += 1 def desapilar(self): """ Retorna el elemento del Tope de la pila y lo elimina :return: El elemento del tope de la pila """ if self.tamano > 0: elemento_auxiliar = self.top.elemento self.top = self.top.Siguiente self.tamano -= 1 return elemento_auxiliar raise IndexError('La pila esta vacía') def mirar(self): """ Ve el elemento del tope de la pila sin eliminarlo :return: El elemento del tope de la pila """ return self.top.elemento def es_vacia(self): return self.tamano == 0 def invertir(self): auxiliar = Pila() nodo_auxiliar = self.top for i in range(self.tamano): auxiliar.apilar(nodo_auxiliar.elemento) nodo_auxiliar = nodo_auxiliar.Siguiente return auxiliar def copiar(self): return self.invertir().invertir() def __repr__(self): resultado = [] auxiliar = self while not auxiliar.es_vacia(): resultado.append(auxiliar.desapilar()) resultado.reverse() return str(resultado)
1,654
527
# Copyright (c) OpenMMLab. All rights reserved. import copy import logging from collections import defaultdict from itertools import chain from torch.nn.utils import clip_grad from mmcv.utils import TORCH_VERSION, _BatchNorm, digit_version # from ..dist_utils import allreduce_grads # from ..fp16_utils import LossScaler, wrap_fp16_model from mmcv.runner.hooks import HOOKS, Hook try: # If PyTorch version >= 1.6.0, torch.cuda.amp.GradScaler would be imported # and used; otherwise, auto fp16 will adopt mmcv's implementation. from torch.cuda.amp import GradScaler except ImportError: pass @HOOKS.register_module() class MyHook(Hook): """A hook contains custom operations for the optimizer. Args: grad_clip (dict, optional): A config dict to control the clip_grad. Default: None. detect_anomalous_params (bool): This option is only used for debugging which will slow down the training speed. Detect anomalous parameters that are not included in the computational graph with `loss` as the root. There are two cases - Parameters were not used during forward pass. - Parameters were not used to produce loss. Default: False. """ def __init__(self, grad_clip=None, detect_anomalous_params=False): self.grad_clip = grad_clip self.detect_anomalous_params = detect_anomalous_params def clip_grads(self, params): params = list( filter(lambda p: p.requires_grad and p.grad is not None, params)) if len(params) > 0: return clip_grad.clip_grad_norm_(params, **self.grad_clip) def after_train_iter(self, runner): runner.optimizer.zero_grad() if self.detect_anomalous_params: self.detect_anomalous_parameters(runner.outputs['loss'], runner) if runner.outputs['loss'] is not None: runner.outputs['loss'].backward() if self.grad_clip is not None: grad_norm = self.clip_grads(runner.model.parameters()) if grad_norm is not None: # Add grad norm to the logger runner.log_buffer.update({'grad_norm': float(grad_norm)}, runner.outputs['num_samples']) runner.optimizer.step() def detect_anomalous_parameters(self, loss, runner): logger = runner.logger parameters_in_graph = set() visited = set() def traverse(grad_fn): if grad_fn is None: return if grad_fn not in visited: visited.add(grad_fn) if hasattr(grad_fn, 'variable'): parameters_in_graph.add(grad_fn.variable) parents = grad_fn.next_functions if parents is not None: for parent in parents: grad_fn = parent[0] traverse(grad_fn) traverse(loss.grad_fn) for n, p in runner.model.named_parameters(): if p not in parameters_in_graph and p.requires_grad: logger.log( level=logging.ERROR, msg=f'{n} with shape {p.size()} is not ' f'in the computational graph \n')
3,354
928
import pytest from news.reporters import ReporterMeta from news.reporters.feed import AtomReporter, RSSReporter @pytest.fixture def rss_reporter(sa_schedule, sa_backend): meta = ReporterMeta(schedule=sa_schedule) return RSSReporter(meta=meta, backend=sa_backend) @pytest.fixture def atom_reporter(sa_schedule, sa_backend): meta = ReporterMeta(schedule=sa_schedule) return AtomReporter(meta=meta, backend=sa_backend)
436
154
#!/usr/bin/env python import sys import os import ConfigParser import json import time import threading from distutils.core import setup import pika import boto import pusherclient from boto.sqs.message import RawMessage # set defaults channel = None global pusher config = ConfigParser.ConfigParser() class MQServer(threading.Thread): """ Message Queue Thread Listener (RabbitMQ, ZeroMQ, other MQs) """ def __init__(self): threading.Thread.__init__(self) def run(self): # get the items in the config endpoint = config.get('MQSERVER', 'endpoint') queue = config.get('MQSERVER', 'queue') # start the connection to the MQ server connection = pika.BlockingConnection(pika.ConnectionParameters(host=endpoint)) channel = connection.channel() channel.basic_consume(self.callback, queue=queue, no_ack=True) print '[*] Waiting for queue messages' # start the consumer for the channel while True: channel.start_consuming() def callback(self, ch, method, properties, body): print "[x] Received %r" % (body,) os.system(config.get('MQSERVER', 'command')) class AmazonSQS(threading.Thread): """ Amazon SQS Thread Listener """ def __init__(self): threading.Thread.__init__(self) def run(self): # start the connection to the SQS endpoint sqs = boto.sqs.connect_to_region(config.get('AWSSQS', 'region'), aws_access_key_id=config.get('AWSSQS', 'key'), aws_secret_access_key=config.get('AWSSQS', 'secret')) q = sqs.lookup(config.get('AWSSQS', 'queue')) q.set_message_class(RawMessage) results = sqs.receive_message(q, number_messages=1) print '[*] Waiting for Amazon SQS messages' last_received = time.time() # loop and look for messages while True: # loop through the results for result in results: # Check to see the last time and the last message offset = int(time.time() - last_received) if offset > 60: # You could set a command to do certain things based upon the config if str(result.get_body()) == 'somvalue': #DO something print str(result.get_body()) # execute your command from the config os.system(config.get('AWSSQS', 'command')) last_received = time.time() # re-get the messages results = sqs.receive_message(q, number_messages=1) time.sleep(30) # at 30 seconds, it's guaranteed to get it at least once class PusherWebSocket(threading.Thread): """ Pusher websocket Thread Listener """ def __init__(self): threading.Thread.__init__(self) def run(self): # get the items in the config key = config.get('PUSHER', 'key') secret = config.get('PUSHER', 'secret') app_id = config.get('PUSHER', 'app_id') # start the connection to the Pusher client pusher = pusherclient.Pusher(key, secret=secret) pusher.connection.bind('pusher:connection_established', self.connect_handler) print '[*] Waiting for Pusher messages' while True: time.sleep(1) def connect_handler(self, data): channel = pusher.subscribe(config.get('PUSHER', 'channel')) channel.bind(config.get('PUSHER', 'event'), self.channel_callback) print '[-] Connected to Pusher' def channel_callback(self, data): # execute your command from the config os.system(config.get('PUSHER', 'command')) def main(): print 'Starting listeners...To exit press CTRL+C' # create a list of threads threads = [] # open the config file config.readfp(open('deployer.cfg')) mq = MQServer() sqs = AmazonSQS() p = PusherWebSocket() try: if len(config.get('MQSERVER', 'endpoint')) > 1: mq.daemon = True # daemonize the thread threads.append(mq) # append the threads to the thread list mq.start() # start the thread if len(config.get('AWSSQS', 'key')) > 1: sqs.daemon = True # daemonize the thread threads.append(sqs) # append the threads to the thread list sqs.start() # start the thread if len(config.get('PUSHER', 'key')) > 1: p.daemon = True # daemonize the thread threads.append(p) # append the threads to the thread list p.start() # start the thread for thread in threads: # check to see the thread is still alive while thread.isAlive(): thread.join(1) except (KeyboardInterrupt, SystemExit): print '\n!Received keyboard interrupt, quitting threads.\n' if __name__ == '__main__': main()
4,945
1,456
# -*- coding: utf-8 -*- from __future__ import absolute_import from .diffusion import create_mrtrix_dti_pipeline from .connectivity_mapping import create_connectivity_pipeline from .group_connectivity import (create_group_connectivity_pipeline)
245
75
import src.lex import src.parse import src.eval import src.tools import src.optimize import src.pre import re import sys def main(file, n=None, first=False): import os debug = False version = "Beta 2.5" l = src.tools.readSembleFile(file) #if first: # l += src.pre.importf("base.smb", l)[0] l, fname = src.pre.process(l) with open("newcode.smb", "w") as newcode: newcode.write(l) x = src.lex.lex(l) with open("lexout.txt", "w") as fw: fw.write(str(x).replace("[", "[\n").replace("]", "]\n")) v = src.parse.parse(x) with open("parseout.txt", "w") as fw: fw.write(str(v).replace("[", "[\n").replace("]", "]\n")) src.eval.cmpf(v, "semble.asm") src.optimize.optimize("semble.asm") os.system("as --32 semble.asm -o semble.o") if n != None: fname = n os.system("ld -m elf_i386 -dynamic-linker /lib/ld-linux.so.2 -o " + fname + " semble.o -lc") if not debug: os.system("rm semble.o parseout.txt lexout.txt newcode.smb console.txt") if __name__ == '__main__': #print(src.lex.checkFuncCall("hello(helo,hello, h)")) #if re.match(r"^\[\d+\]\[(.*\,)*(.*)\]$", "[7][6, 6,]"): # print("hello") #print(src.lex.checkIndexRef("hello[5]")) #print(re.match(r"^(\[[^\n\]]+\])+$", "[56][65][67]")) #src.lex.getIndexs("[56][65][67]") #print(src.lex.checkIndexRef("people.names[3]")) #print(src.lex.lex("5 + 5 * 5\n")) #print("\n") #print(src.parse.organizeEquation(src.lex.lex("5 + 5 * 5 + 5 + 5\n"))) main(sys.argv[1]) print("Compiled Succesfully")
1,528
688
#!/usr/bin/env python """ Data browser for When's My Transport, with caching and JSON/XML parsing """ import json import logging import os import urllib2 import time from xml.dom.minidom import parseString from xml.etree.ElementTree import fromstring from lib.exceptions import WhensMyTransportException # # API URLs - for live APIs and test data we have cached for unit testing # HOME_DIR = os.path.dirname(os.path.abspath(__file__)) + '/..' URL_SETS = { 'live': { 'BUS_URL': "http://countdown.tfl.gov.uk/stopBoard/%s", 'DLR_URL': "http://www.dlrlondon.co.uk/xml/mobile/%s.xml", 'TUBE_URL': "http://cloud.tfl.gov.uk/TrackerNet/PredictionDetailed/%s/%s", 'STATUS_URL': "http://cloud.tfl.gov.uk/TrackerNet/StationStatus/IncidentsOnly", }, 'test': { 'BUS_URL': "file://" + HOME_DIR + "/tests/data/bus/%s.json", 'DLR_URL': "file://" + HOME_DIR + "/tests/data/dlr/%s.xml", 'TUBE_URL': "file://" + HOME_DIR + "/tests/data/tube/%s-%s.xml", 'STATUS_URL': "file://" + HOME_DIR + "/tests/data/tube/status.xml", } } CACHE_MAXIMUM_AGE = 30 # 30 seconds maximum cache age class WMTURLProvider: """ Simple wrapper that provides URLs for the TfL APIs, or test data depending on how we have set this up """ #pylint: disable=R0903 def __init__(self, use_test_data=False): if use_test_data: self.urls = URL_SETS['test'] else: self.urls = URL_SETS['live'] def __getattr__(self, key): return self.urls[key] class WMTBrowser: """ A simple JSON/XML fetcher with caching. Not designed to be used for many thousands of URLs, or for concurrent access """ def __init__(self): self.opener = urllib2.build_opener() self.opener.addheaders = [('User-agent', 'When\'s My Transport?'), ('Accept', 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8')] logging.debug("Starting up browser") self.cache = {} def fetch_url(self, url, default_exception_code): """ Fetch a URL and returns the raw data as a string """ # If URL is in cache and still considered fresh, fetch that if url in self.cache and (time.time() - self.cache[url]['time']) < CACHE_MAXIMUM_AGE: logging.debug("Using cached URL %s", url) url_data = self.cache[url]['data'] # Else fetch URL and store else: logging.debug("Fetching URL %s", url) try: response = self.opener.open(url) url_data = response.read() self.cache[url] = {'data': url_data, 'time': time.time()} # Handle browsing error except urllib2.HTTPError, exc: logging.error("HTTP Error %s reading %s, aborting", exc.code, url) raise WhensMyTransportException(default_exception_code) except Exception, exc: logging.error("%s (%s) encountered for %s, aborting", exc.__class__.__name__, exc, url) raise WhensMyTransportException(default_exception_code) return url_data def fetch_json(self, url, default_exception_code='tfl_server_down'): """ Fetch a JSON URL and returns Python object representation of it """ json_data = self.fetch_url(url, default_exception_code) if json_data: try: obj = json.loads(json_data) return obj # If the JSON parser is choking, probably a 503 Error message in HTML so raise a ValueError except ValueError, exc: del self.cache[url] logging.error("%s encountered when parsing %s - likely not JSON!", exc, url) raise WhensMyTransportException(default_exception_code) else: return None def fetch_xml_tree(self, url, default_exception_code='tfl_server_down'): """ Fetch an XML URL and returns Python object representation of it as an ElementTree """ xml_data = self.fetch_url(url, default_exception_code) if xml_data: try: tree = fromstring(xml_data) namespace = '{%s}' % parseString(xml_data).firstChild.getAttribute('xmlns') # Remove horrible namespace functionality if namespace: for elem in tree.getiterator(): if elem.tag.startswith(namespace): elem.tag = elem.tag[len(namespace):] return tree # If the XML parser is choking, probably a 503 Error message in HTML so raise a ValueError except Exception, exc: del self.cache[url] logging.error("%s encountered when parsing %s - likely not XML!", exc, url) raise WhensMyTransportException(default_exception_code) else: return None
5,005
1,454
import colorsys import numpy as np # ----------------------------------------------------------------------------- # Utilities to generate a HSL colourmap from the magnetisation field data def convert_to_RGB(hls_color): return np.array(colorsys.hls_to_rgb(hls_color[0] / (2 * np.pi), hls_color[1], hls_color[2])) def generate_colours(field_data, colour_model='rgb'): """ field_data :: (n, 3) array """ hls = np.ones_like(field_data) hls[:, 0] = np.arctan2(field_data[:, 1], field_data[:, 0] ) hls[:, 0][hls[:, 0] < 0] = hls[:, 0][hls[:, 0] < 0] + 2 * np.pi hls[:, 1] = 0.5 * (field_data[:, 2] + 1) if colour_model == 'rgb': rgbs = np.apply_along_axis(convert_to_RGB, 1, hls) return rgbs elif colour_model == 'hls': return hls else: raise Exception('Specify a valid colour model: rgb or hls') # -----------------------------------------------------------------------------
1,099
366
class Solution(object): def deleteDuplicates(self, head): initial = head while head: if head.next and head.val == head.next.val: head.next = head.next.next else: head = head.next head = initial return head
299
78
''' Created on Wed Sep 18 2019 @author XXX ''' import numpy as np import scipy.sparse as sps import time import sys import copy from tqdm import tqdm from enum import Enum from RecSysFramework.Utils import seconds_to_biggest_unit from RecSysFramework.Utils.compute_popularity import compute_popularity_user from RecSysFramework.Utils.WriteTextualFile import WriteTextualFile from RecSysFramework.Utils.Log import dict_to_string from scipy.stats import spearmanr import rbo class Comparator(object): """ Abstract comprator Used in general for statistics that take into account more recommenders jointly. eg: variety of recommendations, it measures the difference between items recommended. """ COMPARATOR_NAME = "Comparator" metrics = ['jaccard', 'spearman', 'RBO'] def __init__(self, URM_test, recommenders_to_compare, cutoff=5, metrics_list=['jaccard', 'spearman', 'RBO'], minRatingsPerUser=1, exclude_seen=True, exclude_top_pop=False, save_to_file=True, verbose=True): """ Arguments: recommenders_to_compare {RecommenderBase} -- instances to be compared URM_test {[type]} -- ground thruth Keyword Arguments: cutoff_list {list} -- cutoff for the list of recommendations to consider metrics_list {list} -- list of metrics to be computed (default: []) minRatingsPerUser {int} -- recommendations are provided only for users with at least this values of ratings (default: {1}) exclude_seen {bool} -- shall we exlude from the recommendations the items seen at training time? (default: {True}) exclude_top_pop {bool} -- shall we exlude from the recommendations the most popular items? (default: {False}) save_to_file {bool} -- if True, the results are saved into a file, otherwise just printed """ super(Comparator, self).__init__() self.minRatingsPerUser = minRatingsPerUser self.exclude_seen = exclude_seen self.metrics_list = metrics_list self.exclude_top_pop = exclude_top_pop self.recommenders_to_compare = recommenders_to_compare self.cutoff = cutoff self.save_to_file = save_to_file self.verbose = verbose if save_to_file: recs = '' for r in recommenders_to_compare: recs += r.RECOMMENDER_NAME + '_' self.file = WriteTextualFile( 'ComparationResults', recs, append_datetime=True) if URM_test is not None: self.URM_test = sps.csr_matrix(URM_test) self.n_users, self.n_items = URM_test.shape numRatings = np.ediff1d(self.URM_test.tocsr().indptr) self.usersToEvaluate = np.arange( self.n_users)[numRatings >= self.minRatingsPerUser] def _evaluate_recommender(self, recommender_object): """ should return [{user_x: [rec_items_for_x], user_y: [rec_items_for_y]}, {user_i: [rec_items_for_i], user_j: [rec_items_for_j]}, ... ], [descr1, descr2, ... ] A subclass will give an implementation to this method """ pass def _print(self, text, only_on_file=False): if self.verbose: print(text) if self.save_to_file: self.file.write_line(text) def compare(self): """compares the recommenders according to the metrics defined Returns: [string] -- string describing the results [dict] -- the format is {'descr': descr, 'cutoff': cutoff, 'value': value} """ return_string = '' return_dict = {} for m in self.metrics_list: assert m in self.metrics, "metric provided should be among {}".format( self.metrics) return_string += 'cutoff: {}\n'.format(str(self.cutoff)) return_string += 'computing {}. considering only the first two recommenders provided\n'.format( m) if hasattr(self.recommenders_to_compare[0], 'model_parameters') and hasattr(self.recommenders_to_compare[1], 'model_parameters'): return_string += 'recommender 1: {}\n\n'.format(dict_to_string( self.recommenders_to_compare[0].model_parameters, style='constructor')) return_string += 'recommender 2: {}\n\n'.format(dict_to_string( self.recommenders_to_compare[1].model_parameters, style='constructor')) evaluation_results_first, descr = self._evaluate_recommender( self.recommenders_to_compare[0]) evaluation_results_second, _ = self._evaluate_recommender( self.recommenders_to_compare[1]) for idx in range(len(evaluation_results_second)): if m == 'jaccard': result = self.compute_jaccard( evaluation_results_first[idx], evaluation_results_second[idx]) elif m == 'RBO': result = self.compute_RBO( evaluation_results_first[idx], evaluation_results_second[idx]) return_string += '{} on {}: {}\n'.format( m, descr[idx], result) return_dict['{}_{}_{}'.format( m, descr[idx], self.cutoff)] = result self._print(return_string) return return_string, return_dict def compute_jaccard(self, d1, d2): j = [] for key, l1 in d1.items(): s1 = set(l1) s2 = set(d2[key]) j.append(len(s1 & s2)/len(s1 | s2)) return sum(j)/len(j) def compute_RBO(self, d1, d2): j = [] for key, l1 in d1.items(): l2 = d2[key] j.append(rbo.RankingSimilarity(l1, l2).rbo()) return sum(j)/len(j) class ComparatorHoldout(Comparator): """ComparatorHoldout""" EVALUATOR_NAME = "ComparatorHoldout" def _evaluate_recommender(self, recommender_object): recommended_items_batch_list, _ = recommender_object.recommend(self.usersToEvaluate, remove_seen_flag=self.exclude_seen, cutoff=self.cutoff, remove_top_pop_flag=self.exclude_top_pop, return_scores=True) return [dict(zip(self.usersToEvaluate, recommended_items_batch_list))], ['all_users'] class ComparatorHoldoutUserPopularity(Comparator): """ComparatorHoldoutUserPopularity evaluates the recommender considering different sets of users based on their popularity """ EVALUATOR_NAME = "ComparatorHoldoutUserPopularity" def __init__(self, URM_train, *pos_args, cuts=[0.01, 0.05, 0.1, 0.2, 0.3, 0.4, 0.6, 0.8, 1], **key_args): self.URM_train = URM_train self.cuts = cuts super(ComparatorHoldoutUserPopularity, self).__init__(*pos_args, **key_args) def _recommend_in_batch(self, recommender_object, users, remove_seen_flag, cutoff, remove_top_pop_flag): r = [] size = 1000 n_users = len(users) n_batch = n_users // size for idx in range(n_batch): r += recommender_object.recommend( users[size*idx: size*(idx+1)], remove_seen_flag=remove_seen_flag, cutoff=cutoff, remove_top_pop_flag=remove_top_pop_flag, return_scores=True)[0] r += recommender_object.recommend( users[(size*n_batch) % n_users: n_users], remove_seen_flag=remove_seen_flag, cutoff=cutoff, remove_top_pop_flag=remove_top_pop_flag, return_scores=True)[0] return r def _evaluate_recommender(self, recommender_object): pop = compute_popularity_user(self.URM_train, ordered=True) r = [] descr = [] users, interactions = zip(*pop) users = np.array(users) interactions = np.array(interactions) cum_sum_interactions = np.cumsum(interactions) tot_interactions = np.sum(interactions) recommended_items_all_users = self._recommend_in_batch(recommender_object, np.sort(users), self.exclude_seen, self.cutoff, self.exclude_top_pop, ) recommended_items_all_users = np.array(recommended_items_all_users) for cut in self.cuts: users_in_cut = users[cum_sum_interactions < cut*tot_interactions] recommended_items_batch_list = recommended_items_all_users[users_in_cut, :].tolist() r.append(dict(zip(users_in_cut, recommended_items_batch_list))) descr.append('{}'.format(cut)) return r, descr
9,165
2,720
import numpy as np class CellularAutomationModel: grid_width = 40 grid_height = 40 def __init__(self): self.grid = self._randomised_grid() def evolve(self): """ Evolve the current grid state using Conway's Game of Life algorithm. :returns dict: A dictionary representation of the state of cells in the grid """ base_grid = self.grid.copy() for y in range(self.grid_height): for x in range(self.grid_width): cell_state = base_grid[x, y] n_neighbours = self._calculate_alive_neighbours(x, y, cell_state, grid=base_grid) self.grid[x, y] = self._next_cell_state(cell_state, n_neighbours) return self._json_formatted_grid() def toggle_cell_state(self, x, y): """ Reverses the cell state for a particular cell coordinate. """ self.grid[x][y] = 0 if self.grid[x][y] == 1 else 1 def reset_grid(self): """ Resets the grid array to a random state. :returns dict: A dictionary representation of the state of cells in the grid """ self.grid = self._randomised_grid() return self._json_formatted_grid() def _calculate_alive_neighbours(self, x, y, cell_state, grid): """ Returns the number of alive nearest neighbours. """ surrounding_arr = self._surrounding_arr(x, y, grid) n_alive = sum(sum(surrounding_arr)) return n_alive - cell_state def _json_formatted_grid(self): """ Returns a python dictionary which represents the current state of the cells in the grid. key: An integer that represents a single cell based on the coordinate position. value: The cell state <0 or 1> to represent whether a cell is dead or alive. """ json_grid = {} for x in range(self.grid_width): for y in range(self.grid_height): cell_id = int(x + y*self.grid_width) json_grid[cell_id] = int(self.grid[x, y]) return json_grid def _randomised_grid(self): """ Returns a 2d array with values of randomly assigned values of 0 or 1. """ return np.random.randint(2, size=(self.grid_height, self.grid_width)) @staticmethod def _surrounding_arr(x, y, grid): """ Returns an 2d array containing all the adjacent cells for a particular coordinate (radius = 1 cell). """ if x != 0 and y != 0: return grid[x - 1:x + 2, y - 1:y + 2] elif x == 0: return grid[x:x + 2, y - 1:y + 2] elif y == 0: return grid[x - 1:x + 2, y:y + 2] else: return grid[x:x + 2, y:y + 2] @staticmethod def _next_cell_state(cell_state, n_neighbours): """ Returns the new cell state 0 (dead) or 1 (alive). New state is determined using the current cell state and number of alive neighbours based on the rules in Conway's Game of Life. """ if (cell_state == 1 and (n_neighbours not in range(2, 4))) or (cell_state == 0 and n_neighbours != 3): return 0 return 1
3,233
1,009
from encoder import Encoder from decoder import Decoder from parser import Parser from baseline import * from language_model import LanguageModel from util import Reader import dynet as dy from misc import compute_eval_score, compute_perplexity import os initializers = {'glorot': dy.GlorotInitializer(), 'constant': dy.ConstInitializer(0.01), 'uniform': dy.UniformInitializer(0.1), 'normal': dy.NormalInitializer(mean = 0, var = 1) } optimizers = { "sgd": dy.SimpleSGDTrainer, "adam": dy.AdamTrainer, "adadelta": dy.AdadeltaTrainer, "adagrad": dy.AdagradTrainer } class Session(object): def __init__(self, options): self.reader = Reader(options.data_dir, options.data_augment) self.options = options def supervised_enc(self): encoder = self.create_encoder() if os.path.exists(self.options.result_dir + 'model_enc'): self.load_encoder(encoder) enc_trainer = optimizers[self.options.optimizer](encoder.model) lr = self.options.lr #used only for sgd i = 0 best_f1 = 0 print ('supervised training for encoder...') for epoch in range(self.options.epochs): sents = 0 total_loss = 0.0 train = self.reader.next_example(0) train_size = len(self.reader.data[0]) for data in train: s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] loss = encoder.train(s1, s2, s3, pos, act, self.options.enc_dropout) sents += 1 if loss is not None: total_loss += loss.scalar_value() loss.backward() if self.options.optimizer == 'sgd': enc_trainer.update(lr) else: enc_trainer.update() e = float(i) / train_size if i % self.options.print_every == 0: print('epoch {}: loss per sentence: {}'.format(e, total_loss / sents)) sents = 0 total_loss = 0.0 if i!=0 and i % self.options.save_every == 0: print('computing loss on validation set...') valid = self.reader.next_example(2) #fix this valid_size = len(self.reader.data[2]) rf = open(self.options.result_dir+'result', 'w') for vdata in valid: s1, s2, s3, pos, act = vdata[0], vdata[1], vdata[2], vdata[3], vdata[4] _, output, _ = encoder.parse(s1, s2, s3, pos) rf.write(output + '\n') rf.close() f1 = compute_eval_score(self.options.result_dir) if f1 > best_f1: best_f1 = f1 print ('highest f1: {}'.format(f1)) print ('saving model...') encoder.Save(self.options.result_dir + 'model_enc') else: lr = lr * self.options.decay i += 1 def supervised_dec(self): decoder = self.create_decoder() if os.path.exists(self.options.result_dir + 'model_dec'): self.load_decoder(decoder) dec_trainer = optimizers[self.options.optimizer](decoder.model) lr = self.options.lr #used only for sgd i = 0 lowest_valid_loss = 9999 print ('supervised training for decoder...') for epoch in range(self.options.epochs): sents = 0 total_loss = 0.0 train = self.reader.next_example(0) train_size = len(self.reader.data[0]) for data in train: s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] loss, loss_act, loss_word = decoder.compute_loss(s3, act, self.options.dec_dropout) sents += 1 if loss is not None: total_loss += loss.scalar_value() loss.backward() if self.options.optimizer == 'sgd': dec_trainer.update(lr) else: dec_trainer.update() e = float(i) / train_size if i % self.options.print_every == 0: print('epoch {}: loss per sentence: {}'.format(e, total_loss / sents)) sents = 0 total_loss = 0.0 if i!=0 and i % self.options.save_every == 0: print('computing loss on validation set...') total_valid_loss = 0 valid = self.reader.next_example(1) valid_size = len(self.reader.data[1]) for vdata in valid: s1, s2, s3, pos, act = vdata[0], vdata[1], vdata[2], vdata[3], vdata[4] valid_loss, _, _ = decoder.compute_loss(s3, act) if valid_loss is not None: total_valid_loss += valid_loss.scalar_value() total_valid_loss = total_valid_loss * 1.0 / valid_size if total_valid_loss < lowest_valid_loss: lowest_valid_loss = total_valid_loss print ('saving model...') decoder.Save(self.options.result_dir + 'model_dec') else: lr = lr * self.options.decay i += 1 def unsupervised_with_baseline(self): decoder = self.create_decoder() assert(os.path.exists(self.options.result_dir + 'model_dec')) self.load_decoder(decoder) encoder = self.create_encoder() assert(os.path.exists(self.options.result_dir + 'model_enc')) self.load_encoder(encoder) baseline = self.create_baseline() if os.path.exists(self.options.result_dir + 'baseline'): self.load_baseline(baseline) enc_trainer = optimizers[self.options.optimizer](encoder.model) dec_trainer = optimizers[self.options.optimizer](decoder.model) baseline_trainer = optimizers[self.options.optimizer](baseline.model) lr = self.options.lr #used only for sgd i = 0 lowest_valid_loss = 9999 print ('unsupervised training...') for epoch in range(self.options.epochs): sents = 0 total_loss = 0.0 train = self.reader.next_example(0) train_size = len(self.reader.data[0]) for data in train: s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] sents += 1 # random sample enc_loss_act, _, act = encoder.parse(s1, s2, s3, pos, sample=True) _, dec_loss_act, dec_loss_word = decoder.compute_loss(s3, act) # save reward logpx = -dec_loss_word.scalar_value() total_loss -= logpx # reconstruction and regularization loss backprop to theta_d dec_loss_total = dec_loss_word + dec_loss_act * dy.scalarInput(self.options.dec_reg) dec_loss_total = dec_loss_total * dy.scalarInput(1.0 / self.options.mcsamples) dec_loss_total.scalar_value() dec_loss_total.backward() # update decoder if self.options.optimizer == 'sgd': dec_trainer.update(lr) else: dec_trainer.update() if self.options.enc_update > 0: # compute baseline and backprop to theta_b b = baseline(s3) logpxb = b.scalar_value() b_loss = dy.squared_distance(b, dy.scalarInput(logpx)) b_loss.value() b_loss.backward() # update baseline if self.options.optimizer == 'sgd': baseline_trainer.update(lr) else: baseline_trainer.update() # policy and and regularization loss backprop to theta_e enc_loss_act = encoder.train(s1, s2, s3, pos, act) enc_loss_policy = enc_loss_act * dy.scalarInput((logpx - logpxb) / len(s1)) enc_loss_total = enc_loss_policy * dy.scalarInput(self.options.enc_update) - enc_loss_act * dy.scalarInput(self.options.enc_reg) enc_loss_total = enc_loss_total * dy.scalarInput(1.0 / self.options.mcsamples) enc_loss_total.value() enc_loss_total.backward() # update encoder if self.options.optimizer == 'sgd': enc_trainer.update(lr) else: enc_trainer.update() e = float(i) / train_size if i % self.options.print_every == 0: print('epoch {}: loss per sentence: {}'.format(e, total_loss / sents)) sents = 0 total_loss = 0.0 if i!=0 and i % self.options.save_every == 0: print('computing loss on validation set...') total_valid_loss = 0 valid = self.reader.next_example(1) valid_size = len(self.reader.data[1]) for vdata in valid: s1, s2, s3, pos, act = vdata[0], vdata[1], vdata[2], vdata[3], vdata[4] _, _, valid_word_loss = decoder.compute_loss(s3, act) if valid_word_loss is not None: total_valid_loss += valid_word_loss.scalar_value() total_valid_loss = total_valid_loss * 1.0 / valid_size if total_valid_loss < lowest_valid_loss: lowest_valid_loss = total_valid_loss print ('saving model...') encoder.Save(self.options.result_dir + 'model_enc') decoder.Save(self.options.result_dir + 'model_dec') baseline.Save(self.options.result_dir + 'baseline') else: lr = lr * self.options.decay i += 1 def unsupervised_without_baseline(self): decoder = self.create_decoder() assert(os.path.exists(self.options.result_dir + 'model_dec')) self.load_decoder(decoder) encoder = self.create_encoder() assert(os.path.exists(self.options.result_dir + 'model_enc')) self.load_encoder(encoder) enc_trainer = optimizers[self.options.optimizer](encoder.model) dec_trainer = optimizers[self.options.optimizer](decoder.model) lr = self.options.lr #used only for sgd i = 0 lowest_valid_loss = 9999 print ('unsupervised training...') for epoch in range(self.options.epochs): sents = 0 total_loss = 0.0 train = self.reader.next_example(0) train_size = len(self.reader.data[0]) for data in train: s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] sents += 1 # max sample enc_loss_act, _, act = encoder.parse(s1, s2, s3, pos, sample=False) _, dec_loss_act, dec_loss_word = decoder.compute_loss(s3, act) logpxb = -dec_loss_word.scalar_value() total_loss -= logpxb # random sample enc_loss_act, _, act = encoder.parse(s1, s2, s3, pos, sample=True) _, dec_loss_act, dec_loss_word = decoder.compute_loss(s3, act) # save reward logpx = -dec_loss_word.scalar_value() # reconstruction and regularization loss backprop to theta_d dec_loss_total = dec_loss_word + dec_loss_act * dy.scalarInput(self.options.dec_reg) dec_loss_total = dec_loss_total * dy.scalarInput(1.0 / self.options.mcsamples) dec_loss_total.scalar_value() dec_loss_total.backward() # update decoder if self.options.optimizer == 'sgd': dec_trainer.update(lr) else: dec_trainer.update() if self.options.enc_update > 0: # policy and and regularization loss backprop to theta_e enc_loss_act = encoder.train(s1, s2, s3, pos, act) enc_loss_policy = enc_loss_act * dy.scalarInput((logpx - logpxb) / len(s1)) enc_loss_total = enc_loss_policy * dy.scalarInput(self.options.enc_update) - enc_loss_act * dy.scalarInput(self.options.enc_reg) enc_loss_total = enc_loss_total * dy.scalarInput(1.0 / self.options.mcsamples) enc_loss_total.value() enc_loss_total.backward() if self.options.optimizer == 'sgd': enc_trainer.update(lr) else: enc_trainer.update() e = float(i) / train_size if i % self.options.print_every == 0: print('epoch {}: loss per sentence: {}'.format(e, total_loss / sents)) sents = 0 total_loss = 0.0 if i!=0 and i % self.options.save_every == 0: print('computing loss on validation set...') total_valid_loss = 0 valid = self.reader.next_example(1) valid_size = len(self.reader.data[1]) for vdata in valid: s1, s2, s3, pos, act = vdata[0], vdata[1], vdata[2], vdata[3], vdata[4] _, _, valid_word_loss = decoder.compute_loss(s3, act) if valid_word_loss is not None: total_valid_loss += valid_word_loss.scalar_value() total_valid_loss = total_valid_loss * 1.0 / valid_size if total_valid_loss < lowest_valid_loss: lowest_valid_loss = total_valid_loss print ('saving model...') encoder.Save(self.options.result_dir + 'model_enc') decoder.Save(self.options.result_dir + 'model_dec') else: lr = lr * self.options.decay i += 1 def pretrain_baseline(self): baseline = self.create_baseline() if os.path.exists(self.options.result_dir + 'baseline'): self.load_baseline(baseline) baseline_trainer = optimizers[self.options.optimizer](baseline.model) lr = self.options.lr #used only for sgd i = 0 lowest_valid_loss = 9999 print ('train baseline, for simplicity use the same data here') for epoch in range(self.options.epochs): sents = 0 total_loss = 0.0 train = self.reader.next_example(0) train_size = len(self.reader.data[0]) for data in train: s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] sents += 1 loss = -baseline(s3) if loss is not None: total_loss += loss.scalar_value() loss.backward() if self.options.optimizer == 'sgd': baseline_trainer.update(lr) else: baseline_trainer.update() e = float(i) / train_size if i % self.options.print_every == 0: print('epoch {}: loss per sentence: {}'.format(e, total_loss / sents)) sents = 0 total_loss = 0.0 if i!=0 and i % self.options.save_every == 0: print('computing loss on validation set...') total_valid_loss = 0 valid = self.reader.next_example(1) valid_size = len(self.reader.data[1]) for vdata in valid: s1, s2, s3, pos, act = vdata[0], vdata[1], vdata[2], vdata[3], vdata[4] valid_loss = -baseline(s3) if valid_loss is not None: total_valid_loss += valid_loss.scalar_value() total_valid_loss = total_valid_loss * 1.0 / valid_size if total_valid_loss < lowest_valid_loss: lowest_valid_loss = total_valid_loss print ('saving model...') baseline.Save(self.options.result_dir + 'baseline') else: lr = lr * self.options.decay i += 1 def parsing(self): decoder = self.create_decoder() assert(os.path.exists(self.options.result_dir + 'model_dec')) self.load_decoder(decoder) encoder = self.create_encoder() assert(os.path.exists(self.options.result_dir + 'model_enc')) self.load_encoder(encoder) print('parsing...') rf = open(os.path.join(self.options.result_dir, 'result'), 'w') test = self.reader.next_example(2) p = Parser(encoder, decoder) for dataid, data in enumerate(test): s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] output = p(s1, s2, s3, pos, self.options.nsamples) rf.write(output + '\n') rf.close() f1 = compute_eval_score(self.options.result_dir) print('bracket F1 score is {}'.format(f1)) def language_modeling(self): decoder = self.create_decoder() assert(os.path.exists(self.options.result_dir + 'model_dec')) self.load_decoder(decoder) encoder = self.create_encoder() assert(os.path.exists(self.options.result_dir + 'model_enc')) self.load_encoder(encoder) print('computing language model score...') test = self.reader.next_example(2) lm = LanguageModel(encoder, decoder) total_ll = 0 total_tokens = 0 for dataid, data in enumerate(test): s1, s2, s3, pos, act = data[0], data[1], data[2], data[3], data[4] if len(s1) <= 1: continue total_ll += lm(s1, s2, s3, pos, self.options.nsamples) total_tokens += len(s1) perp = compute_perplexity(total_ll, total_tokens) print('perplexity: {}'.format(perp)) def create_decoder(self): return Decoder(self.reader, self.options.nlayers, self.options.word_dim, self.options.pretrained_dim, self.options.action_dim, self.options.dec_lstm_dim, self.options.embedding_file) def create_encoder(self): return Encoder(self.reader, self.options.nlayers, self.options.word_dim, self.options.pretrained_dim, self.options.pos_dim, self.options.action_dim, self.options.enc_lstm_dim, self.options.embedding_file) def create_baseline(self): baseline = None if self.options.baseline == 'rnnlm': baseline = LanguageModelBaseline(self.reader, self.options.word_dim, self.options.pretrained_dim, self.options.dec_lstm_dim, self.options.embedding_file) elif self.options.baseline == 'rnnauto': baseline = RNNAutoencBaseline(self.reader, self.options.word_dim, self.options.pretrained_dim, self.options.dec_lstm_dim, self.options.embedding_file) elif self.options.baseline == 'mlp': baseline = MLPAutoencBaseline(self.reader, self.options.word_dim, self.options.pretrained_dim, self.options.embedding_file) else: raise NotImplementedError("Baseline Not Implmented") return baseline def load_decoder(self, decoder): decoder.Load(self.options.result_dir + 'model_dec') def load_encoder(self, encoder): encoder.Load(self.options.result_dir + 'model_enc') def load_baseline(self, baseline): baseline.Load(self.options.result_dir + 'baseline')
21,365
6,285
import random def mutacion(indiv): alea_1=random.randint(0,len(indiv)-1) alea_2 = random.randint(0, len(indiv)-1) interc_1=indiv[alea_1] interc_2=indiv[alea_2] indiv[alea_1] = interc_2 indiv[alea_2] = interc_1 return indiv
254
122
#!/usr/bin/env python # Copyright (c) 2019 Riverbed Technology, Inc. # # This software is licensed under the terms and conditions of the MIT License # accompanying the software ("License"). This software is distributed "AS IS" # as set forth in the License. import csv import sys import string import optparse from collections import defaultdict from steelscript.netprofiler.core.app import NetProfilerApp from steelscript.netprofiler.core.hostgroup import HostGroupType, HostGroup from steelscript.commands.steel import prompt_yn from steelscript.common.exceptions import RvbdException # This script will take a file with subnets and SiteNames # and create a HostGroupType on the target NetProfiler. # If the HostGroupType already exists, it will be deleted, # before creating a new one with the same name. # # See the EXAMPLE text below for the format of the input # file. Note that multiple SiteNames with different # IP address spaces can be included. EXAMPLE_WARN = """ Invalid file format Ensure file has correct header. example file: subnet SiteName 10.143.58.64/26 CZ-Prague-HG 10.194.32.0/23 MX-SantaFe-HG 10.170.55.0/24 KR-Seoul-HG 10.234.9.0/24 ID-Surabaya-HG 10.143.58.63/23 CZ-Prague-HG """ class HostGroupImport(NetProfilerApp): def add_options(self, parser): super(HostGroupImport, self).add_options(parser) group = optparse.OptionGroup(parser, "HostGroup Options") group.add_option('--hostgroup', action='store', help='Name of hostgroup to overwrite') group.add_option('-i', '--input-file', action='store', help='File path to hostgroup file') parser.add_option_group(group) def validate_args(self): """Ensure all arguments are present.""" super(HostGroupImport, self).validate_args() if not self.options.input_file: self.parser.error('Host group file is required, specify with ' '"-i" or "--input-file"') if not self.options.hostgroup: self.parser.error('Hostgroup name is required, specify with ' '"--hostgroup"') def validate(self, name): valid = set(string.letters + string.digits + '.-_') return all(c in valid for c in name) def import_file(self): """Process the input file and load into dict.""" groups = defaultdict(list) with open(self.options.input_file, 'rb') as f: dialect = csv.Sniffer().sniff(f.read(1024)) f.seek(0) reader = csv.reader(f, dialect) header = reader.next() if header != ['subnet', 'SiteName']: print(EXAMPLE_WARN) for i, row in enumerate(reader): cidr, group = row if not self.validate(group): print('Invalid group name on line {0}: {1}' ''.format(i+2, group)) sys.exit() groups[group].append(cidr) return groups def update_hostgroups(self, groups): """Replace existing HostGroupType with contents of groups dict.""" # First find any existing HostGroupType try: hgtype = HostGroupType.find_by_name(self.netprofiler, self.options.hostgroup) hgtype.config = [] hgtype.groups = {} print('Existing HostGroupType "{0}" found.' ''.format(self.options.hostgroup)) except RvbdException: print('No existing HostGroupType found, creating a new one.') hgtype = HostGroupType.create(self.netprofiler, self.options.hostgroup) # Add new values for group, cidrs in groups.items(): hg = HostGroup(hgtype, group) hg.add(cidrs) # Save to NetProfiler hgtype.save() print ('HostGroupType "%s" configuration saved.' % self.options.hostgroup) def main(self): """Confirm overwrite then update hostgroups.""" confirm = ('The contents of hostgroup {0} will be overwritten ' 'by the file {1}, are you sure?' ''.format(self.options.hostgroup, self.options.input_file)) if not prompt_yn(confirm): print('Okay, aborting.') sys.exit() groups = self.import_file() self.update_hostgroups(groups) print('Successfully updated {0} on {1}'.format(self.options.hostgroup, self.netprofiler.host)) if __name__ == '__main__': HostGroupImport().run()
4,724
1,374
from clint.textui import * def Write(): i = 0 while True: i += 1 text = input(colored.cyan("%d: "%i)) if text == "q": sys.exit(0) def Print(): print(colored.cyan("-- Bug list --")) print(colored.cyan("-- ") + colored.red("<q:quit>") + colored.cyan(" --")) Write() Print()
330
121
from datasets.hscic.hscic_datasets import scrape as datasets_scrape from datasets.hscic.hscic_indicators import scrape as indicators_scrape def main(workspace): datasets_scrape(workspace) indicators_scrape(workspace)
229
84
import copy import json import requests import datetime from src.tests.behave.common.config import get_conflagration class APIClient(object): def __init__(self, base_url, request_kwargs=None, client_kwargs=None): self.cfg = get_conflagration() token = self.get_identity_token() timeout = None ssl_certificate_verify = False verbose = True if request_kwargs: timeout = request_kwargs.get("timeout", timeout) ssl_certificate_verify = request_kwargs.get( "ssl_certificate_verify", ssl_certificate_verify ) verbose = request_kwargs.get("verbose", verbose) self.base_url = base_url self.headers = { "Content-Type": "application/json", "X-Auth-Token": token, } if client_kwargs: if not client_kwargs["authorized"]: self.headers["X-Auth-Token"] = "" self.verify = ssl_certificate_verify self.verbose = verbose self.request_kwargs = dict() self.request_kwargs["url"] = self.base_url self.request_kwargs["headers"] = self.headers self.request_kwargs["verify"] = self.verify if timeout: self.request_kwargs["timeout"] = timeout def get( self, resource_id=None, params=None, url=None, url_suffix=None, headers=None, ): request_kwargs = copy.deepcopy(self.request_kwargs) if url: request_kwargs["url"] = url if resource_id: resource_url = request_kwargs["url"] request_kwargs["url"] = "{0}/{1}".format(resource_url, resource_id) if url_suffix: resource_url = request_kwargs["url"] request_kwargs["url"] = "{0}/{1}".format(resource_url, url_suffix) if params: request_kwargs["params"] = params if headers: request_kwargs["headers"].update(headers) resp = requests.get(**request_kwargs) if self.verbose: d = datetime.datetime.now() s = "{}-{}-{} {}:{}:{}".format( d.year, d.month, d.day, d.hour, d.minute, d.second ) print("{0}GET REQUEST{1}".format("*" * 20, "*" * 24)) print("{0}{1}{2}".format("*" * 16, s, "*" * 20)) print(request_kwargs) print("{0}RESPONSE{1}".format("*" * 20, "*" * 27)) print(resp.content) print("*" * 48) return resp def post(self, data, url_suffix=None, headers=None, resource_id=None): request_kwargs = copy.deepcopy(self.request_kwargs) if resource_id: resource_url = request_kwargs["url"] request_kwargs["url"] = "{0}/{1}".format(resource_url, resource_id) if url_suffix: resource_url = request_kwargs["url"] request_kwargs["url"] = "{0}/{1}".format(resource_url, url_suffix) if headers: request_kwargs["headers"].update(headers) request_kwargs["data"] = data resp = requests.post(**request_kwargs) if self.verbose: d = datetime.datetime.now() s = "{}-{}-{} {}:{}:{}".format( d.year, d.month, d.day, d.hour, d.minute, d.second ) print("{0}POST REQUEST{1}".format("*" * 20, "*" * 24)) print("{0}{1}{2}".format("*" * 16, s, "*" * 21)) print(request_kwargs) print("{0}RESPONSE{1}".format("*" * 20, "*" * 27)) print(resp.content) print("*" * 48) return resp def delete(self, url_suffix=None, headers=None, resource_id=None): request_kwargs = copy.deepcopy(self.request_kwargs) if resource_id: resource_url = request_kwargs["url"] request_kwargs["url"] = "{0}/{1}".format(resource_url, resource_id) if url_suffix: resource_url = request_kwargs["url"] request_kwargs["url"] = "{0}/{1}".format(resource_url, url_suffix) if headers: request_kwargs["headers"].update(headers) resp = requests.delete(**request_kwargs) if self.verbose: d = datetime.datetime.now() s = "{}-{}-{} {}:{}:{}".format( d.year, d.month, d.day, d.hour, d.minute, d.second ) print("{0}DELETE REQUEST{1}".format("*" * 20, "*" * 24)) print("{0}{1}{2}".format("*" * 18, s, "*" * 21)) print(request_kwargs) print("{0}RESPONSE{1}".format("*" * 20, "*" * 27)) print(resp.content) print("*" * 48) return resp def get_identity_token(self): internal_idenity_url = self.cfg.INTERNAL_IDENTITY.internal_identity_url internal_identity_username = ( self.cfg.INTERNAL_IDENTITY.internal_identity_username ) internal_identity_password = ( self.cfg.INTERNAL_IDENTITY.internal_identity_password ) domain = self.cfg.INTERNAL_IDENTITY.domain domain_name = self.cfg.INTERNAL_IDENTITY.domain_name identity_headers = {"Content-Type": "application/json"} identity_data = { "auth": { "passwordCredentials": { "username": internal_identity_username, "password": internal_identity_password, }, domain: {"name": domain_name}, } } token_resp = requests.post( internal_idenity_url, headers=identity_headers, data=json.dumps(identity_data), ) token = token_resp.json()["access"]["token"]["id"] return token
5,754
1,753
"""*********************************************************** *** Copyright Tektronix, Inc. *** *** See www.tek.com/sample-license for licensing terms. *** ***********************************************************""" import socket import struct import math import time import sys echo_cmd = 0 """********************************************************************************* Function: instrument_connect(my_socket, ip_address string, my_port int, timeout do_reset, do_id_query) Purpose: Open an instance of an instrument object for remote communication over LAN/Ethernet. Parameters: my_socket - Instance of a socket object. ip_address (string) - The TCP/IP address string associated with the target instrument. my_port (int) - The instrument connection port. timeout (int) - The timeout limit for query/communication exchanges. do_reset (int) - Determines whether the instrument is to be reset upon connection to the instrument. Setting to 1 will perform the reset; setting to zero avoids it. do_clear (int) - Determines whether the instrument is to be cleared do_id_query (int) - Deterines when the instrument is to echho its identification string after it is initialized. Returns: my_socket - Updated instance of a socket object that includes attributes of a valid connection. Revisions: 2019-07-30 JJB Initial revision. *********************************************************************************""" def instrument_connect(my_socket, my_address, my_port, timeout, do_reset, do_clear, do_id_query): my_socket.connect((my_address, my_port)) # input to connect must be a tuple my_socket.settimeout(timeout) my_socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1) if do_reset == 1: instrument_write(my_socket, "*RST") if do_clear == 1: instrument_write(my_socket, "*CLS") if do_id_query == 1: tmp_id = instrument_query(my_socket, "*IDN?", 100) print(tmp_id) return my_socket """********************************************************************************* Function: instrument_disconnect(my_socket) Purpose: Break the LAN/Ethernet connection between the controlling computer and the target instrument. Parameters: my_socket - The TCP instrument connection object used for sending and receiving data. Returns: None Revisions: 2019-07-30 JJB Initial revision. *********************************************************************************""" def instrument_disconnect(my_socket): my_socket.close() return """********************************************************************************* Function: instrument_write(my_socket, my_command) Purpose: This function issues control commands to the target instrument. Parameters: my_socket - The TCP instrument connection object used for sending and receiving data. my_command (string) - The command issued to the instrument to make it perform some action or service. Returns: None Revisions: 2019-07-30 JJB Initial revision. *********************************************************************************""" def instrument_write(my_socket, my_command): if echo_cmd == 1: print(my_command) cmd = "{0}\n".format(my_command) my_socket.send(cmd.encode()) return """********************************************************************************* Function: instrument_read(my_socket, receive_size) Purpose: This function asks the connected instrument to reply with some previously requested information, typically queued up from a call to instrument_write(). Parameters: my_socket - The TCP instrument connection object used for sending and receiving data. receive_size (int) - Size of the data/string to be returned to the caller. Returns: reply_string (string) - The requested information returned from the target instrument. Revisions: 2019-07-30 JJB Initial revision. *********************************************************************************""" def instrument_read(my_socket, receive_size): return my_socket.recv(receive_size).decode() """********************************************************************************* Function: instrument_query(my_socket, my_command, receive_size) Purpose: This function issues control commands to the target instrument with the expectation that data will be returned. For this function instance, the returned data is (typically) in string format. Parameters: my_socket - The TCP instrument connection object used for sending and receiving data. my_command (string) - The command issued to the instrument to make it perform some action or service. receive_size (int) - The approximate number of bytes of data the caller expects to be returned in the response from the instrument. Returns: reply_string (string) - The requested information returned from the target instrument. Obtained by way of a caller to instrument_read(). Revisions: 2019-07-30 JJB Initial revision. *********************************************************************************""" def instrument_query(my_socket, my_command, receive_size): instrument_write(my_socket, my_command) return instrument_read(my_socket, receive_size) """********************************************************************************* Function: write_data(output_data_path, data_str) Purpose: This function issues control commands to the target instrument. Parameters: output_data_path (string) - The file name and path of the file to be written to. Note that the file is opened in append mode and no previously existing data will be over- written. data_str (string) - The data to be written to file. It is up to the user to format this data external to this function prior to using it. Returns: None Revisions: 2020-01-03 JJB Initial revision. *********************************************************************************""" def write_data(output_data_path, data_str): # This function writes the floating point data to the # target file. # for f in floats: ofile = open(output_data_path, "a") # append the target data dataStr = "{0}\n".format(data_str) ofile.write(dataStr) ofile.close() # Close the data file. return def upload_test_sequence(instrument_object, file_and_path): with open(file_and_path) as file_in: n = 1 # The first line in the sequence file is the header and not intended to be part of the test sequence for line in file_in: if n != 1: instrument_write(instrument_object, "append_test_to_global_table(\"{0}\")".format(line.rstrip('\r\n'))) n += 1 return """********************************************************************************* This example shows how the user of a Keithley DMM6500 can load a TSP script file and execute embedded functions. This allow the user to customize test operations at the instrument level. In particular, this example shows how a user might create a direct socket connection to the Series 2260B power supply and execute a supply output test sequence that defines voltage/current levels, durations for each defined step, and slew control. This program is dependendent on two additional files: A. The series_2260B_sequence_control.tsp script which.... 1. Promotes the transfer of the test sequence file to a Lua table on the DMM. 2. Initiates the sockets connection to the 2260B 3. Executes the uploaded test sequence. B. A 2260B test sequence in *.csv format. *********************************************************************************""" my_ip_address = "192.168.1.104" # Define your instrument's IP address here. my_port = 5025 # Define your instrument's port number here. do_instr_reset = 1 do_instr_clear = 1 do_instr_id_query = 1 t1 = time.time() # Open the socket connections... my_instr = socket.socket(socket.AF_INET, socket.SOCK_STREAM) # Establish a TCP/IP socket object instrument_connect(my_instr, my_ip_address, my_port, 20000, do_instr_reset, do_instr_clear, do_instr_id_query) # Ready the instrument to receive the target TSP file contents file = "series_2260B_sequence_control.tsp" func_file = open(file, "r") contents = func_file.read() func_file.close() instrument_write(my_instr, "if loadfuncs ~= nil then script.delete('loadfuncs') end") # Load the script file in one large chunk then close out the loadfuncs wrapper script. instrument_write(my_instr, "loadscript loadfuncs\n{0}\nendscript\n".format(contents)) # Call loadfuncs to load the contents of the script file into active memory print(instrument_query(my_instr, "loadfuncs()", 32)) # Note that we are echoing a queried function here. # You will note that the final line in the functions.tsp # script file is a print() command that will push its # contents to the output data queue. instrument_write(my_instr, "do_beep(0.250, 1000, 3)") file = "Test_Sequence_06.csv" upload_test_sequence(my_instr, file) ip_address_2260B = "192.168.1.117" instrument_write(my_instr, "connect_to_2260B(\"{0}\")".format(ip_address_2260B)) instrument_write(my_instr, "enable_2260B_output({0}, {1}, {2})".format(0.0, 1.0, "ON")) instrument_write(my_instr, "ps2260_execute_test_sequence()") instrument_write(my_instr, "enable_2260B_output({0}, {1}, {2})".format(0.0, 1.0, "OFF")) instrument_write(my_instr, "disconnect_from_2260B()") instrument_disconnect(my_instr) t2 = time.time() # Stop the timer... # Notify the user of completion and the data streaming rate achieved. print("done") print("Total Time Elapsed: {0:.3f} s".format(t2 - t1)) input("Press Enter to continue...") exit()
10,556
2,996
import cv2 from dev_autopilot import autopilot from emulation import get_bindings, clear_input import threading import kthread from pynput import keyboard from programInfo import showInfo STATE = 0 def start_action(): stop_action() kthread.KThread(target = autopilot, name = "EDAutopilot").start() def stop_action(): cv2.destroyAllWindows() for thread in threading.enumerate(): if thread.getName() == 'EDAutopilot': thread.kill() clear_input(get_bindings()) def on_press(key): try: if key == keyboard.Key.home: print('start action') start_action() if key == keyboard.Key.end: print('stop action') stop_action() except AttributeError: print('special key {0} pressed'.format(key)) def on_release(key): if key == keyboard.Key.esc: # Stop listener cv2.destroyAllWindows() stop_action() return False showInfo(); # Collect events until released with keyboard.Listener(on_press=on_press,on_release=on_release) as listener: listener.join()
1,143
359
""" Showcases correlated colour temperature computations. """ import colour from colour.utilities import message_box message_box("Correlated Colour Temperature Computations") cmfs = colour.MSDS_CMFS["CIE 1931 2 Degree Standard Observer"] illuminant = colour.SDS_ILLUMINANTS["D65"] xy = colour.XYZ_to_xy(colour.sd_to_XYZ(illuminant, cmfs) / 100) uv = colour.UCS_to_uv(colour.XYZ_to_UCS(colour.xy_to_XYZ(xy))) message_box( f'Converting to "CCT" and "D_uv" from given "CIE UCS" colourspace "uv" ' f'chromaticity coordinates using "Ohno (2013)" method:\n\n\t{uv}' ) print(colour.uv_to_CCT(uv, cmfs=cmfs)) print(colour.temperature.uv_to_CCT_Ohno2013(uv, cmfs=cmfs)) print("\n") message_box("Faster computation with 3 iterations but a lot less precise.") print(colour.uv_to_CCT(uv, cmfs=cmfs, iterations=3)) print(colour.temperature.uv_to_CCT_Ohno2013(uv, cmfs=cmfs, iterations=3)) print("\n") message_box( f'Converting to "CCT" and "D_uv" from given "CIE UCS" colourspace "uv" ' f'chromaticity coordinates using "Robertson (1968)" method:\n\n\t{uv}' ) print(colour.uv_to_CCT(uv, method="Robertson 1968")) print(colour.temperature.uv_to_CCT_Robertson1968(uv)) print("\n") CCT_D_uv = [6503.49254150, 0.00320598] message_box( f'Converting to "CIE UCS" colourspace "uv" chromaticity coordinates from ' f'given "CCT" and "D_uv" using "Ohno (2013)" method:\n\n\t{CCT_D_uv}' ) print(colour.CCT_to_uv(CCT_D_uv, cmfs=cmfs)) print(colour.temperature.CCT_to_uv_Ohno2013(CCT_D_uv, cmfs=cmfs)) print("\n") message_box( f'Converting to "CIE UCS" colourspace "uv" chromaticity coordinates from ' f'given "CCT" and "D_uv" using "Robertson (1968)" method:\n\n\t{CCT_D_uv}' ) print(colour.CCT_to_uv(CCT_D_uv, method="Robertson 1968")) print(colour.temperature.CCT_to_uv_Robertson1968(CCT_D_uv)) print("\n") CCT = 6503.49254150 message_box( f'Converting to "CIE UCS" colourspace "uv" chromaticity coordinates from ' f'given "CCT" using "Krystek (1985)" method:\n\n\t({CCT})' ) print(colour.CCT_to_uv(CCT, method="Krystek 1985")) print(colour.temperature.CCT_to_uv_Krystek1985(CCT)) print("\n") xy = colour.CCS_ILLUMINANTS["CIE 1931 2 Degree Standard Observer"]["D65"] message_box( f'Converting to "CCT" from given "CIE xy" chromaticity coordinates using ' f'"McCamy (1992)" method:\n\n\t{xy}' ) print(colour.xy_to_CCT(xy, method="McCamy 1992")) print(colour.temperature.xy_to_CCT_McCamy1992(xy)) print("\n") message_box( f'Converting to "CCT" from given "CIE xy" chromaticity coordinates using ' f'"Hernandez-Andres, Lee and Romero (1999)" method:\n\n\t{xy}' ) print(colour.xy_to_CCT(xy, method="Hernandez 1999")) print(colour.temperature.xy_to_CCT_Hernandez1999(xy)) print("\n") CCT = 6503.49254150 message_box( f'Converting to "CIE xy" chromaticity coordinates from given "CCT" using ' f'"Kang, Moon, Hong, Lee, Cho and Kim (2002)" method:\n\n\t{CCT}' ) print(colour.CCT_to_xy(CCT, method="Kang 2002")) print(colour.temperature.CCT_to_xy_Kang2002(CCT)) print("\n") message_box( f'Converting to "CIE xy" chromaticity coordinates from given "CCT" using ' f'"CIE Illuminant D Series" method:\n\n\t{CCT}' ) print(colour.CCT_to_xy(CCT, method="CIE Illuminant D Series")) print(colour.temperature.CCT_to_xy_CIE_D(CCT))
3,284
1,495
#!/usr/bin/env python import drmaa import shlex from optparse import OptionParser from sys import stderr, stdin, exit from datetime import datetime import traceback def Stop(pool, jt, info): '''Job failure function that stops synchronization.''' pool.shall_stop = True pool.all_done = False pool.failed_jobs.append(jt) def Proceed(pool, jt, info): '''Job failure function that proceeds with the remaining jobs.''' pool.all_done = False pool.failed_jobs.append(jt) def Resubmit(max_tries, fail): '''Job failure function factory that resubmits a failed job. :Parameters: max_tries: maximum number of submissions for a job. fail: failure function to call if the maximum number of tries has been reached. ''' def resubmit_function(pool, jt, info): if jt.failures >= max_tries: fail(pool, jt, info) else: jt.jobid = pool.session.runJob(jt) pool.log('job specified at ' + jt.source + ' resubmitted with id ' + jt.jobid) pool.current_jobs[jt.jobid] = jt return resubmit_function class JobPool: ''' A pool of jobs. :Members: session: DRMAA session logfile: file where actions and status are written current_jobs: jobs that have been submitted and that are not finished all_done: either all finished jobs were successful shall_stop: either this object should stop the synchronization ''' def __init__(self, session, logfile): self.session = session self.logfile = logfile self.current_jobs = {} self.all_done = True self.shall_stop = False self.failed_jobs = [] def log(self, msg=''): '''Logs a message''' d = datetime.now() self.logfile.write('[' + d.strftime('%Y-%m-%d %H:%M:%S') + '] ' + msg + '\n') self.logfile.flush() def createJobTemplate(self): '''Creates a job template (delegates to self.session)''' return self.session.createJobTemplate() def runJob(self, jt): '''Submits a job. This method delegates to self.session, then keeps track of the submitted job :Parameters: jt: job template, with a member 'source' indicating where this template was specified ''' jt.jobid = self.session.runJob(jt) if jt.source is None: jt.source = jobid jt.failures = 0 self.log('job specified at ' + jt.source + ' submitted with id ' + jt.jobid) self.current_jobs[jt.jobid] = jt return jt.jobid def waitall(self, fail=Proceed, interval=60): '''Waits for all submitted jobs to finish. :Parameters: fail: function called in case of failure, the function must accept 3 paremeters: this object, the JobTemplate object and the DRMAA JobInfo object. interval: check for job status every number of seconds. ''' start = datetime.now() running = 0 while self.current_jobs: joblist = list(self.current_jobs.keys()) # create fresh list to work around Python 3 iterator try: self.log('synchronizing %d jobs (%d running), see you in %d seconds' % (len(joblist), running, interval)) self.session.synchronize(joblist, interval, False) except drmaa.errors.ExitTimeoutException: pass running = 0 for jobid in joblist: status = self.session.jobStatus(jobid) if status == drmaa.JobState.DONE: try: info = self.session.wait(jobid, drmaa.Session.TIMEOUT_NO_WAIT) jt = self.current_jobs[jobid] except drmaa.errors.ExitTimeoutException: pass if info.wasAborted: self.log('job specified at %s with id %s aborted' % (self.current_jobs[jobid].source, jobid)) self._failed(jobid, fail, info) elif info.hasSignal: self.log('job specified at %s with id %s aborted received signal %d' % (self.current_jobs[jobid].source, jobid, info.terminatedSignal)) self._failed(jobid, fail, info) elif info.exitStatus != 0: self.log('job specified at %s with id %s aborted exited with status %d' % (self.current_jobs[jobid].source, jobid, info.exitStatus)) self._failed(jobid, fail, info) else: self.log('job specified at %s with id %s is done' % (self.current_jobs[jobid].source, jobid)) del self.current_jobs[jobid] elif status == drmaa.JobState.FAILED: self.log('job specified at %s with id %s failed somehow' % (self.current_jobs[jobid].source, jobid)) self._failed(jobid, fail, None) elif status == drmaa.JobState.RUNNING: running += 1 if self.shall_stop: break if self.all_done: delta = datetime.now() - start self.log('all jobs completed successfully in ' + str(delta) + ', you\'re welcome') else: self.log('sorry, the following jobs have failed:') for job in self.failed_jobs: self.log(job.source + ' with id ' + str(job.jobid)) def _failed(self, jobid, fail, info): jt = self.current_jobs[jobid] jt.failures += 1 del self.current_jobs[jobid] fail(self, jt, info) def runall(self, jobs, fail=Proceed, interval=60): '''Submits jobs and waits for them to finish. :Parameters: jobs: a sequence of job templates fail: job failure function interval: job status check interval in seconds :Return value: True if all jobs finished successfully, False otherwise. ''' for jt in jobs: self.runJob(jt) self.waitall(fail, interval) return self.all_done def terminate(self): '''Terminates all remaining jobs.''' self.log('terminating remaining jobs') self.session.control(drmaa.Session.JOB_IDS_SESSION_ALL, drmaa.JobControlAction.TERMINATE) self.current_jobs = {} class QSyncBase: def __init__(self): pass def create_jobs(self, session): raise NotImplemented() def go(self, interval=60, force_interval=False, fail=Proceed, logfile=stderr): if interval < 1: raise Exception('illegal interval: %d' % interval) if interval <= 10 and not force_interval: raise Exception('unwise interval: %d (use force interval if you want this anyway') session = drmaa.Session() session.initialize() jobs = self.create_jobs(session) pool = JobPool(session, logfile) try: r = pool.runall(jobs, fail, interval) if not r: pool.terminate() return r except BaseException as e: pool.log('wow, some exception here...') traceback.print_exc() pool.terminate() finally: session.exit() class QSync(OptionParser, QSyncBase): def __init__(self): OptionParser.__init__(self, usage='Usage: %prog [OPTIONS] [FILE...]') self.set_defaults(fail=Proceed) self.add_option('-s', '--stop-on-failure', action='store_const', const=Stop, dest='fail', help='if one job fails, stop synchronization and terminate all remaining jobs') self.add_option('-p', '--proceed-on-failure', action='store_const', const=Proceed, dest='fail', help='continue running jobs even if some fail (default behaviour)') self.add_option('-r', '--resubmit-on-failure', action='store', type='int', dest='resubmit', help='resubmit failed jobs at most N times each', metavar='N') self.add_option('-l', '--log-file', action='store', type='string', dest='logfile', default=None, help='write log into FILE (default: stderr)', metavar='FILE') self.add_option('-i', '--interval', action='store', type='int', dest='interval', default=60, help='wait T seconds before polling job status, values below 10 require --force-interval (default: %default)', metavar='T') self.add_option('--force-interval', action='store_true', dest='force_interval', default=False, help='accept poll intervals below 10 seconds') def run(self): options, self.filenames = self.parse_args() fail = options.fail if options.resubmit: if options.resubmit < 1: raise Exception('illegal number of resubmissions: %d' % options.resubmit) fail = Resubmit(options.resubmit, fail) logfile = stderr if options.logfile: logfile = open(options.logfile, 'w') self.go(interval=options.interval, force_interval=options.force_interval, fail=fail, logfile=logfile) @staticmethod def _create_job(session, filename, f): for n, line in enumerate(f): jt = session.createJobTemplate() b, dd, a = line.partition('--') if dd != '': jt.nativeSpecification = b line = a args = shlex.split(line) jt.remoteCommand = args[0] jt.args = args[1:] jt.source = '%s:%d' % (filename, n + 1) yield jt def create_jobs(self, session): if self.filenames: for filename in self.filenames: f = open(filename) for p in QSync._create_job(session, filename, f): yield p f.close() else: for p in QSync._create_job(session, '<stdin>', stdin): yield p if __name__ == '__main__': if not QSync().run(): exit(1)
9,940
2,886
import os import setuptools with open("README.md", "r") as fh: long_description = fh.read() here = os.path.dirname(os.path.abspath(__file__)) version_ns = {} with open(os.path.join(here, "jupyterlab_heroku", "_version.py")) as f: exec(f.read(), {}, version_ns) setuptools.setup( name="jupyterlab_heroku", version=version_ns["__version__"], author="Jeremy Tuloup", description="A server extension for the JupyterLab Heroku extension", long_description=long_description, long_description_content_type="text/markdown", packages=setuptools.find_packages(), install_requires=["jupyterlab"], package_data={"jupyterlab_heroku": ["*"]}, )
678
236
from rest_framework import serializers from .models import Placement class PlacementSerializer(serializers.ModelSerializer): class Meta: model = Placement fields = ("key", "placement_name", "company", "role", "description", "deadline")
258
68
# set random number generator np.random.seed(2020) # initialize step_end and v step_end = int(t_max / dt) v = el t = 0 with plt.xkcd(): # initialize the figure plt.figure() plt.title('$V_m$ with random I(t)') plt.xlabel('time (s)') plt.ylabel(r'$V_m$ (V)') # loop for step_end steps for step in range(step_end): t = step * dt plt.plot(t, v, 'k.') i = i_mean * (1 + 0.1 * (t_max / dt)**(0.5) * (2 * np.random.random() - 1)) v = v + (dt / tau) * (el - v + r * i) plt.show()
509
234
import sys from collections import defaultdict import torch from varclr.utils.infer import MockArgs from varclr.data.preprocessor import CodePreprocessor if __name__ == "__main__": ret = torch.load(sys.argv[2]) vars, embs = ret["vars"], ret["embs"] embs /= embs.norm(dim=1, keepdim=True) embs = embs.cuda() var2idx = dict([(var, idx) for idx, var in enumerate(vars)]) processor = CodePreprocessor(MockArgs()) Ks = [1, 5, 10, 25, 50, 100, 250, 500, 1000] topk_succ = defaultdict(int) tot = 0 with open(sys.argv[1], "r") as f: for line in f: try: var1, var2 = line.strip().split() except ValueError: print("skpped: ", line) def canon(var): return "".join( [ word.capitalize() if idx > 0 else word for idx, word in enumerate(processor(var).split()) ] ) var1, var2 = canon(var1), canon(var2) if var1 not in var2idx or var2 not in var2idx: print(f"variable {var1} or {var2} not found") continue tot += 1 for k in Ks: result = torch.topk(embs @ embs[var2idx[var1]], k=k + 1) topk_succ[k] += var2 in [vars[idx] for idx in result.indices][1:] print(f"Total {tot} variable pairs") for k in Ks: print(f"Recall@{k} = {100 * topk_succ[k] / tot:.1f}")
1,509
517
import os import numpy as np from . import imtools, datprops from .datfile import DatFile from .chiptype import ChipType moduleDir = os.path.abspath( os.path.dirname( __file__ ) ) class FlowCorr: def __init__( self, chiptype, xblock=None, yblock=None, rootdir='.', method='' ): ''' Initialize a flowcorr object chiptype: a ChipType object xblock: The full-chip column origin; setting to None returns a full chip yblock: The full-chip row origin; setting to None returns a full chip rootdir: root directory to look for flowcorr files. search will also look up a level, within the module directory, and in the dats directory method: if specified, automaticaly loads the corresponding flowcorr 'buffer' 'file' if advanced options need to be passed into the load functions, they should be called separatly with method being left empty ''' self.chiptype = ChipType(chiptype) self.xblock = xblock self.yblock = yblock self.searchpath = [ rootdir, os.path.join( rootdir, '..' ), os.path.join( moduleDir, '../dats' ), moduleDir, os.path.join( moduleDir, 'dats' ) ] if method.lower() == 'buffer': self.frombuffer() elif method.lower() == 'file': self.fromfile() elif not method: pass else: raise ValueError( 'Flowcorr method "%s" is undefined' % method ) def frombuffer(self, flow_file='C2_step.dat', force=False, framerate=15): ''' Returns the flow correction measured from a buffered flow flowfile: measurement file used to calculate the flowcorr force: calculate the data from raw, even if an existing analysis is present framerate: fps ''' try: if force: raise IOError self.filename = os.path.join( self.searchpath[0], 'flowcorr_slopes.dat' ) self.flowcorr = datprops.read_dat( self.filename, 'flowcorr', chiptype=self.chiptype ) except IOError: # Read the dat file found = False for dirname in self.searchpath: self.filename = os.path.join( dirname, flow_file ) if os.path.exists( self.filename ): found = True break if not found: raise IOError( '%s was not found' % self.filename ) data = DatFile( self.filename, chiptype=self.chiptype ) # Calculate properties self.flowcorr = data.measure_slope( method='maxslope' ) self.time_offset = np.min(data.measure_t0( method='maxslope' )) #TODO: This is not very robust. should just shift t0 here and record the offest instead of trying to do things later with it self.pinned = data.measure_pinned() # remove pins self.flowcorr[ self.pinned ] = 1 # Save a few more variables self.t0 = data.measure_t0( meathod='maxslope' ) self.actpix = data.measure_actpix self.phpoint = data.measure_plateau() return self.flowcorr def fromfile( self, fc_type ): ''' Loads the flow correction from file based on the chip type and scales up from miniblocks to full chips or analysis blocks. This method only differentiates based on thumbnail or full chip/analysis block. All other differences are rolled into ChipType. fc_type: can be 'ecc' or 'wt'. flowcorr file is defined by self.chiptype.flowcorr_<fc_type> ''' # Thumbnails are enough different to have their own function if self.chiptype.tn == 'self': return self.tn_fromfile( fc_type ) # Spatial thumbnails are just subsampled data. We don't need special loading # Calculate the size of the flowcorr files xMiniBlocks = self.chiptype.chipC / self.chiptype.miniC yMiniBlocks = self.chiptype.chipR / self.chiptype.miniR # Set the flowcorr path starting local before using the default for path in self.searchpath: filename = os.path.join( path, '%s.dat' % getattr( self.chiptype, 'flowcorr_%s' % fc_type ) ) try: flowcorr = datprops.read_dat( filename , metric='flowcorr' ) break except IOError: continue raise IOError( 'Could not find a flowcorr file' ) # Scale the flowcorr data to the entire well sizes = [ ( 96, 168 ), # This is an unscaled P1-sized flowcorr file. This is the most likely size when reading fc_flowcorr.dat ( yMiniBlocks, xMiniBlocks ), # This is the historical per-chip file. This is ( 96, 168 ) for a P1/540 chip ( self.chiptype.chipR, self.chiptype.chipC ) ] # This is the pre-compiled value try: fc_xMiniBlocks = self.chiptype.fullchip.chipC / self.chiptype.fullchip.miniC fc_yMiniBlocks = self.chiptype.fullchip.chipR / self.chiptype.fullchip.miniR sizes.append( ( fc_yMiniBlocks, fc_xMiniBlocks ) ) sizes.append( ( self.chiptype.fullchip.chipR, self.chiptype.fullchip.chipC ) ) except AttributeError: pass for size in sizes: try: flowcorr = flowcorr.reshape( size ) break except ValueError: # Keep going until you itterate through all possible sizes. If you still get an error, then die if size == sizes[-1]: print 'Possible Sizes' print sizes print 'Elements' print flowcorr.shape raise ValueError( 'Could not determine flowcorr size' ) continue # Resize the image to the current size if self.chiptype.burger is None: # This is a standard resize operation flowcorr = imtools.imresize( flowcorr, ( self.chiptype.chipR, self.chiptype.chipC ) ) elif self.chiptype.spatn != 'self': # This is burger mode on a full size chip flowcorr = imtools.imresize( flowcorr, ( self.chiptype.burger.chipR, self.chiptype.burger.chipC ) ) # Clip off the top and bottom first = ( flowcorr.shape[0] - self.chiptype.chipR ) / 2 last = first + self.chiptype.chipR flowcorr = flowcorr[ first:last, : ] else: # This is burger mode on a spatial thumbnail # This has the effect of adding more rows beyond the 800 typically used for a spatial thumbnail rows = self.chiptype.chipR * self.chiptype.burger.chipR / self.chiptype.fullchip.chipR flowcorr = imtools.imresize( flowcorr, ( rows, self.chiptype.chipC ) ) # Clip off the top and bottom first = ( flowcorr.shape[0] - self.chiptype.chipR ) / 2 last = first + self.chiptype.chipR flowcorr = flowcorr[ first:last, : ] # Reduce to a single analysis block if ( self.xblock is not None and self.yblock is not None and self.xblock != -1 and self.yblock != -1 ): flowcorr = flowcorr[ self.yblock: self.chiptype.blockR + self.yblock, self.xblock: self.chiptype.blockC + self.xblock ] self.flowcorr = flowcorr return flowcorr def tn_fromfile( self, fc_type ): ''' Gets the per-well flowcorrection for a STANDARD (not spatial) thumbnail ''' # Calculate the size of the flowcorr files xMiniBlocks = self.chiptype.chipC / self.chiptype.miniC yMiniBlocks = self.chiptype.chipR / self.chiptype.miniR # Set the flowcorr path starting local before using the default for path in self.searchpath: filename = os.path.join( path, '%s.dat' % getattr( self.chiptype, 'flowcorr_%s' % fc_type ) ) try: flowcorr = datprops.read_dat( filename , metric='flowcorr' ) break except IOError: continue raise IOError( 'Could not find a flowcorr file' ) # Scale the flowcorr data to the entire well sizes = ( ( 96, 168 ), # This is an unscaled P1-sized flowcorr file. ( 48, 96 ) , # This is an unscaled P0-sized flowcorr file. ( yMiniBlocks, xMiniBlocks ), # This is the historical thumbnail flowcorr (swapped x & y - STP 7/13/2015) ( self.chiptype.fullchip.chipR, self.chiptype.fullchip.chipC ) ) # This is the pre-compiled value for size in sizes: try: flowcorr = flowcorr.reshape( size ) break except ValueError: # Keep going until you itterate through all possible sizes. If you still get an error, then die if size == sizes[-1]: raise ValueError( 'Could not determine flowcorr size' ) continue # Resize the image to the full chip size if self.chiptype.burger is None: # This is a standard resize operation based on the full chip flowcorr = imtools.imresize( flowcorr, ( self.chiptype.fullchip.chipR, self.chiptype.fullchip.chipC ) ) else: # This is burger mode on a regular thumbnail. Full chip is actually specified by burger and then we have to clip flowcorr = imtools.imresize( flowcorr, ( self.chiptype.burger.chipR, self.chiptype.burger.chipC ) ) # Clip off the top and bottom first = ( flowcorr.shape[0] - self.chiptype.fullchip.chipR ) / 2 last = first + self.chiptype.fullchip.chipR flowcorr = flowcorr[ first:last, : ] # Reduce to thumbnail data tnflowcorr = np.zeros( ( self.chiptype.chipR, self.chiptype.chipC ) ) for r in range( self.chiptype.yBlocks ): tn_rstart = r*self.chiptype.blockR tn_rend = tn_rstart + self.chiptype.blockR #fc_rstart = int( (r+0.5)*self.chiptype.fullchip.blockR ) - self.chiptype.blockR/2 # middle of block in case the thumbnail different yBlocks center within the block fc_rstart = int( (r+0.5)*(self.chiptype.fullchip.chipR/self.chiptype.yBlocks) ) - self.chiptype.blockR/2 fc_rend = fc_rstart + self.chiptype.blockR for c in range( self.chiptype.xBlocks ): tn_cstart = c*self.chiptype.blockC tn_cend = tn_cstart + self.chiptype.blockC fc_cstart = int( (c+0.5)*self.chiptype.fullchip.blockC ) - self.chiptype.blockC/2 fc_cend = fc_cstart + self.chiptype.blockC tnflowcorr[ tn_rstart:tn_rend, tn_cstart:tn_cend ] = flowcorr[ fc_rstart:fc_rend, fc_cstart:fc_cend ] self.flowcorr = tnflowcorr return self.flowcorr
11,285
3,361
#!/usr/bin/env python import numpy as np def ini(): '''initialization Args: Returns: ''' pass def accuracy(Yte_predict, Yte): '''verification of the correct label with the predicted one Args: Yte_predict (list): predicted labels (by computer) Yte (list): true labels (by human) Returns: accuracy (list): congruity ''' return np.mean(Yte_predict == Yte) def ratio(Yte_predict): '''Percentage portion of array Args: Yte_predict (list): predicted labels (by computer) Returns: accuracy (list): relative part of the labels "1" in the array ''' return np.mean(Yte_predict == 1)
688
219
# AUTOGENERATED! DO NOT EDIT! File to edit: nbs/models/models.dnn.ipynb (unless otherwise specified). __all__ = ['Multi_Layer_Perceptron', 'CollabFNet'] # Cell import torch import torch.nn as nn import torch.nn.functional as F # Cell class Multi_Layer_Perceptron(nn.Module): def __init__(self, args, num_users, num_items): super(Multi_Layer_Perceptron, self).__init__() self.num_users = num_users self.num_items = num_items self.factor_num = args.factor_num self.layers = args.layers self.embedding_user = nn.Embedding(num_embeddings=self.num_users, embedding_dim=self.factor_num) self.embedding_item = nn.Embedding(num_embeddings=self.num_items, embedding_dim=self.factor_num) self.fc_layers = nn.ModuleList() for idx, (in_size, out_size) in enumerate(zip(self.layers[:-1], self.layers[1:])): self.fc_layers.append(nn.Linear(in_size, out_size)) self.affine_output = nn.Linear(in_features=self.layers[-1], out_features=1) self.logistic = nn.Sigmoid() def forward(self, user_indices, item_indices): user_embedding = self.embedding_user(user_indices) item_embedding = self.embedding_item(item_indices) vector = torch.cat([user_embedding, item_embedding], dim=-1) # the concat latent vector for idx, _ in enumerate(range(len(self.fc_layers))): vector = self.fc_layers[idx](vector) vector = nn.ReLU()(vector) # vector = nn.BatchNorm1d()(vector) # vector = nn.Dropout(p=0.5)(vector) logits = self.affine_output(vector) rating = self.logistic(logits) return rating def init_weight(self): pass # Cell class CollabFNet(nn.Module): def __init__(self, num_users, num_items, emb_size=100, n_hidden=10): super(CollabFNet, self).__init__() self.user_emb = nn.Embedding(num_users, emb_size) self.item_emb = nn.Embedding(num_items, emb_size) self.lin1 = nn.Linear(emb_size*2, n_hidden) self.lin2 = nn.Linear(n_hidden, 1) self.drop1 = nn.Dropout(0.1) def forward(self, u, v): U = self.user_emb(u) V = self.item_emb(v) x = F.relu(torch.cat([U, V], dim=1)) x = self.drop1(x) x = F.relu(self.lin1(x)) x = self.lin2(x) return x
2,354
855
import unittest import pandas as pd import numpy as np import src.features.outlier_correction as oc class TestFlagNormalOutliers(unittest.TestCase): def setUp(self): self.series = pd.Series(np.random.normal(0, 0.1, 1000)) self.series.loc[500] = -5 self.series.loc[42] = 5 def test_flag(self): outliers = oc.flag_normal_outliers(self.series, 5) self.assertTrue(outliers.loc[500]) self.assertTrue(outliers.loc[42])
477
175
import csv import operator if __name__ == "__main__": # We use try-except statements to properly handle errors, as in case the # input file does not exist in the directory. try: # We open the .csv file loading the filds separated by a ; delimiter: csv_archivo_locales = open("Locales.csv", encoding="utf8", errors='ignore') locales = csv.reader(csv_archivo_locales, delimiter=";") csv_archivo_terrazas = open("Terrazas.csv", encoding="utf8", errors='ignore') terrazas = csv.reader(csv_archivo_terrazas, delimiter=";") # We skip the first line before saving the results into a list for later # processing by using the next() statement. The reason why we do not use # the line_num function is because we would need to include it in a loop # and read file line by line whereas this way, with just two instructions # we get the same result: next(locales, None) lista_locales = list(locales) next(terrazas, None) lista_terrazas = list(terrazas) # When we read the fields from the CSV, they are stored as strings, so we # need to explicitely convert the ID to int to be able to sort them: for i in lista_locales: i[0] = int(i[0]) for j in lista_terrazas: j[0] = int(j[0]) # We sort the lists taking into account the ID column which is in the # first position to get it with the itemergetter function: sorted_lista_locales = sorted(lista_locales, key=operator.itemgetter(0), reverse = False) sorted_lista_terrazas = sorted(lista_terrazas, key=operator.itemgetter(0), reverse = False) # For each entry in lista_terrazas, we check where is its corresponding # entry in lista_locales. As they are sorted ascendently, all the entries # before the wanted one are included in the result list as they do not # appear in lista_terrazas. Moreover, we keep the index counter to just # traverse the lista_locales once. index = 0 no_terrazas = [] for terraza in sorted_lista_terrazas: while (terraza[0] > sorted_lista_locales[index][0]): no_terrazas.append(sorted_lista_locales[index]) index += 1 # It is important to perform this step once we have reached the local # entry from the lista_locales, as if not, the next entry from # lista_terrazas will be greater than the found one in lista_locales # and will wrongly include it in the result list: if (terraza[0] == sorted_lista_locales[index][0]): index += 1 # We open the output file to store the data retrieved: csvFileObj = open("NoTerrazas.csv", "w") csvWriter = csv.writer(csvFileObj, delimiter=";") # For each row in the final list, we write it to the CSV output file: for row in no_terrazas: csvWriter.writerow(row) print ("The file has been successfully created.") # We close the opened CSV file: csvFileObj.close() # NOTE: we have realized that a few error occurs when comparing the total # amount of entries in NoTerrazas.csv with the entries in Locales.csv and # Terrazas.csv, as it is expected to have NoTerrazas = Locales - Terrazas # However, as the exercice statement indicates that some inconsistency # may exist in the files, we assume the error is due to this unsteadiness. except IOError: print("The file does not exist.")
3,613
994
""" Copyright (c) College of Mechatronics and Control Engineering, Shenzhen University. All rights reserved. Description : Author:Team Li """ import glob, os, sys, time from threading import Thread import cv2 from carla_utils.world_ops import * from carla_utils.sensor_ops import * try: sys.path.append(glob.glob('**/carla-*%d.%d-%s.egg' % ( sys.version_info.major, sys.version_info.minor, 'win-amd64' if os.name == 'nt' else 'linux-x86_64'))[0]) import carla except: raise ImportError('Please check your carla file') def check_whether_respawn_actors(world, vehicles): """check whether to respawn the static acotors in a frequency""" while True: if carla_actors_static(vehicles, bigger_than=0.8): respawn_static_actors(world, vehicles) time.sleep(20) if __name__ == '__main__': client = carla.Client('127.0.0.1', 2000) client.set_timeout(10.0) # seconds logger.info('connect success...') world = client.get_world() ## spawn vehicles in carla world spawn_vehicles(world, number=5) actor_list = world.get_actors() vehicles = list(actor_list.filter('vehicle*')) bgr_camera_config = {'data_type': 'sensor.camera.rgb', 'image_size_x': 418, 'image_size_y': 278, 'fov': 110, 'sensor_tick': 0.05, 'transform': carla.Transform(carla.Location(x=0.8, z=1.7)), 'attach_to':vehicles[0]} bgr_sensor = bgr_camera(world, bgr_camera_config) depth_camera_config = {'data_type': 'sensor.camera.depth', 'image_size_x': 418, 'image_size_y': 278, 'fov': 110, 'sensor_tick': 0.02, 'transform': carla.Transform(carla.Location(x=0.8, z=1.7)), 'attach_to':vehicles[0]} depth_sensor = depth_camera(world, depth_camera_config) semantic_camera_config = {'data_type': 'sensor.camera.semantic_segmentation', 'image_size_x': 418, 'image_size_y': 278, 'fov': 110, 'sensor_tick': 0.02, 'transform': carla.Transform(carla.Location(x=0.8, z=1.7)), 'attach_to':vehicles[0]} semantic_sensor = semantic_camera(world, semantic_camera_config) ## start check t = Thread(target=check_whether_respawn_actors, args=(world, vehicles)) t.daemon = True t.start() logger.info('Press key a to stop...') while True: bgr = bgr_sensor.get() depth = depth_sensor.get() semantic = semantic_sensor.get() cv2.imshow('Color', bgr) cv2.imshow('Depth', depth) cv2.imshow('Semantic', semantic) a = cv2.waitKey(10) if a == 97: cv2.destroyAllWindows() break vehicles = list(actor_list.filter('vehicle*')) for vehicle in vehicles: vehicle.destroy() logger.info('Destroy all vehicles...') sensors = list(actor_list.filter('sensor*')) for sensor in sensors: sensor.destroy() logger.info('Destroy all sensors...')
3,076
1,080
#!/usr/bin/env python import os if not os.path.isfile("/proc/sys/fs/binfmt_misc/WSLInterop"): # windows LOGFILE = os.environ['LOCALAPPDATA'] + "\\AGS\\New World\\Game.log" else: # wsl LOGFILE = os.popen('cmd.exe /c "echo %LocalAppData%"').read().strip() + "\\AGS\\New World\\Game.log" LOGFILE = os.popen("wslpath '{}'".format(LOGFILE)).read().strip() # INSTRUCTIONS # 1. Uncomment the names of the plugins you wish to use # 2. Configure the relevant variables below # Need help? Check out the wiki for a sample config file - https://github.com/Rawr0/nwwatch/wiki ############################### ## PLUGIN SETTINGS ## ############################### # It is recommended you leave these two as default, no testing has been performed with different values NW_FILE_CHECK_FREQUENCY = 60 # How frequently the script checks your queue position. Default: 60 NW_ALERT_AT_QUEUE_POSITION = 25 # Send a notification when you are at this position in queue (or less). It is recommended this be 25 or greater. PLUGINS_ENABLED = { # Remove the hash (#) before a line to enable it #"NotifyByPushover", #"NotifyBySMS", #"NotifyByDiscord" } # Want to test your notifications? Enable it above and then set "TEST_MODE" to True. A notification will be triggered as soon as the script starts TEST_MODE = True ############################### ## PLUGIN SPECIFIC VARIABLES ## ############################### # Plugin: NotifyByPushover PUSHOVER_TOKEN = "<VALUEHERE>" PUSHOVER_USER = "<VALUEHERE>" PUSHOVER_DEVICE = "<VALUEHERE>" PUSHOVER_HIGHPRIORITY = True # Plugin: NotifyBySMS (sinch.com) (Note: Paid service) # Note: The SMS provider is currently going through a rebranding and, as a result, the APIs below could stop working. # If any issues are encountered, please raise an issue on Github SMS_PLAN_ID = "" SMS_TOKEN = "" SMS_SOURCE = "New World" # Source phone number, including country code (eg. +15551231234) or alphanumeric string SMS_TARGET = "" # Destination phone number, including country code (eg. +15555551234 or +61411000000) # Plugin: NotifyByDiscord # In a Discord server you own/manage, navigate to Server Settings -> Integrations -> Webooks -> New Webhook. Click "Copy Webhook URL" and paste it below DISCORD_WEBHOOKURL = "https://discord.com/api/webhooks/xxxxxxxxxxxxxx/xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" DISCORD_TTS = False # Use Discord "Text to speech" to speak the announcement ############################### ## INTERNAL VARIABLES ## ############################### # Don't change these unless you're having issues NW_LOGFILE_CHECK_LENGTH = 100 # Number of lines to monitor in the logfile NW_SEARCH_REGEX = ".*Waiting in login queue.*Position \((.*)\)$" NW_SEARCH_REGEX_INDEX = 1
2,792
938
#author SANKALP SAXENA def arrays(arr): arr.reverse() l = numpy.array(arr, float) return l
109
46
# -*- coding: utf-8 -*- import os import time import pytest from pip_services3_commons.config import ConfigParams from pip_services3_kafka.connect.KafkaConnection import KafkaConnection broker_host = os.environ.get('KAFKA_SERVICE_HOST') or 'localhost' broker_port = os.environ.get('KAFKA_SERVICE_PORT') or 9092 broker_topic = os.environ.get('KAFKA_TOPIC') or 'test' broker_user = os.environ.get('KAFKA_USER') # or 'kafka' broker_pass = os.environ.get('KAFKA_PASS') # or 'pass123' @pytest.mark.skipif(not broker_host and not broker_port, reason="Kafka server is not configured") class TestKafkaConnection: connection: KafkaConnection def setup_method(self): config = ConfigParams.from_tuples( 'topic', broker_topic, 'connection.protocol', 'tcp', 'connection.host', broker_host, 'connection.port', broker_port, 'credential.username', broker_user, 'credential.password', broker_pass, 'credential.mechanism', 'plain' ) self.connection = KafkaConnection() self.connection.configure(config) def test_open_close(self): self.connection.open(None) assert self.connection.is_open() is True assert self.connection.get_connection() is not None self.connection.close(None) assert self.connection.is_open() is False assert self.connection.get_connection() is None def test_list_topics(self): self.connection.open(None) assert self.connection.is_open() is True assert self.connection.get_connection() is not None topics = self.connection.read_queue_names() assert isinstance(topics, list) self.connection.close(None) assert self.connection.is_open() is False assert self.connection.get_connection() is None def test_create_delete_topics(self): topics = ['new_topic1', 'new_topic2'] self.connection.open(None) self.connection.create_queue(topics[0]) self.connection.create_queue(topics[1]) time.sleep(0.5) kafka_topics = self.connection.read_queue_names() assert topics[0] in kafka_topics assert topics[1] in kafka_topics self.connection.delete_queue(topics[0]) self.connection.delete_queue(topics[1]) time.sleep(0.5) kafka_topics = self.connection.read_queue_names() assert topics[0] not in kafka_topics assert topics[1] not in kafka_topics
2,508
783
# -------------------------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # -------------------------------------------------------------------------------------------- import argparse from azure.cli.command_modules.monitor.util import get_aggregation_map, get_operator_map def period_type(value): import re def _get_substring(indices): if indices == tuple([-1, -1]): return '' return value[indices[0]: indices[1]] regex = r'(p)?(\d+y)?(\d+m)?(\d+d)?(t)?(\d+h)?(\d+m)?(\d+s)?' match = re.match(regex, value.lower()) match_len = match.regs[0] if match_len != tuple([0, len(value)]): raise ValueError # simply return value if a valid ISO8601 string is supplied if match.regs[1] != tuple([-1, -1]) and match.regs[5] != tuple([-1, -1]): return value # if shorthand is used, only support days, minutes, hours, seconds # ensure M is interpretted as minutes days = _get_substring(match.regs[4]) minutes = _get_substring(match.regs[6]) or _get_substring(match.regs[3]) hours = _get_substring(match.regs[7]) seconds = _get_substring(match.regs[8]) return 'P{}T{}{}{}'.format(days, minutes, hours, seconds).upper() # pylint: disable=too-few-public-methods class ConditionAction(argparse.Action): def __call__(self, parser, namespace, values, option_string=None): from azure.mgmt.monitor.models import ThresholdRuleCondition, RuleMetricDataSource # get default description if not specified if namespace.description is None: namespace.description = ' '.join(values) if len(values) == 1: # workaround because CMD.exe eats > character... Allows condition to be # specified as a quoted expression values = values[0].split(' ') if len(values) < 5: from knack.util import CLIError raise CLIError('usage error: --condition METRIC {>,>=,<,<=} THRESHOLD {avg,min,max,total,last} DURATION') metric_name = ' '.join(values[:-4]) operator = get_operator_map()[values[-4]] threshold = int(values[-3]) aggregation = get_aggregation_map()[values[-2].lower()] window = period_type(values[-1]) metric = RuleMetricDataSource(None, metric_name) # target URI will be filled in later condition = ThresholdRuleCondition(operator, threshold, metric, window, aggregation) namespace.condition = condition # pylint: disable=protected-access class AlertAddAction(argparse._AppendAction): def __call__(self, parser, namespace, values, option_string=None): action = self.get_action(values, option_string) super(AlertAddAction, self).__call__(parser, namespace, action, option_string) def get_action(self, values, option_string): # pylint: disable=no-self-use from knack.util import CLIError _type = values[0].lower() if _type == 'email': from azure.mgmt.monitor.models import RuleEmailAction return RuleEmailAction(custom_emails=values[1:]) elif _type == 'webhook': from azure.mgmt.monitor.models import RuleWebhookAction uri = values[1] try: properties = dict(x.split('=', 1) for x in values[2:]) except ValueError: raise CLIError('usage error: {} webhook URI [KEY=VALUE ...]'.format(option_string)) return RuleWebhookAction(uri, properties) raise CLIError('usage error: {} TYPE KEY [ARGS]'.format(option_string)) class AlertRemoveAction(argparse._AppendAction): def __call__(self, parser, namespace, values, option_string=None): action = self.get_action(values, option_string) super(AlertRemoveAction, self).__call__(parser, namespace, action, option_string) def get_action(self, values, option_string): # pylint: disable=no-self-use # TYPE is artificially enforced to create consistency with the --add-action argument # but it could be enhanced to do additional validation in the future. from knack.util import CLIError _type = values[0].lower() if _type not in ['email', 'webhook']: raise CLIError('usage error: {} TYPE KEY [KEY ...]'.format(option_string)) return values[1:] class MultiObjectsDeserializeAction(argparse._AppendAction): # pylint: disable=protected-access def __call__(self, parser, namespace, values, option_string=None): type_name = values[0] type_properties = values[1:] try: super(MultiObjectsDeserializeAction, self).__call__(parser, namespace, self.get_deserializer(type_name)(*type_properties), option_string) except KeyError: raise ValueError('usage error: the type "{}" is not recognizable.'.format(type_name)) except TypeError: raise ValueError( 'usage error: Failed to parse "{}" as object of type "{}".'.format(' '.join(values), type_name)) except ValueError as ex: raise ValueError( 'usage error: Failed to parse "{}" as object of type "{}". {}'.format( ' '.join(values), type_name, str(ex))) def get_deserializer(self, type_name): raise NotImplementedError() class ActionGroupReceiverParameterAction(MultiObjectsDeserializeAction): def get_deserializer(self, type_name): from azure.mgmt.monitor.models import EmailReceiver, SmsReceiver, WebhookReceiver return {'email': EmailReceiver, 'sms': SmsReceiver, 'webhook': WebhookReceiver}[type_name]
5,964
1,653
import itertools import numpy as np import torch import unittest from mogwai.data_loading import one_hot from mogwai.models import GremlinPseudolikelihood class TestGremlinPL(unittest.TestCase): def setUp(self): torch.manual_seed(0) N = 100 L = 20 A = 8 msa = torch.randint(0, A, [N, L]) msa = torch.FloatTensor(one_hot(msa.numpy())) msa_counts = msa.sum(0) self.msa = msa self.model = GremlinPseudolikelihood(N, L, msa_counts, vocab_size=A) # Need nonzero weights but don't want to take a grad for this test wt = self.model.weight.data self.model.weight.data = torch.randn_like(wt) # Used for data leakage test. self.A = A def test_parameter_shapes(self): self.assertTupleEqual(self.model.weight.shape, (20, 8, 20, 8)) self.assertTupleEqual(self.model.bias.shape, (20, 8)) def test_forward_shape(self): batch = self.msa[:64] loss, logits = self.model(batch) self.assertTupleEqual(logits.shape, (64, 20, 8)) def onehot_vector(self, idx: int): oh = torch.zeros(self.A) oh[idx] = 1.0 return oh @torch.no_grad() def test_data_leakage(self): # Confirm that logits for position 0 do not change # when sequence at position 0 is exhaustively changed. logits_list = [] example = self.msa[0] seq_pos = 0 for i in range(self.A): example[seq_pos] = self.onehot_vector(i) _, logits = self.model(example.unsqueeze(0)) logits_list.append(logits[0, seq_pos]) all_pairs = itertools.combinations(logits_list, 2) for x, y in all_pairs: np.testing.assert_array_almost_equal(x.numpy(), y.numpy()) class TestGremlinPLGrad(unittest.TestCase): def setUp(self): torch.manual_seed(0) N = 100 L = 20 A = 8 msa = torch.randint(0, A, [N, L]) msa = torch.FloatTensor(one_hot(msa.numpy())) msa_counts = msa.sum(0) self.msa = msa self.model = GremlinPseudolikelihood(N, L, msa_counts, vocab_size=A) def test_gradient(self): # Tests that backward runs. batch = self.msa[:64] loss, _ = self.model(batch) loss.backward() # TODO: Presumably there's a less stupid approach self.assertTrue(True) if __name__ == "__main__": unittest.main()
2,465
906
from . import cupy_wrapper as acp from .cupy_vjps import ( untake, balanced_eq, match_complex, replace_zero, dot_adjoint_0, dot_adjoint_1, tensordot_adjoint_0, tensordot_adjoint_1, nograd_functions, ) from autograd.extend import ( defjvp, defjvp_argnum, def_linear, vspace, JVPNode, register_notrace, ) from autograd.util import func from .cupy_boxes import ArrayBox for fun in nograd_functions: register_notrace(JVPNode, fun) defjvp(func(ArrayBox.__getitem__), "same") defjvp(untake, "same") defjvp_argnum( acp.array_from_args, lambda argnum, g, ans, args, kwargs: untake(g, argnum - 2, vspace(ans)), ) defjvp( acp._array_from_scalar_or_array, None, None, lambda g, ans, args, kwargs, _: acp._array_from_scalar_or_array( args, kwargs, g ), ) # ----- Functions that are constant w.r.t. continuous inputs ----- # defjvp(acp.nan_to_num, lambda g, ans, x: acp.where(acp.isfinite(x), g, 0.)) # ----- Binary ufuncs (linear) ----- def_linear(acp.multiply) # ----- Binary ufuncs ----- defjvp( acp.add, lambda g, ans, x, y: broadcast(g, ans), lambda g, ans, x, y: broadcast(g, ans), ) defjvp( acp.subtract, lambda g, ans, x, y: broadcast(g, ans), lambda g, ans, x, y: broadcast(-g, ans), ) defjvp(acp.divide, "same", lambda g, ans, x, y: -g * x / y ** 2) defjvp( acp.maximum, lambda g, ans, x, y: g * balanced_eq(x, ans, y), lambda g, ans, x, y: g * balanced_eq(y, ans, x), ) defjvp( acp.minimum, lambda g, ans, x, y: g * balanced_eq(x, ans, y), lambda g, ans, x, y: g * balanced_eq(y, ans, x), ) defjvp( acp.fmax, lambda g, ans, x, y: g * balanced_eq(x, ans, y), lambda g, ans, x, y: g * balanced_eq(y, ans, x), ) defjvp( acp.fmin, lambda g, ans, x, y: g * balanced_eq(x, ans, y), lambda g, ans, x, y: g * balanced_eq(y, ans, x), ) defjvp( acp.logaddexp, lambda g, ans, x, y: g * acp.exp(x - ans), lambda g, ans, x, y: g * acp.exp(y - ans), ) defjvp( acp.logaddexp2, lambda g, ans, x, y: g * 2 ** (x - ans), lambda g, ans, x, y: g * 2 ** (y - ans), ) defjvp(acp.true_divide, "same", lambda g, ans, x, y: -g * x / y ** 2) defjvp( acp.mod, lambda g, ans, x, y: broadcast(g, ans), lambda g, ans, x, y: -g * acp.floor(x / y), ) defjvp( acp.remainder, lambda g, ans, x, y: broadcast(g, ans), lambda g, ans, x, y: -g * acp.floor(x / y), ) defjvp( acp.power, lambda g, ans, x, y: g * y * x ** acp.where(y, y - 1, 1.0), lambda g, ans, x, y: g * acp.log(replace_zero(x, 1.0)) * x ** y, ) defjvp( acp.arctan2, lambda g, ans, x, y: g * y / (x ** 2 + y ** 2), lambda g, ans, x, y: g * -x / (x ** 2 + y ** 2), ) # ----- Simple grads (linear) ----- defjvp(acp.negative, "same") defjvp(acp.rad2deg, "same") defjvp(acp.degrees, "same") defjvp(acp.deg2rad, "same") defjvp(acp.radians, "same") defjvp(acp.reshape, "same") defjvp(acp.roll, "same") defjvp(acp.array_split, "same") defjvp(acp.split, "same") defjvp(acp.vsplit, "same") defjvp(acp.hsplit, "same") defjvp(acp.dsplit, "same") defjvp(acp.ravel, "same") defjvp(acp.expand_dims, "same") defjvp(acp.squeeze, "same") defjvp(acp.diag, "same") defjvp(acp.diagonal, "same") defjvp(acp.make_diagonal, "same") defjvp(acp.flipud, "same") defjvp(acp.fliplr, "same") defjvp(acp.rot90, "same") defjvp(acp.trace, "same") defjvp(acp.full, "same", argnums=(1,)) defjvp(acp.triu, "same") defjvp(acp.tril, "same") defjvp(acp.swapaxes, "same") defjvp(acp.rollaxis, "same") defjvp(acp.moveaxis, "same") # def_linear(acp.cross) # ----- Simple grads ----- defjvp( acp.abs, lambda g, ans, x: acp.real(g * replace_zero(acp.conj(x), 0.0)) / replace_zero(ans, 1.0), ) # defjvp(acp.fabs, lambda g, ans, x : acp.sign(x) * g) # fabs doesn't take complex # numbers. defjvp(acp.absolute, lambda g, ans, x: acp.real(g * acp.conj(x)) / ans) defjvp(acp.reciprocal, lambda g, ans, x: -g / x ** 2) defjvp(acp.exp, lambda g, ans, x: ans * g) defjvp(acp.exp2, lambda g, ans, x: ans * acp.log(2) * g) defjvp(acp.expm1, lambda g, ans, x: (ans + 1) * g) defjvp(acp.log, lambda g, ans, x: g / x) defjvp(acp.log2, lambda g, ans, x: g / x / acp.log(2)) defjvp(acp.log10, lambda g, ans, x: g / x / acp.log(10)) defjvp(acp.log1p, lambda g, ans, x: g / (x + 1)) defjvp(acp.sin, lambda g, ans, x: g * acp.cos(x)) defjvp(acp.cos, lambda g, ans, x: -g * acp.sin(x)) defjvp(acp.tan, lambda g, ans, x: g / acp.cos(x) ** 2) defjvp(acp.arcsin, lambda g, ans, x: g / acp.sqrt(1 - x ** 2)) defjvp(acp.arccos, lambda g, ans, x: -g / acp.sqrt(1 - x ** 2)) defjvp(acp.arctan, lambda g, ans, x: g / (1 + x ** 2)) defjvp(acp.sinh, lambda g, ans, x: g * acp.cosh(x)) defjvp(acp.cosh, lambda g, ans, x: g * acp.sinh(x)) defjvp(acp.tanh, lambda g, ans, x: g / acp.cosh(x) ** 2) defjvp(acp.arcsinh, lambda g, ans, x: g / acp.sqrt(x ** 2 + 1)) defjvp(acp.arccosh, lambda g, ans, x: g / acp.sqrt(x ** 2 - 1)) defjvp(acp.arctanh, lambda g, ans, x: g / (1 - x ** 2)) defjvp(acp.square, lambda g, ans, x: g * 2 * x) defjvp(acp.sqrt, lambda g, ans, x: g * 0.5 * x ** -0.5) # defjvp(acp.sinc, lambda g, ans, x : g * (acp.cos(acp.pi*x)*acp.pi*x - # acp.sin(acp.pi*x))/(acp.pi*x**2)) defjvp( acp.clip, lambda g, ans, x, a_min, a_max: g * acp.logical_and(ans != a_min, ans != a_max), ) # defjvp(acp.real_if_close, lambda g, ans, x : match_complex(ans, g)) defjvp(acp.real, lambda g, ans, x: acp.real(g)) defjvp(acp.imag, lambda g, ans, x: match_complex(ans, -1j * g)) defjvp(acp.conj, lambda g, ans, x: acp.conj(g)) defjvp( acp.angle, lambda g, ans, x: match_complex( ans, g * acp.conj(x * 1j) / acp.abs(x) ** 2 ), ) defjvp( acp.where, None, lambda g, ans, c, x=None, y=None: acp.where(c, g, acp.zeros(g.shape)), lambda g, ans, c, x=None, y=None: acp.where(c, acp.zeros(g.shape), g), ) # ----- Trickier grads ----- defjvp(acp.kron, "same", "same") # defjvp(acp.diff, 'same') defjvp(acp.repeat, "same") defjvp(acp.tile, "same") defjvp(acp.transpose, "same") defjvp(acp.sum, "same") defjvp(acp.mean, "same") defjvp( acp.prod, lambda g, ans, x, axis=None, keepdims=False: ans * acp.sum(g / x, axis=axis, keepdims=keepdims), ) defjvp( acp.linspace, lambda g, ans, start, stop, *args, **kwargs: acp.linspace( g, 0, *args, **kwargs ), lambda g, ans, start, stop, *args, **kwargs: acp.linspace( 0, g, *args, **kwargs ), ) def forward_grad_np_var(g, ans, x, axis=None, ddof=0, keepdims=False): if axis is None: num_reps = acp.size(g) elif isinstance(axis, int): num_reps = g.shape[axis] elif isinstance(axis, tuple): num_reps = acp.prod(acp.array(np.shape(g))[list(axis)]) x_minus_mean = acp.conj(x - acp.mean(x, axis=axis, keepdims=True)) return ( 2.0 * acp.sum(acp.real(g * x_minus_mean), axis=axis, keepdims=keepdims) / (num_reps - ddof) ) defjvp(acp.var, forward_grad_np_var) def forward_grad_np_std(g, ans, x, axis=None, ddof=0, keepdims=False): if axis is None: num_reps = acp.size(g) elif isinstance(axis, int): num_reps = g.shape[axis] elif isinstance(axis, tuple): num_reps = acp.prod(acp.array(g.shape)[list(axis)]) if num_reps <= 1: return acp.zeros_like(ans) x_minus_mean = acp.conj(x - acp.mean(x, axis=axis, keepdims=True)) return acp.sum( acp.real(g * x_minus_mean), axis=axis, keepdims=keepdims ) / ((num_reps - ddof) * ans) defjvp(acp.std, forward_grad_np_std) def fwd_grad_chooser(g, ans, x, axis=None, keepdims=False): if acp.isscalar(x): return g if not keepdims: if isinstance(axis, int): ans = acp.expand_dims(ans, axis) elif isinstance(axis, tuple): for ax in sorted(axis): ans = acp.expand_dims(ans, ax) chosen_locations = x == ans return acp.sum( (g * chosen_locations), axis=axis, keepdims=keepdims ) / acp.sum(chosen_locations, axis=axis, keepdims=keepdims) defjvp(acp.max, fwd_grad_chooser) defjvp(acp.min, fwd_grad_chooser) defjvp(acp.amax, fwd_grad_chooser) defjvp(acp.amin, fwd_grad_chooser) defjvp(acp.cumsum, "same") def_linear(acp.inner) def_linear(acp.matmul) def_linear(acp.dot) def_linear(acp.tensordot) def_linear(acp.outer) def_linear(dot_adjoint_0) def_linear(dot_adjoint_1) def_linear(tensordot_adjoint_0) def_linear(tensordot_adjoint_1) def fwd_grad_concatenate_args(argnum, g, ans, axis_args, kwargs): result = [] for i in range(1, len(axis_args)): if i == argnum: result.append(g) else: result.append(acp.zeros_like(axis_args[i])) return acp.concatenate_args(axis_args[0], *result) defjvp_argnum(acp.concatenate_args, fwd_grad_concatenate_args) def fwd_grad_sort(g, ans, x, axis=-1, kind="quicksort", order=None): sort_perm = acp.argsort(x, axis, kind, order) return g[sort_perm] defjvp(acp.sort, fwd_grad_sort) defjvp(acp.msort, lambda g, ans, x: fwd_grad_sort(g, ans, x, axis=0)) def fwd_grad_partition( g, ans, x, kth, axis=-1, kind="introselect", order=None ): partition_perm = acp.argpartition(x, kth, axis, kind, order) return g[partition_perm] defjvp(acp.partition, fwd_grad_partition) def atleast_jvpmaker(fun): def jvp(g, ans, *arys): if len(arys) > 1: raise NotImplementedError("Can't handle multiple arguments yet.") return fun(g) return jvp defjvp(acp.atleast_1d, atleast_jvpmaker(acp.atleast_1d)) defjvp(acp.atleast_2d, atleast_jvpmaker(acp.atleast_2d)) defjvp(acp.atleast_3d, atleast_jvpmaker(acp.atleast_3d)) def_linear(acp.einsum) # TODO(mattjj): can we call np.broadcast_to or a related function instead? def broadcast(x, target): target_shape, target_ndim, target_dtype, target_iscomplex = acp.metadata( target ) while acp.ndim(x) < target_ndim: x = acp.expand_dims(x, 0) for axis, size in enumerate(x.shape): if size == 1: x = acp.repeat(x, target_shape[axis], axis=axis) if target_iscomplex and not acp.iscomplexobj(x): x = x + 0j # TODO(mattjj): this might promote the dtype return x
10,233
4,775
# !/usr/bin/env python # -*- coding = utf-8 -*- # @Author:wanghui # @Time: # @File:getStartActivityConfig.py import configparser import os BASE_DIR = os.path.dirname(os.path.abspath(__file__)) class GetStartActivityConfig(object): def __init__(self): self.config = configparser.ConfigParser() self.read_config() def read_config(self): """读取配置文件""" try: self.config.read(os.path.join(BASE_DIR, '')) except Exception as e: print(f'配置文件不正确: ', e) return None def get_set_up(self): """获取setUp的信息""" return self.config.items('SETUP') def get_tear_down(self): """获取setUp的信息""" return self.config.items('TEARDOWN') def get_sections(self): """ 获取所有的section """ return self.config.sections() def get_option(self, section): """获取当前section下的所有options""" return self.config.options(section) def get_section_items(self, section): """获取当前section下的所有键值对""" return self.config.items(section) def get_section_password(self, section, option): """获取当前option对应的值""" return self.config.get(section, option) def check_config(self, *arg): """检查配置文件信息是否正确""" try: self.read_config() if len(arg) == 1: # 判断是否有section return self.config.has_section(arg[0]) elif len(arg) == 3: # 判断section下 option是否正确 if self.config[arg[0]][arg[1]] == arg[2]: return True else: return False else: return False except Exception as e: return False def main(): config = GetStartActivityConfig() print(config.get_set_up()) # print(config.get_section_items('360')) # print(config.get_section_password('360', 'platformVersion')) if __name__ == '__main__': main()
1,950
663
# MIT License # # Copyright (c) 2019 Daniel Brotsky # # Portions copyright (c) 2017 by Peter Norvig and Naoki Shibuya # (See https://towardsdatascience.com/peter-norvigs-sudoku-solver-25779bb349ce) # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. from typing import Dict from .board import Board class Solver: """This solver was originally published by Peter Norvig. It was later updated to py3 by Naoki Shibuya. See https://towardsdatascience.com/peter-norvigs-sudoku-solver-25779bb349ce for the code and an explanation of how it works. Small adaptations have been made to adapt for multiple sizes of puzzle and to invoke the solver from the generator. """ # digits, row labels and col labels up to 16 (max puzzle size) all_digits = '123456789ABCDEFG' all_rows = 'ABCDEFGHIJKLMNOP' all_cols = 'abcdefghijklmnop' def __init__(self, board: Board): """Captures the board so you can solve it non-destructively.""" self.digits = self.all_digits[:board.side_length] self.rows = self.all_rows[:board.side_length] self.cols = self.all_cols[:board.side_length] self.squares = self.cross(self.rows, self.cols) self.unit_list = ([self.cross(self.rows, c) for c in self.cols] + [self.cross(r, self.cols) for r in self.rows] + [self.cross(rs, cs) for rs in self.rank_groups(self.rows, board.rank, board.side_length) for cs in self.rank_groups(self.cols, board.rank, board.side_length)]) self.units = dict((s, [u for u in self.unit_list if s in u]) for s in self.squares) self.peers = dict((s, set(sum(self.units[s], [])) - {s}) for s in self.squares) self.values = self.board2values(board) self.solution = {} def board2values(self, board): return self.parse_grid(''.join([str(c) for c in board.cells])) def values2board(self, values): try: return Board([self.all_digits.index(values[s][0])+1 for s in self.squares]) except ValueError: return {} def is_solution(self): return all(len(self.values[s]) == 1 for s in self.squares) def cross(self, aa: [str], bb: [str]): """Cross product of elements in aa and elements in bb.""" return [a + b for a in aa for b in bb] def rank_groups(self, indices: str, step: int, end: int) -> [str]: return [indices[start:start + step] for start in range(0, end, step)] def parse_grid(self, grid: str) -> Dict[str, str]: """Convert grid to a dict of possible values, {square: digits}, or return {} if a contradiction is detected.""" # To start, every square can be any digit; then assign values from the grid. values = dict((s, self.digits) for s in self.squares) for s, d in self.grid_values(grid).items(): if d in self.digits and not self.assign(values, s, d): return {} # (Fail if we can't assign d to square s.) return values def grid_values(self, grid: str) -> Dict[str, str]: """Convert grid into a dict of {square: char} with '0' or '.' for empties.""" chars = [c for c in grid if c in self.digits or c in '0.-'] if len(chars) != len(self.squares): raise ValueError("Grid contains illegal characters") return dict(zip(self.squares, chars)) def assign(self, values: Dict[str, str], s: str, d: str): """Eliminate all the other values (except d) from values[s] and propagate. Return values, except return {} if a contradiction is detected.""" other_values = values[s].replace(d, '') if all(self.eliminate(values, s, d2) for d2 in other_values): return values else: return {} def eliminate(self, values: Dict[str, str], s: str, d: str): """Eliminate d from values[s]; propagate when values or places <= 2. Return values, except return {} if a contradiction is detected.""" if d not in values[s]: return values # Already eliminated values[s] = values[s].replace(d, '') # (1) If a square s is reduced to one value d2, then eliminate d2 from the peers. if len(values[s]) == 0: return {} # Contradiction: removed last value elif len(values[s]) == 1: d2 = values[s] if not all(self.eliminate(values, s2, d2) for s2 in self.peers[s]): return {} # (2) If a unit u is reduced to only one place for a value d, then put it there. for u in self.units[s]: places = [s for s in u if d in values[s]] if len(places) == 0: return {} # Contradiction: no place for this value elif len(places) == 1: # d can only be in one place in unit; assign it there if not self.assign(values, places[0], d): return {} return values def can_solve(self): values = self.search(self.values) if values: self.solution = self.values2board(values) return True return False def search(self, values: Dict[str, str]) -> Dict[str, str]: """Using depth-first search and propagation, try all possible values.""" if not values: return {} # Failed earlier if all(len(values[s]) == 1 for s in self.squares): return values # Solved! # Chose the unfilled square s with the fewest possibilities n, s = min((len(values[s]), s) for s in self.squares if len(values[s]) > 1) return next(filter(len, (self.search(self.assign(values.copy(), s, d)) for d in values[s])), {})
6,773
2,131
"""Decorators used for adding functionality to the library.""" from deserialize.exceptions import NoDefaultSpecifiedException def default(key_name, default_value): """A decorator function for mapping default values to key names.""" def store_defaults_map(class_reference): """Store the defaults map.""" if not hasattr(class_reference, "__deserialize_defaults_map__"): setattr(class_reference, "__deserialize_defaults_map__", {}) class_reference.__deserialize_defaults_map__[key_name] = default_value return class_reference return store_defaults_map def _has_default(class_reference, key_name): """Returns True if this key has a default, False otherwise. :returns: True if this key has a default, False otherwise. """ if not hasattr(class_reference, "__deserialize_defaults_map__"): return False return key_name in class_reference.__deserialize_defaults_map__ def _get_default(class_reference, key_name): """Get the default value for the given class and key name. :raises NoDefaultSpecifiedException: If a default hasn't been specified """ if not hasattr(class_reference, "__deserialize_defaults_map__"): raise NoDefaultSpecifiedException() if key_name in class_reference.__deserialize_defaults_map__: return class_reference.__deserialize_defaults_map__[key_name] raise NoDefaultSpecifiedException()
1,442
418
from matplotlib import pyplot as plt import gpflow import tensorflow as tf import os import numpy as np import cProfile def outputGraph(model, dirName, fileName): model.compile() if not(os.path.isdir(dirName)): os.mkdir(dirName) fullFileName = os.path.join(dirName, fileName) if os.path.isfile(fullFileName): os.remove(fullFileName) tf.train.write_graph(model.session.graph_def, dirName+'/', fileName, as_text=False) # build a very simple data set: def getData(): rng = np.random.RandomState(1) N = 30 X = rng.rand(N,1) Y = np.sin(12*X) + 0.66*np.cos(25*X) + rng.randn(N,1)*0.1 + 3 return X,Y def plotData(X,Y): plt.figure() plt.plot(X, Y, 'kx', mew=2) def getRegressionModel(X,Y): #build the GPR object k = gpflow.kernels.Matern52(1) meanf = gpflow.mean_functions.Linear(1,0) m = gpflow.models.GPR(X, Y, k, meanf) m.likelihood.variance = 0.01 print "Here are the parameters before optimization" m return m def optimizeModel(m): m.optimize() print "Here are the parameters after optimization" m def plotOptimizationResult(X,Y,m): #plot! xx = np.linspace(-0.1, 1.1, 100)[:,None] mean, var = m.predict_y(xx) plt.figure() plt.plot(X, Y, 'kx', mew=2) plt.plot(xx, mean, 'b', lw=2) plt.plot(xx, mean + 2*np.sqrt(var), 'b--', xx, mean - 2*np.sqrt(var), 'b--', lw=1.2) def setModelPriors(m): #we'll choose rather arbitrary priors. m.kern.lengthscales.prior = gpflow.priors.Gamma(1., 1.) m.kern.variance.prior = gpflow.priors.Gamma(1., 1.) m.likelihood.variance.prior = gpflow.priors.Gamma(1., 1.) m.mean_function.A.prior = gpflow.priors.Gaussian(0., 10.) m.mean_function.b.prior = gpflow.priors.Gaussian(0., 10.) print "model with priors ", m def getSamples(m): samples = m.sample(100, epsilon = 0.1) return samples def plotSamples(X, Y, m, samples): xx = np.linspace(-0.1, 1.1, 100)[:,None] plt.figure() plt.plot(samples) f, axs = plt.subplots(1,3, figsize=(12,4), tight_layout=True) axs[0].plot(samples[:,0], samples[:,1], 'k.', alpha = 0.15) axs[0].set_xlabel('noise_variance') axs[0].set_ylabel('signal_variance') axs[1].plot(samples[:,0], samples[:,2], 'k.', alpha = 0.15) axs[1].set_xlabel('noise_variance') axs[1].set_ylabel('lengthscale') axs[2].plot(samples[:,2], samples[:,1], 'k.', alpha = 0.1) axs[2].set_xlabel('lengthscale') axs[2].set_ylabel('signal_variance') #an attempt to plot the function posterior #Note that we should really sample the function values here, instead of just using the mean. #We are under-representing the uncertainty here. # TODO: get full_covariance of the predictions (predict_f only?) plt.figure() for s in samples: m.set_state(s) mean, _ = m.predict_y(xx) plt.plot(xx, mean, 'b', lw=2, alpha = 0.05) plt.plot(X, Y, 'kx', mew=2) def showAllPlots(): plt.show() def runExperiments(plotting=True,outputGraphs=False): X,Y = getData() if plotting: plotData(X,Y) m = getRegressionModel(X,Y) if outputGraphs: modelDir = 'models' outputGraph(m, modelDir, 'pointHypers') optimizeModel(m) if plotting: plotOptimizationResult(X,Y,m) setModelPriors(m) if outputGraphs: outputGraph(m, modelDir, 'bayesHypers') samples = getSamples(m) if plotting: plotSamples(X, Y, m, samples) showAllPlots() if __name__ == '__main__': runExperiments() #cProfile.run('runExperiments(plotting=False)')
3,608
1,443
import bpy import bmesh import base64 import math import mathutils import random import struct import zlib from array import array from itertools import islice from bpy.props import IntProperty, PointerProperty, StringProperty from bpy.types import Object, Operator, Panel, PropertyGroup from mathutils import Vector from mathutils.bvhtree import BVHTree #---------------------------------------------------------------------------# # Operators class AgnosiaCreatePointcloudOperator(Operator): bl_idname = "object.create_pointcloud" bl_label = "Create pointcloud" bl_options = {'REGISTER'} def execute(self, context): if context.mode != "OBJECT": self.report({'WARNING'}, "Create pointcloud: must be in Object mode.") return {'CANCELLED'} target = context.object if (target is None) or (target.type != 'MESH'): self.report({'WARNING'}, "Create pointcloud: must select a Mesh object.") return {'CANCELLED'} if target.pointclouds: self.report({'WARNING'}, "Create pointcloud: can't create a pointcloud from a pointcloud.") return {'CANCELLED'} # Deselect and hide the sampled object. target.select_set(False) target.hide_set(True) # Create a new pointcloud. o = create_pointcloud_from(context, target) # Make the pointcloud active, and select it. context.view_layer.objects.active = o o.select_set(True) # And begin updating the new pointcloud. bpy.ops.object.update_pointcloud() return {'FINISHED'} class AgnosiaUpdatePointcloudOperator(Operator): bl_idname = "object.update_pointcloud" bl_label = "Update pointcloud" bl_options = set() _timer = None _generator = None _finished = False _cancelled = False _object = None # Class variable _running_on = {} def execute(self, context): self._object = context.object # Only allow one instance of the operator to run on any given object at a time. prior_op = self.__class__._running_on.get(self._object) if prior_op is not None: prior_op.abort() self.__class__._running_on[self._object] = self self._generator = update_pointcloud_iter(self._object) self._cancelled = False self._finished = False wm = context.window_manager self._timer = wm.event_timer_add(0.1, window=context.window) wm.modal_handler_add(self) return {'RUNNING_MODAL'} def modal(self, context, event): if event.type in {'RIGHTMOUSE', 'ESC'}: self._cancelled = True if self._cancelled or self._finished: # Remove ourselves if self.__class__._running_on.get(self._object) == self: del self.__class__._running_on[self._object] # Remove the timer wm = context.window_manager wm.event_timer_remove(self._timer) self._timer = None if self._cancelled: return {'CANCELLED'} elif self._finished: return {'FINISHED'} elif event.type == 'TIMER': try: next(self._generator) except StopIteration: self._finished = True return {'PASS_THROUGH'} def abort(self): self._cancelled = True class AgnosiaPointcloudExportOperator(Operator): bl_idname = "object.export_pointcloud" bl_label = "Export pointcloud" bl_options = {'REGISTER'} filepath : bpy.props.StringProperty(subtype="FILE_PATH") @classmethod def poll(cls, context): o = context.object return ( (context.mode == 'OBJECT') and (o is not None) and (len(o.pointclouds) > 0) ) def invoke(self, context, event): context.window_manager.fileselect_add(self) return {'RUNNING_MODAL'} def execute(self, context): o = context.object pc = o.pointclouds[0] def to_uint8(f): return min(max(0, int(f * 255.0)), 255) with PointcloudBinWriter(self.filepath) as f: vertices = pc.raw_vertices normals = pc.raw_normals colors = pc.raw_colors def records(): v_it = iter(vertices) n_it = iter(normals) c_it = iter(colors) while True: v_xyz = list(islice(v_it, 3)) n_xyz = list(islice(n_it, 3)) c_rgb = [to_uint8(f) for f in list(islice(c_it, 4))][:3] if v_xyz and c_rgb: yield v_xyz + c_rgb else: break for r in records(): f.write(*r) return {'FINISHED'} #---------------------------------------------------------------------------# # Panels class AGNOSIA_PT_pointcloud(Panel): bl_label = "Pointcloud" bl_idname = "AGNOSIA_PT_pointcloud" bl_space_type = "VIEW_3D" bl_region_type = "UI" bl_category = "Agnosia" bl_context = "objectmode" @classmethod def poll(self, context): o = context.object if o is None: return False if not o.select_get(): return False if not o.pointclouds: return False return True def draw(self, context): o = context.object pc = o.pointclouds[0] layout = self.layout row = layout.row(align=True) box = row.box() box.label(text="There is nothing here that you recognise. Yet."); box = layout.box() box.prop(pc, 'target') box.prop(pc, 'point_count') box.prop(pc, 'seed') layout.operator('object.export_pointcloud', text="Export .bin") #---------------------------------------------------------------------------# # Pointcloud property and data def _pointcloud_property_update(self, context): bpy.ops.object.update_pointcloud() class PointcloudProperty(PropertyGroup): target : PointerProperty(name="Sample", type=Object, update=_pointcloud_property_update) point_count : IntProperty(name="Point count", default=1024, min=128, step=64, update=_pointcloud_property_update) seed : IntProperty(name="Seed", default=0, update=_pointcloud_property_update) raw_vertices_string : StringProperty(name="_RawVerticesString", default="") raw_normals_string : StringProperty(name="_RawNormalsString", default="") raw_colors_string : StringProperty(name="_RawColorsString", default="") _raw_cache = None @staticmethod def _pack_array(a): if a: b = a.tobytes() c = zlib.compress(b) d = base64.encodebytes(c) return d.decode('ascii') else: return "" @staticmethod def _unpack_array(s, typecode): if s: a = array(typecode) b = bytes(s, 'ascii') c = base64.decodebytes(b) d = zlib.decompress(c) a.frombytes(d) return a else: return array(typecode) @property def raw_cache(self): cache = self.__dict__.get('_raw_cache') if cache is None: cache = {} self.__dict__['_raw_cache'] = cache return cache @property def raw_vertices(self): if self.raw_vertices_string: value = self.raw_cache.get('vertices') if value is None: value = self._unpack_array(self.raw_vertices_string, 'f') self.raw_cache['vertices'] = value return value else: return array('f') @property def raw_normals(self): if self.raw_normals_string: value = self.raw_cache.get('normals') if value is None: value = self._unpack_array(self.raw_normals_string, 'f') self.raw_cache['normals'] = value return value else: return array('f') @property def raw_colors(self): if self.raw_colors_string: value = self.raw_cache.get('colors') if value is None: value = self._unpack_array(self.raw_colors_string, 'f') self.raw_cache['colors'] = value return value else: return array('f') def set_raw_data(self, vertices, normals=None, colors=None): if (not isinstance(vertices, array)) or (vertices.typecode != 'f'): raise ValueError("vertices must be type array('f')") if len(vertices) % 3 != 0: raise ValueError("vertices length must be multiple of 3") vertex_count = len(vertices) // 3 if (normals is not None): if (not isinstance(normals, array)) or (normals.typecode != 'f'): raise ValueError("normals must be type array('f')") if len(normals) != (3 * vertex_count): raise ValueError("len(normals) must be 3 * vertex_count") if (colors is not None): if (not isinstance(colors, array)) or (colors.typecode != 'f'): raise ValueError("colors must be type array('f')") if len(colors) != 4 * vertex_count: raise ValueError("len(colors) must be 4 * vertex_count") self.raw_vertices_string = self._pack_array(vertices) self.raw_normals_string = self._pack_array(normals) self.raw_colors_string = self._pack_array(colors) # Cached self.raw_cache['vertices'] = vertices self.raw_cache['normals'] = normals self.raw_cache['colors'] = colors #---------------------------------------------------------------------------# # Material def layout_nodes(node_tree, root_node): """Make all the nodes in node_tree, starting from root_node, nice and tidy.""" from collections import defaultdict from math import ceil # Lookup table of nodes to their incoming links incoming = defaultdict(list) for l in node_tree.links: incoming[l.to_node].append(l) # Lookup table of nodes to their sort keys sort_keys = {} sort_keys[root_node] = ('_root',) all_columns = [[root_node]] links = list(incoming[root_node]) # Arrange all the nodes from the root nodes into columns, # with each column's nodes in order by the outputs and nodes they feed into. while links: # Drop all the nodes on all the links into this column column = [] for l in links: # k = ((l.to_socket.name, l.from_socket.name), ) + sort_keys[l.to_node] k = (l.to_socket.name, ) + sort_keys[l.to_node] other_k = sort_keys.get(l.from_node, None) if other_k is not None: k = max(k, other_k) sort_keys[l.from_node] = k if l.from_node not in column: column.append(l.from_node) column.sort(key=sort_keys.get) all_columns.append(column) # Get the next set of links to sort links = [] for n in column: links.extend(incoming[n]) # Now lay out all the nodes right-to-left, with each column vertically # centered with respect to all the other columns. Coordinates are +Y up, +X right. grid_size = 20.0 def total_height(n): # Height of a node including its title bar. Not exact numbers, but good enough. return (20.0 if n.hide else (n.height + 30.0)) column_location = Vector((0.0, 0.0)) # x: right edge, y: center. spacing = Vector((3.0, 2.0)) * grid_size for i, column in enumerate(all_columns): # Calculate the total size max_node_width = max(ceil(n.width) for n in column) total_node_height = sum(ceil(total_height(n)) for n in column) total_spacing = spacing[1] * (len(column) - 1) column_width = ceil(max_node_width / grid_size) * grid_size column_height = total_node_height + total_spacing # Lay out these nodes vertically down the column. x = column_location[0] - (column_width / 2.0) y = column_location[1] + (column_height / 2.0) for n in column: node_x = round(x - n.width / 2.0) node_y = y #round(y - total_height(n) / 2.0) n.location = Vector((node_x, node_y)) y -= (ceil(total_height(n) + spacing[1])) column_location[0] -= (column_width + spacing[0]) def define_pointcloud_material(material): material.use_nodes = True tree = material.node_tree nodes = tree.nodes links = tree.links nodes.clear() output = nodes.new(type='ShaderNodeOutputMaterial') # output.location = self._grid_location(6, 4) diffuse = nodes.new(type='ShaderNodeBsdfDiffuse') diffuse.inputs['Color'].default_value = (1.0, 1.0, 1.0, 1.0) diffuse.inputs['Roughness'].default_value = 0.5 links.new(diffuse.outputs['BSDF'], output.inputs['Surface']) # Get colors and normals from the vertex color layers. colors = nodes.new(type='ShaderNodeAttribute') colors.label = "PointColor Attribute" colors.attribute_name = 'PointColor' links.new(colors.outputs['Color'], diffuse.inputs['Color']) normals = nodes.new(type='ShaderNodeAttribute') colors.label = "PointNormal Attribute" normals.attribute_name = 'PointNormal' # Create nodes to unpack the normals from the second vertex color layer. combine = nodes.new(type='ShaderNodeCombineXYZ') combine.hide = True links.new(combine.outputs['Vector'], diffuse.inputs['Normal']) separate = nodes.new(type='ShaderNodeSeparateXYZ') separate.hide = True links.new(normals.outputs['Vector'], separate.inputs['Vector']) # Each of the X, Y, and Z channels needs (foo - 0.5) * 2.0 for i in range(3): sub = nodes.new(type='ShaderNodeMath') sub.label = " - 0.5" sub.operation = 'SUBTRACT' sub.hide = True sub.inputs[1].default_value = 0.5 links.new(separate.outputs[i], sub.inputs[0]) mul = nodes.new(type='ShaderNodeMath') mul.label = " * 2.0" mul.operation = 'MULTIPLY' mul.hide = True mul.inputs[1].default_value = 2.0 links.new(sub.outputs[0], mul.inputs[0]) links.new(mul.outputs[0], combine.inputs[i]) layout_nodes(tree, output) def get_pointcloud_material(): name = 'PointcloudMaterial' m = bpy.data.materials.get(name) if not m: m = bpy.data.materials.new(name) define_pointcloud_material(m); return m def assign_material(o, mat): if (o.data.materials): o.data.materials[0] = mat else: o.data.materials.append(mat) #---------------------------------------------------------------------------# # Pointcloud objects. def create_pointcloud_from(context, target): o = create_empty_mesh_obj(context, 'Pointcloud') pc = o.pointclouds.add() pc.target = target pc.seed = random.randint(-2**31, 2**31) return o def create_empty_mesh_obj(context, name): mesh = bpy.data.meshes.new(name + 'Mesh') o = bpy.data.objects.new(name, mesh) o.show_name = True context.scene.collection.objects.link(o) return o def update_pointcloud_iter(o): if not o.pointclouds: return pc = o.pointclouds[0] target = pc.target if (target is None) or (target.type != 'MESH') or (target.pointclouds): return seed = pc.seed rng = random.Random(seed) for data in generate_points(pc.target, pc.point_count, rng, step_count=4096): yield vertices_arr = array('f', (f for vec in data[0] for f in vec)) normals_arr = array('f', (f for vec in data[1] for f in vec)) colors_arr = array('f', (f for vec in data[2] for f in vec)) pc.set_raw_data(vertices_arr, normals=normals_arr, colors=colors_arr) o.data = create_pointcloud_mesh(o.data.name, data) assign_material(o, get_pointcloud_material()) def generate_points(target, count, rng=random, step_count=0): if not step_count: step_count = count total_count = 0 total_data = [[], [], []] while total_count < count: step_count = min(step_count, (count - total_count)) # data = sphere_sample_obj(target, step_count, rng) # data = volume_sample_obj(target, step_count, rng) data = surface_sample_obj(target, step_count, rng) for i in range(len(total_data)): total_data[i] += data[i] total_count += step_count if total_count < count: yield list(total_data) yield total_data #---------------------------------------------------------------------------# # Meshes for in-Blender visualization. def create_pointcloud_mesh(name, data): mesh = bpy.data.meshes.new(name) (vertices, normals, colors) = data # Expand each vertex to make a quad facing the -y axis. if vertices: (vertices, faces, normals, colors) = \ expand_vertex_data_to_mesh(vertices, normals, colors) mesh.from_pydata(vertices, [], faces) mesh.validate(verbose=True, clean_customdata=False) mesh.update() # Apply per-vertex colors and normals color_layer = mesh.vertex_colors.new(name='PointColor') for (i, color) in enumerate(colors): color_layer.data[i].color = color normal_layer = mesh.vertex_colors.new(name='PointNormal') for (i, normal) in enumerate(normals): # Pack the normals into the color data n = (normal / 2.0) + Vector((0.5, 0.5, 0.5)) normal_layer.data[i].color = (n[0], n[1], n[2], 0.0) return mesh def expand_vertex_data_to_mesh(vertices, normals, colors): expanded_vertices = [] expanded_normals = [] expanded_colors = [] faces = [] # Size of the mesh representing a point. scale = 0.05 quad = ( Vector((1, 0, 1)) * scale, Vector((-1, 0, 1)) * scale, Vector((-1, 0, -1)) * scale, Vector((1, 0, -1)) * scale, ) # Expand the source data to a quad. for v in vertices: expanded_vertices.extend((v + quad[0], v + quad[1], v + quad[2], v + quad[3])) for n in normals: expanded_normals.extend((n, n, n, n)) for c in colors: expanded_colors.extend((c, c, c, c)) # Generate faces for i in range(len(vertices)): base = (4 * i) faces.append(( base + 0, base + 1, base + 2, base + 3, )) return (expanded_vertices, faces, expanded_normals, expanded_colors) #---------------------------------------------------------------------------# # Sampling. def sphere_sample_obj(o, count, rng): # Sample the object by raycasting from a sphere surrounding it # towards the origin. vertices = [] normals = [] colors = [] radius = object_bounding_radius(o) + 0.1 it = iter(sphere_surface_points(radius, rng)) while len(vertices) < count: pt = next(it) result, position, normal, index = raycast_to_origin(o, pt) if result: vertices.append(position) normals.append(normal) colors.append((1.0, 0.0, 1.0, 1.0)) return (vertices, normals, colors) def volume_sample_obj(o, count, rng): # Sample the object by generating points within its bounds and # testing if they're inside it. Assumes the mesh is watertight. vertices = [] normals = [] colors = [] bm = bmesh.new() bm.from_mesh(o.data) bvh = BVHTree.FromBMesh(bm) halfwidth = object_bounding_halfwidth(o) + 0.1 it = iter(cube_volume_points(halfwidth, rng)) while len(vertices) < count: pt = next(it) (location, normal, index, distance) = raycast_to_exterior(bvh, pt) pt_is_inside = (location is not None) if pt_is_inside: vertices.append(location) normals.append(normal) # TEMP: color each point by its coordinates r = (abs(location[0]) / halfwidth) g = (abs(location[1]) / halfwidth) b = (abs(location[2]) / halfwidth) colors.append((r, g, b, 1.0)) return (vertices, normals, colors) def surface_sample_obj(o, count, rng): # Sample the object by generating points on the surfaces of its tris. vertices = [] normals = [] colors = [] mesh = o.data # Find the surface area of each poly and the whole mesh. poly_areas = [p.area for p in mesh.polygons] surface_area = sum(poly_areas) # Generate uniform random area targets. area_targets = sorted(rng.uniform(0, surface_area) for _ in range(count)) # Iterate the polys to see which reaches the target. area_so_far = 0 target = area_targets.pop(0) for i, poly in enumerate(mesh.polygons): area_so_far += poly_areas[i] while target <= area_so_far: # Spawn a point. poly_vertices = [Vector(mesh.vertices[j].co) for j in poly.vertices] location = polygon_surface_point(poly_vertices, rng) normal = poly.normal # Save the point. vertices.append(location) normals.append(normal) # TEMP: color each point by its coordinates halfwidth = object_bounding_halfwidth(o) + 0.1 r = (abs(location[0]) / halfwidth) g = (abs(location[1]) / halfwidth) b = (abs(location[2]) / halfwidth) colors.append((r, g, b, 1.0)) # Get a new target if area_targets: target = area_targets.pop(0) else: # If we've run out of targets, then we have enough points. break if not vertices: print(f"ERROR: didn't generate any vertices!") return (vertices, normals, colors) def object_bounding_radius(o): from math import sqrt radius = 0.0 for (x, y, z) in o.bound_box: radius = max(radius, sqrt(x*x + y*y + z*z)) return radius def object_bounding_halfwidth(o): halfwidth = 0.0 for (x, y, z) in o.bound_box: halfwidth = max(halfwidth, abs(x), abs(y), abs(z)) return halfwidth def sphere_surface_points(radius, rng): # Generate Vectors randomly distributed on the surface of # a sphere with the given radius. while True: from math import acos, cos, pi, sin, sqrt u = rng.random() v = rng.random() theta = 2 * pi * u phi = acos(2 * v - 1) x = radius * cos(theta) * sin(phi) y = radius * sin(theta) * sin(phi) z = radius * cos(phi) yield Vector((x, y, z)) def cube_volume_points(halfwidth, rng): # Generate Vectors randomly distributed within the volume # of a cube with the given halfwidth. while True: u = rng.uniform(-1, 1) v = rng.uniform(-1, 1) w = rng.uniform(-1, 1) x = halfwidth * u y = halfwidth * v z = halfwidth * w yield Vector((x, y, z)) def polygon_surface_point(vertices, rng): # Return a random point on the surface of the polygon. # Determine the vertices abc of each triangle, and the vectors ab and ac. tris = [ ( vertices[0], # a vertices[i+1], # b vertices[i+2], # c (vertices[i+1] - vertices[0]), # ab (vertices[i+2] - vertices[0]), # ac ) for i in range(len(vertices) - 2) ] # Find the area of each tri and the total polygon area. tri_areas = [ (ab.cross(ac)).length / 2.0 for (_, _, _, ab, ac) in tris ] surface_area = sum(tri_areas) # Pick a target tri by area target = rng.uniform(0, surface_area) area_so_far = 0 for (i, (a, b, c, ab, ac)) in enumerate(tris): area_so_far += tri_areas[i] while target <= area_so_far: # Pick a point in this tri r1 = rng.random() r2 = rng.random() r1root = math.sqrt(r1) pt = ((1 - r1root) * a + r1root * (1 - r2) * b + r1root * r2 * c) return pt # Shouldn't get here, but if we do, just return a vertex print("ERROR: failed to find target tri.") return vertices[0] def raycast_to_origin(o, pt): # Raycast the object o from pt (in object space) to its origin. # Return a tuple: (result, position, normal, index) origin = Vector((0.0, 0.0, 0.0)) direction = (origin - pt).normalized() return o.ray_cast(pt, direction) def raycast_to_exterior(bvh, pt): """Raycast the BVHTree bvh from pt to the object's exterior. If pt is on the object's interior, return (location, normal, index, distance); if it's on the exterior, return (None, None, None, None).""" NO_HIT = (None, None, None, None) # If the point's too close to the origin, we can't get a proper direction, # so just skip it. origin = Vector((0, 0, 0)) from_origin = (pt - origin) if (from_origin.length < 0.0001): return NO_HIT ray_origin = pt direction = from_origin.normalized() # Raycast from the point towards the exterior, iterating # until we don't hit any faces. tiny_step = (direction * 0.0001) first_outward = None outward_crossings = 0 (location, normal, index, distance) = bvh.ray_cast(ray_origin, direction) if location is None: # Didn't hit anything, so we're done. return NO_HIT # Check if the face is oriented towards the ray or away from it. inward_facing = (direction.dot(normal) < 0) if inward_facing: # Must have been outside the object. return NO_HIT return (location, normal, index, distance) #---------------------------------------------------------------------------# # Pointcloud file types ## Structures in binary pointcloud file def bin_size(size): return struct.pack('=L', size) def bin_point(x, y, z, r, g, b): return struct.pack('=fffBBBx', x, y, z, r, g, b) ## Binary pointcloud writing class PointcloudBinWriter: # File format: # uint32_t size_of_data # struct record { # float x, y, z; # uint8_t r, g, b; # uint8_t pad; # } records[size / sizeof(struct record)] def __init__(self, filename): self.filename = filename self.file = None self.count = 0 self.size = 0 def write(self, x, y, z, r, g, b): assert (self.file is not None), "File is not open." blob = bin_point(x, y, z, r, g, b) self.file.write(blob) self.size += len(blob) self.count += 1 def __len__(self): return self.count def __enter__(self): self.file = open(self.filename, 'wb') # The file starts with the size of its data. We write a zero # initially, and fill in the actual size on __exit__(). self.file.write(bin_size(0)) return self def __exit__(self, exc_type, exc_value, traceback): if not exc_type and not exc_value: self.file.seek(0) self.file.write(bin_size(self.size)) self.file.close() #---------------------------------------------------------------------------# # Utils. from contextlib import contextmanager @contextmanager def tempfile(suffix='', dir=None): """ Context for temporary file. Will find a free temporary filename upon entering and will try to delete the file on leaving, even in case of an exception. Parameters ---------- suffix : string optional file suffix dir : string optional directory to save temporary file in """ # From: https://stackoverflow.com/a/29491523 import os import tempfile as tmp tf = tmp.NamedTemporaryFile(delete=False, suffix=suffix, dir=dir) tf.file.close() try: yield tf.name finally: try: os.remove(tf.name) except OSError as e: if e.errno == 2: pass else: raise @contextmanager def file_atomic(filepath, *args, **kwargs): """ Open temporary file object that atomically moves to destination upon exiting. Allows reading and writing to and from the same filename. The file will not be moved to destination in case of an exception. Parameters ---------- filepath : string the file path to be opened fsync : bool whether to force write the file to disk *args : mixed Any valid arguments for :code:`open` **kwargs : mixed Any valid keyword arguments for :code:`open` """ # From: https://stackoverflow.com/a/29491523 import os fsync = kwargs.get('fsync', False) with tempfile(dir=os.path.dirname(os.path.abspath(filepath))) as tmppath: with open(tmppath, *args, **kwargs) as file: try: yield file finally: if fsync: file.flush() os.fsync(file.fileno()) os.rename(tmppath, filepath)
29,160
9,168
import configparser from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker config = configparser.ConfigParser() config.read('alembic.ini') connection_url = config['alembic']['sqlalchemy.url'] Engine = create_engine(connection_url, connect_args={'check_same_thread': False}) Session = sessionmaker(bind=Engine)
335
101
# Copyright 2014-2016 ARM Limited # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # #pylint: disable=E1101,W0201 import os import re from wlauto import Workload, Parameter, Executable from wlauto.common.resources import File from wlauto.exceptions import ConfigError regex_map = { "Richards": (re.compile(r'Richards: (\d+.*)')), "DeltaBlue": (re.compile(r'DeltaBlue: (\d+.*)')), "Crypto": (re.compile(r'Crypto: (\d+.*)')), "RayTrace": (re.compile(r'RayTrace: (\d+.*)')), "EarleyBoyer": (re.compile(r'EarleyBoyer: (\d+.*)')), "RegExp": (re.compile(r'RegExp: (\d+.*)')), "Splay": (re.compile(r'Splay: (\d+.*)')), "SplayLatency": (re.compile(r'SplayLatency: (\d+.*)')), "NavierStokes": (re.compile(r'NavierStokes: (\d+.*)')), "PdfJS": (re.compile(r'PdfJS: (\d+.*)')), "Mandreel": (re.compile(r'Mandreel: (\d+.*)')), "MandreelLatency": (re.compile(r'MandreelLatency: (\d+.*)')), "Gameboy": (re.compile(r'Gameboy: (\d+.*)')), "CodeLoad": (re.compile(r'CodeLoad: (\d+.*)')), "Box2D": (re.compile(r'Box2D: (\d+.*)')), "zlib": (re.compile(r'zlib: (\d+.*)')), "Score": (re.compile(r'Score .*: (\d+.*)')) } class Octaned8(Workload): name = 'octaned8' description = """ Runs the Octane d8 benchmark. This workload runs d8 binaries built from source and placed in the dependencies folder along with test assets from https://github.com/chromium/octane which also need to be placed in an assets folder within the dependencies folder. Original source from:: https://github.com/v8/v8/wiki/D8%20on%20Android """ parameters = [ Parameter('run_timeout', kind=int, default=180, description='Timeout, in seconds, for the test execution.'), ] supported_platforms = ['android'] executables = ['d8', 'natives_blob.bin', 'snapshot_blob.bin'] def initialize(self, context): # pylint: disable=no-self-use assets_dir = self.device.path.join(self.device.working_directory, 'assets') self.device.execute('mkdir -p {}'.format(assets_dir)) assets_tar = 'octaned8-assets.tar' fpath = context.resolver.get(File(self, assets_tar)) self.device.push_file(fpath, assets_dir, timeout=300) self.command = 'cd {}; {} busybox tar -x -f {}'.format(assets_dir, self.device.busybox, assets_tar) self.output = self.device.execute(self.command, timeout=self.run_timeout) for f in self.executables: binFile = context.resolver.get(Executable(self, self.device.abi, f)) self.device_exe = self.device.install(binFile) def setup(self, context): self.logger.info('Copying d8 binaries to device') assets_dir = self.device.path.join(self.device.working_directory, 'assets') device_file = self.device.path.join(self.device.working_directory, 'octaned8.output') self.command = 'cd {}; {}/d8 ./run.js >> {} 2>&1'.format(assets_dir, self.device.binaries_directory, device_file) def run(self, context): self.logger.info('Starting d8 tests') self.output = self.device.execute(self.command, timeout=self.run_timeout) def update_result(self, context): host_file = os.path.join(context.output_directory, 'octaned8.output') device_file = self.device.path.join(self.device.working_directory, 'octaned8.output') self.device.pull_file(device_file, host_file) with open(os.path.join(host_file)) as octaned8_file: for line in octaned8_file: for label, regex in regex_map.iteritems(): match = regex.search(line) if match: context.result.add_metric(label, float(match.group(1))) self.device.execute('rm {}'.format(device_file)) def finalize(self, context): for f in self.executables: self.device.uninstall_executable(f) self.device.execute('rm {}'.format(self.device.path.join(self.device.working_directory, f))) assets_dir = self.device.path.join(self.device.working_directory, 'assets') self.device.execute('rm -rf {}'.format(assets_dir))
4,682
1,606
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'variable_design.ui' # # Created by: PyQt5 UI code generator 5.11.3 # # WARNING! All changes made in this file will be lost! from PyQt5 import QtCore, QtGui, QtWidgets class Ui_MainWindow(object): def setupUi(self, MainWindow): MainWindow.setObjectName("MainWindow") MainWindow.setWindowModality(QtCore.Qt.WindowModal) MainWindow.resize(929, 461) MainWindow.setMinimumSize(QtCore.QSize(929, 461)) MainWindow.setMaximumSize(QtCore.QSize(929, 461)) MainWindow.setStyleSheet("background-color: rgb(70, 70, 70);") self.centralwidget = QtWidgets.QWidget(MainWindow) self.centralwidget.setObjectName("centralwidget") self.x_variable_lineEdit = QtWidgets.QLineEdit(self.centralwidget) self.x_variable_lineEdit.setGeometry(QtCore.QRect(10, 40, 651, 51)) font = QtGui.QFont() font.setFamily("Courier") font.setPointSize(10) self.x_variable_lineEdit.setFont(font) self.x_variable_lineEdit.setStyleSheet("color: rgb(255, 255, 255);\n" "background-color: rgb(53, 53, 53);\n" "\n" "") self.x_variable_lineEdit.setObjectName("x_variable_lineEdit") self.newVariableButton = QtWidgets.QPushButton(self.centralwidget) self.newVariableButton.setGeometry(QtCore.QRect(10, 380, 35, 35)) font = QtGui.QFont() font.setPointSize(8) self.newVariableButton.setFont(font) self.newVariableButton.setStyleSheet("color: rgb(255, 255, 255);") self.newVariableButton.setText("") self.newVariableButton.setObjectName("newVariableButton") self.variablesList_treeWidget = QtWidgets.QTreeWidget(self.centralwidget) self.variablesList_treeWidget.setGeometry(QtCore.QRect(10, 190, 651, 181)) self.variablesList_treeWidget.setStyleSheet("color: rgb(1, 1, 1);\n" "background-color: rgb(120, 120, 120);") self.variablesList_treeWidget.setObjectName("variablesList_treeWidget") self.variablesList_treeWidget.header().setDefaultSectionSize(120) self.RemoveVar_Button = QtWidgets.QPushButton(self.centralwidget) self.RemoveVar_Button.setGeometry(QtCore.QRect(50, 380, 35, 35)) self.RemoveVar_Button.setStyleSheet("color: rgb(255, 255, 255);") self.RemoveVar_Button.setText("") self.RemoveVar_Button.setObjectName("RemoveVar_Button") self.label = QtWidgets.QLabel(self.centralwidget) self.label.setGeometry(QtCore.QRect(10, 10, 91, 16)) font = QtGui.QFont() font.setPointSize(8) self.label.setFont(font) self.label.setStyleSheet("color: rgb(255, 255, 255);") self.label.setObjectName("label") self.x_expr_lineEdit = QtWidgets.QLineEdit(self.centralwidget) self.x_expr_lineEdit.setGeometry(QtCore.QRect(110, 10, 551, 20)) font = QtGui.QFont() font.setFamily("Courier") self.x_expr_lineEdit.setFont(font) self.x_expr_lineEdit.setStyleSheet("color: rgb(204, 204, 204);") self.x_expr_lineEdit.setReadOnly(True) self.x_expr_lineEdit.setObjectName("x_expr_lineEdit") self.y_variable_lineEdit = QtWidgets.QLineEdit(self.centralwidget) self.y_variable_lineEdit.setGeometry(QtCore.QRect(10, 130, 651, 51)) font = QtGui.QFont() font.setFamily("Courier") font.setPointSize(10) self.y_variable_lineEdit.setFont(font) self.y_variable_lineEdit.setStyleSheet("color: rgb(255, 255, 255);\n" "background-color: rgb(53, 53, 53);\n" "\n" "") self.y_variable_lineEdit.setObjectName("y_variable_lineEdit") self.label_2 = QtWidgets.QLabel(self.centralwidget) self.label_2.setGeometry(QtCore.QRect(10, 100, 91, 16)) font = QtGui.QFont() font.setPointSize(8) self.label_2.setFont(font) self.label_2.setStyleSheet("color: rgb(255, 255, 255);") self.label_2.setObjectName("label_2") self.y_expr_lineEdit = QtWidgets.QLineEdit(self.centralwidget) self.y_expr_lineEdit.setGeometry(QtCore.QRect(110, 100, 551, 20)) font = QtGui.QFont() font.setFamily("Courier") self.y_expr_lineEdit.setFont(font) self.y_expr_lineEdit.setStyleSheet("color: rgb(204, 204, 204);") self.y_expr_lineEdit.setReadOnly(True) self.y_expr_lineEdit.setObjectName("y_expr_lineEdit") self.apply_pushButton = QtWidgets.QPushButton(self.centralwidget) self.apply_pushButton.setGeometry(QtCore.QRect(500, 380, 75, 35)) self.apply_pushButton.setStyleSheet("color: rgb(255, 255, 255);") self.apply_pushButton.setObjectName("apply_pushButton") self.cancel_pushButton = QtWidgets.QPushButton(self.centralwidget) self.cancel_pushButton.setGeometry(QtCore.QRect(580, 380, 75, 35)) self.cancel_pushButton.setStyleSheet("color: rgb(255, 255, 255);") self.cancel_pushButton.setObjectName("cancel_pushButton") self.ok_btn = QtWidgets.QPushButton(self.centralwidget) self.ok_btn.setGeometry(QtCore.QRect(420, 380, 75, 35)) self.ok_btn.setStyleSheet("color: rgb(255, 255, 255);") self.ok_btn.setObjectName("ok_btn") self.groupBox = QtWidgets.QGroupBox(self.centralwidget) self.groupBox.setGeometry(QtCore.QRect(670, 0, 251, 181)) self.groupBox.setStyleSheet("color: rgb(200, 200, 200);") self.groupBox.setObjectName("groupBox") self.grab_gbl_btn = QtWidgets.QPushButton(self.groupBox) self.grab_gbl_btn.setGeometry(QtCore.QRect(90, 150, 75, 23)) self.grab_gbl_btn.setObjectName("grab_gbl_btn") self.rm_gbl_btn = QtWidgets.QPushButton(self.groupBox) self.rm_gbl_btn.setGeometry(QtCore.QRect(170, 150, 71, 23)) self.rm_gbl_btn.setObjectName("rm_gbl_btn") self.global_treeview = QtWidgets.QTreeView(self.groupBox) self.global_treeview.setGeometry(QtCore.QRect(10, 20, 231, 121)) self.global_treeview.setObjectName("global_treeview") self.set_gbl_btn = QtWidgets.QPushButton(self.groupBox) self.set_gbl_btn.setGeometry(QtCore.QRect(10, 150, 75, 23)) self.set_gbl_btn.setToolTip("") self.set_gbl_btn.setObjectName("set_gbl_btn") self.groupBox_2 = QtWidgets.QGroupBox(self.centralwidget) self.groupBox_2.setGeometry(QtCore.QRect(670, 190, 251, 221)) self.groupBox_2.setStyleSheet("color: rgb(200, 200, 200);") self.groupBox_2.setObjectName("groupBox_2") self.search_alf_le = QtWidgets.QLineEdit(self.groupBox_2) self.search_alf_le.setGeometry(QtCore.QRect(10, 190, 110, 20)) self.search_alf_le.setStyleSheet("color: rgb(200, 200, 200);") self.search_alf_le.setObjectName("search_alf_le") self.search_num_le = QtWidgets.QLineEdit(self.groupBox_2) self.search_num_le.setGeometry(QtCore.QRect(130, 190, 110, 20)) self.search_num_le.setStyleSheet("color: rgb(200, 200, 200);") self.search_num_le.setObjectName("search_num_le") self.num_view = QtWidgets.QListView(self.groupBox_2) self.num_view.setGeometry(QtCore.QRect(130, 60, 110, 121)) self.num_view.setStyleSheet("color: rgb(150, 150, 150);\n" "background-color: rgb(60, 60, 60);\n" "") self.num_view.setObjectName("num_view") self.alf_view = QtWidgets.QListView(self.groupBox_2) self.alf_view.setGeometry(QtCore.QRect(10, 60, 110, 121)) self.alf_view.setStyleSheet("color: rgb(150, 150, 150);\n" "background-color: rgb(60, 60, 60);\n" "") self.alf_view.setObjectName("alf_view") self.alf_label_2 = QtWidgets.QLabel(self.groupBox_2) self.alf_label_2.setGeometry(QtCore.QRect(140, 40, 51, 16)) font = QtGui.QFont() font.setPointSize(8) self.alf_label_2.setFont(font) self.alf_label_2.setStyleSheet("color: rgb(255, 255, 255);") self.alf_label_2.setObjectName("alf_label_2") self.alf_label = QtWidgets.QLabel(self.groupBox_2) self.alf_label.setGeometry(QtCore.QRect(10, 40, 51, 16)) font = QtGui.QFont() font.setPointSize(8) self.alf_label.setFont(font) self.alf_label.setStyleSheet("color: rgb(255, 255, 255);") self.alf_label.setObjectName("alf_label") self.add2Y_checkBox = QtWidgets.QCheckBox(self.groupBox_2) self.add2Y_checkBox.setGeometry(QtCore.QRect(90, 20, 61, 17)) self.add2Y_checkBox.setStyleSheet("color: rgb(255, 255, 255);") self.add2Y_checkBox.setObjectName("add2Y_checkBox") self.add2XcheckBox = QtWidgets.QCheckBox(self.groupBox_2) self.add2XcheckBox.setGeometry(QtCore.QRect(10, 20, 70, 17)) self.add2XcheckBox.setStyleSheet("color: rgb(255, 255, 255);") self.add2XcheckBox.setObjectName("add2XcheckBox") MainWindow.setCentralWidget(self.centralwidget) self.menubar = QtWidgets.QMenuBar(MainWindow) self.menubar.setGeometry(QtCore.QRect(0, 0, 929, 21)) self.menubar.setObjectName("menubar") MainWindow.setMenuBar(self.menubar) self.statusbar = QtWidgets.QStatusBar(MainWindow) self.statusbar.setObjectName("statusbar") MainWindow.setStatusBar(self.statusbar) self.retranslateUi(MainWindow) QtCore.QMetaObject.connectSlotsByName(MainWindow) def retranslateUi(self, MainWindow): _translate = QtCore.QCoreApplication.translate MainWindow.setWindowTitle(_translate("MainWindow", "Variables")) self.variablesList_treeWidget.headerItem().setText(0, _translate("MainWindow", "Variable Name")) self.variablesList_treeWidget.headerItem().setText(1, _translate("MainWindow", "Label")) self.variablesList_treeWidget.headerItem().setText(2, _translate("MainWindow", "x")) self.variablesList_treeWidget.headerItem().setText(3, _translate("MainWindow", "y")) self.variablesList_treeWidget.headerItem().setText(4, _translate("MainWindow", "Restart")) self.variablesList_treeWidget.headerItem().setText(5, _translate("MainWindow", "csv")) self.variablesList_treeWidget.headerItem().setText(6, _translate("MainWindow", "Plot")) self.label.setText(_translate("MainWindow", "Current xVariable")) self.label_2.setText(_translate("MainWindow", "Current yVariable")) self.apply_pushButton.setText(_translate("MainWindow", "Apply")) self.cancel_pushButton.setText(_translate("MainWindow", "Cancel")) self.ok_btn.setText(_translate("MainWindow", "Ok")) self.groupBox.setTitle(_translate("MainWindow", "Global expressions")) self.grab_gbl_btn.setText(_translate("MainWindow", "grab")) self.rm_gbl_btn.setText(_translate("MainWindow", "remove")) self.set_gbl_btn.setText(_translate("MainWindow", "set")) self.groupBox_2.setTitle(_translate("MainWindow", "Plot_alf | Plot_num")) self.search_alf_le.setPlaceholderText(_translate("MainWindow", "find plot_alfa")) self.search_num_le.setPlaceholderText(_translate("MainWindow", "find plot_num")) self.alf_label_2.setText(_translate("MainWindow", "plot_num")) self.alf_label.setText(_translate("MainWindow", "plot_alf")) self.add2Y_checkBox.setText(_translate("MainWindow", "Add to Y")) self.add2XcheckBox.setText(_translate("MainWindow", "Add to X")) if __name__ == "__main__": import sys app = QtWidgets.QApplication(sys.argv) MainWindow = QtWidgets.QMainWindow() ui = Ui_MainWindow() ui.setupUi(MainWindow) MainWindow.show() sys.exit(app.exec_())
11,872
4,550
''' n-gram은 이와 같은 확률적 언어 모델의 대표적인 것으로서, n개 단어의 연쇄를 확률적으로 표현해 두면 실제로 발성된 문장의 기록을 계산할 수 있다. ''' # Step 1 : Bag of Words from nltk.corpus import reuters from collections import Counter, defaultdict counts = Counter(reuters.words()) total_count = len(reuters.words()) # 공통적으로 가장 많이 나타나는 20개의 단어 print(counts.most_common(n=20)) # 빈도 비율 계산 for word in counts: counts[word] /= float(total_count) # 빈도 비율의 총합 계산 print(sum(counts.values())) import random # 100개의 단어 생성 text = [] for _ in range(100) : r = random.random() accumulator = 0 for word, freq in counts.items(): accumulator += freq if accumulator >= r: text.append(word) break print(" ".join(text)) # 텍스트의 확률 계산 from operator import mul from functools import reduce print(reduce(mul, [counts[w] for w in text], 1.0)) # Step 2 : Bi-gram & Tri-gram from nltk import bigrams, trigrams first_sentence = reuters.sents()[0] print(first_sentence) ### Bi-gram 결과 print("\n----- Bi-gram 결과 확 확인인 -----") print(list(bigrams(first_sentence))) ### Bi-gram 결과 print("\n----- 패딩된 Bi-gram 결과 확인 -----") print((list(bigrams(first_sentence, pad_left=True, pad_right=True)))) ### Bi-gram 결과 print("\n----- Tri-gram 결과 확인 -----") print(list(trigrams(first_sentence))) ### Bi-gram 결과 print("\n----- 패딩된 Tri-gram 결과 확인 -----") print((list(trigrams(first_sentence, pad_left=True, pad_right=True)))) ### Reuters 데이터(말뭉치)를 이용한 Tri-gram 모델 생성 print("\n----- Tri-gram 모델 생성 -----") model = defaultdict(lambda : defaultdict(lambda : 0)) for sentence in reuters.sents() : print("\n문장 : ", sentence) for w1, w2, w3 in trigrams(sentence, pad_right=True, pad_left=True): model[(w1,w2)][w3] += 1 # 'what the' 다음에 'economists'가 나오는 것이 2개 존재 print(model['what','the']['economists']) for w1_w2 in model: total_count = float(sum(model[w1_w2].values())) for w3 in model[w1_w2]: model[w1_w2][w3] /= total_count print(model['what', 'the']['economists']) ### Language Model을 이용해 텍스트 생성하기 print("\n----- 언어 모델을 이용해 텍스트 생성하기 -----") import random text = [None, None] prob = 1.0 sentence_finished = False while not sentence_finished: r = random.random() accumulator = .0 for word in model[tuple(text[-2:])].keys(): accumulator += model[tuple(text[-2:])][word] if accumulator >= r: prob *= model[tuple(text[-2:])][word] text.append(word) print("aa : " ,text) break if text[-2:] == [None, None]: sentence_finished = True print("텍스트의 확률 : ", prob) print(" ".join([t for t in text if t]))
2,597
1,288
#!/usr/bin/python import numpy class Color(object): lower_bound = 0 upper_bound = 255 def __init__(self, r, g, b, a): self.r = r self.g = g self.b = b self.a = a def __str__(self): return '(%s, %s, %s, %s)' % (self.r, self.g, self.b, self.a) def composite_pixels(src, dest): new_color = Color(0,0,0,0) new_color.r = src.a * src.r * dest.a + src.r * (1 - dest.a) + dest.r * dest.a * (1 - src.a) new_color.g = src.a * src.g * dest.a + src.g * (1 - dest.a) + dest.g * dest.a * (1 - src.a) new_color.b = src.a * src.b * dest.a + src.b * (1 - dest.a) + dest.b * dest.a * (1 - src.a) new_color.a = src.a * dest.a + src.a * (1 - dest.a) + dest.a * (1 - src.a) return new_color def simple_composite_pixels(src, dest): return Color( (src.r * src.a)/255 + ((dest.r * dest.a) * (255 - src.a))/255**2, (src.g * src.a)/255 + ((dest.g * dest.a) * (255 - src.a))/255**2, (src.b * src.a)/255 + ((dest.b * dest.a) * (255 - src.a))/255**2, src.a + dest.a - (src.a * dest.a)/255, ) def overlay_pixel(src, dest): a = numpy.array([ (src[0] * src[3]) / 255 + ((dest[0] * dest[3]) * (255 - src[3])) / 255 ** 2, (src[1] * src[3]) / 255 + ((dest[1] * dest[3]) * (255 - src[3])) / 255 ** 2, (src[2] * src[3]) / 255 + ((dest[2] * dest[3]) * (255 - src[3])) / 255 ** 2, src[3] + dest[3] - (src[3] * dest[3]) / 255, ], dtype=numpy.uint8) print [ (src[0] * src[3]) / 255 + ((dest[0] * dest[3]) * (255 - src[3])) / 255 ** 2, (src[1] * src[3]) / 255 + ((dest[1] * dest[3]) * (255 - src[3])) / 255 ** 2, (src[2] * src[3]) / 255 + ((dest[2] * dest[3]) * (255 - src[3])) / 255 ** 2, src[3] + dest[3] - (src[3] * dest[3]) / 255, ] print src, dest, a return a if __name__ == '__main__': a = (120,120,120,255) b = (117,176,73,255) c = (134,96,67,255) d = (38,92,255,100) e = (0,0,0,0) print overlay_pixel(a,e) print (a[0] * a[3]) / 255 + ((e[0] * e[3]) * (255 - a[3])) / 255 ** 2
1,977
1,071
import os import requests import urllib3.util.retry import logging import subprocess import re import pprint from datetime import datetime, timedelta import pygit2 from github import Github from github.GithubException import UnknownObjectException # Create logger with logging level set to all LOGGER = logging.getLogger(__name__) logging.basicConfig(level=logging.INFO) def get_project_and_feedstock_repos(github_client, repo_name): """ Get the project and feedstock repos from the repo name. Parameters ---------- github_client : github.MainClass.Github The Github client. repo_name : str The repo name. Returns ------- project_repo : github.Repository.Repository The project repository. feedstock_repo : github.Repository.Repository The feedstock repository. feedstock_repo_name : str The feedstock repository name. """ feedstock_repo_name = repo_name + "-feedstock" project_repo = github_client.get_repo(repo_name) # Get feedstock repository try: feedstock_repo = github_client.get_repo(feedstock_repo_name) # If feedstock repo does not exist, log an error and exit except UnknownObjectException: LOGGER.error( "repository_dispatch event: feedstock repository of '%s' not found" % repo_name) return None, None, None return project_repo, feedstock_repo, feedstock_repo_name def get_project_version(repo_dir, VERSION_PEP440=re.compile(r'(?P<major>\d+)\.(?P<minor>\d+)\.(?P<patch>\d+)(\.(?P<release>[a-z]+)(?P<dev>\d+))?')): """ Get the project version from the version file in a given repository. Parameters ---------- repo_dir : str The path to the repo directory. VERSION_PEP440 : re.compile The regex to match the version. Returns ------- str The project version. """ # Get version from version file in project repo with open(os.path.join(repo_dir, "version"), 'r') as fp: version = fp.read().rstrip("\n") # Match version from file to pep440 and retrieve groups match = VERSION_PEP440.match(version) if match: major = match.group("major") minor = match.group("minor") patch = match.group("patch") release = match.group("release") dev = match.group("dev") if release: version = f"{major}.{minor}.{patch}.{release}{dev}" else: version = f"{major}.{minor}.{patch}" LOGGER.info("version: %s", version) LOGGER.info("major: %s", major) LOGGER.info("minor: %s", minor) LOGGER.info("patch: %s", patch) LOGGER.info("release: %s", release) LOGGER.info("dev: %s", dev) return version else: LOGGER.error( "repository_dispatch event: could not parse version") return None def get_commit_tags(repo, commit_hash, supported_tags=["ci", "rerender"]): """ Get the tags of a commit. Parameters ---------- repo : github.Repository.Repository The repository. commit_hash : str The commit hash. supported_tags : list[str] List of tags that are to be searched for. Returns ------- tags : dict[str, bool] Dictionary of tags found, with supported tags as keys and the value denoting wether they were found. commit_message : str New commit message cleaned of the tag in brackets, but with the tag in front instead. """ # Get commit from its sha commit = repo.get_commit(sha=commit_hash) message = commit.raw_data["commit"]["message"] # Extract commit tag if there is one tag = re.search(r'\[(.*?)\]', message) if tag: tag = tag.group(1) LOGGER.info("tag: %s", tag) if tag.lower() in supported_tags: # Remove tag from message, and add it in front commit_message = message.replace(f'[{tag}]', "") # Clean excess spaces commit_message = re.sub(r'\s+', " ", commit_message).strip() # Add tag in front of commit message commit_message = "%s: %s" % (tag.upper(), commit_message) # Return True for the tag that was found return {possible_tag: tag.lower() == possible_tag for possible_tag in supported_tags}, commit_message else: # Quit if the tag is not in the list of supported ones LOGGER.info( "no supported tag detected (was '%s', supported are %s" % (tag, supported_tags) ) return {possible_tag: False for possible_tag in supported_tags}, None else: # Quit if there is not tag LOGGER.info("no tag detected") return {possible_tag: False for possible_tag in supported_tags}, None def was_branch_last_commit_recent(repo, branch_name, time_treshold=timedelta(hours=24)): """ Check if the last commit of a branch is recent. Parameters ---------- repo : github.Repository.Repository The repository. branch_name : str The branch name. time_treshold : datetime.timedelta The time threshold under which the last commit will be considered recent. Returns ------- bool True if the last commit is recent, False otherwise. """ # Get info of latest commit for given branch branch = repo.get_branch(branch_name) date_string = branch.commit.raw_data["commit"]["author"]["date"] last_commit_time = datetime.strptime(date_string, "%Y-%m-%dT%H:%M:%SZ") # Trigger release if last commit time was less than some time ago if last_commit_time > datetime.now() - time_treshold: LOGGER.info( "since is within specified time, nightly release will follow") return True return False def push_all_to_github(repo, branch_name, directory, commit_message): """ Push all files in a directory to a github repository. Parameters ---------- repo : github.Repository.Repository The repository. branch_name : str The branch name. directory : str The directory to push. commit_message : str The commit message. """ # Add all files subprocess.run(["git", "add", "."], cwd=directory) # Commit with proper commit message subprocess.run(["git", "commit", "-m", commit_message], cwd=directory) # Get url to push to repo_auth_url = "https://%s@github.com/%s.git" % (os.environ["GH_TOKEN"], repo) # Push changes and tags subprocess.run(["git", "push", "--all", "-f", repo_auth_url], cwd=directory) subprocess.run(["git", "push", repo_auth_url, branch_name, "--tags"], cwd=directory) def create_api_sessions(github_token): """Create API sessions for GitHub. Parameters ---------- github_token : str The GitHub access token. Returns ------- session : requests.Session A `requests` session w/ the beta `check_run` API configured. gh : github.MainClass.Github A `Github` object from the PyGithub package. """ # based on # https://alexwlchan.net/2019/03/ # creating-a-github-action-to-auto-merge-pull-requests/ # with lots of edits sess = requests.Session() sess.headers = { "Accept": "; ".join([ "application/vnd.github.v3+json", # special beta api for check_suites endpoint "application/vnd.github.antiope-preview+json", ]), "Authorization": f"token {github_token}", "User-Agent": f"GitHub Actions script in {__file__}" } def raise_for_status(resp, *args, **kwargs): try: resp.raise_for_status() except Exception as e: print('ERROR:', resp.text) raise e sess.hooks["response"].append(raise_for_status) # build a github object too gh = Github( github_token, retry=urllib3.util.retry.Retry(total=10, backoff_factor=0.1)) return sess, gh def clone_repo(clone_url, clone_path, branch, auth_token): # Use pygit2 to clone the repo to disk # if using github app pem key token, use x-access-token like below # if you were using a personal access token, use auth_method = 'x-oauth-basic' AND reverse the auth_method and token parameters auth_method = 'x-access-token' callbacks = pygit2.RemoteCallbacks(pygit2.UserPass(auth_method, auth_token)) pygit2_repo = pygit2.clone_repository(clone_url, clone_path, callbacks=callbacks) pygit2_branch = pygit2_repo.branches['origin/' + branch] pygit2_ref = pygit2_repo.lookup_reference(pygit2_branch.name) pygit2_repo.checkout(pygit2_ref) # Checkout correct branch subprocess.run(["git", "checkout", branch], cwd=clone_path) return pygit2_repo, pygit2_ref def get_var_values(var_retrieve, root=''): ret = {} for var, file, regex in var_retrieve: with open(os.path.join(root, file), 'r') as f: s = f.read() m = regex.search(s) v = m.group(1) ret[var] = v return ret def update_var_values(var_retrieved, version_tag, git_rev=None, root=''): ret = {} git_rev = version_tag if git_rev is None else git_rev for k, v in var_retrieved.items(): if k == 'build': if var_retrieved['version'] == version_tag: # NOTE(Geoffrey): we are only bumping build number at the moment. v = int(v) + 1 else: v = 0 elif k == 'git_rev': v = git_rev elif k == 'version': v = version_tag ret[k] = v return ret def substitute_vars_in_file(vars_substitute, directory): # Substitute variables in files for file, regex, subst, val in vars_substitute: path = os.path.join(directory, file) # Read file with open(path, "r") as f: s = f.read() # Substitute s = regex.sub(subst.replace("{}", str(val)), s) # Write file with open(path, "w") as f: f.write(s)
10,139
3,053
from django.core.validators import MinLengthValidator from django.contrib.auth.models import User from django.contrib.auth.validators import UnicodeUsernameValidator from django.db import models from django.db.models.signals import pre_save from django.dispatch import receiver from django.utils.encoding import iri_to_uri from mongo.objectid import ObjectId from mur.commonmark import commonmark def _objectid(): return str(ObjectId()) class Post(models.Model): path_validators = [MinLengthValidator(6), UnicodeUsernameValidator()] objectid = models.CharField(max_length=24, default=_objectid, editable=False, unique=True) user = models.ForeignKey(User, on_delete=models.PROTECT, related_name='posts') path = models.CharField(max_length=127, validators=path_validators) contents = models.TextField() contents_html = models.TextField(default='', editable=False) created = models.DateTimeField(auto_now_add=True) updated = models.DateTimeField(auto_now=True) def get_absolute_url(self): # The string returned from get_absolute_url() must contain only ASCII characters. return iri_to_uri(f'/{self.user}/{self.path}/') def __str__(self): return f'Post({self.user}/{self.path})' class Meta: unique_together = ('user', 'path') @receiver(pre_save, sender=Post) def update_html(sender, instance, update_fields, **kwargs): if update_fields and 'contents' not in update_fields: return instance.contents_html = commonmark(instance.contents)
1,538
450
# Import external modules. from google.appengine.ext import ndb import logging # Import local modules. from configAutocomplete import const as conf from constants import Constants class Survey(ndb.Model): surveyId = ndb.StringProperty() # Primary key title = ndb.StringProperty() introduction = ndb.StringProperty() creator = ndb.StringProperty() allowEdit = ndb.BooleanProperty() freezeUserInput = ndb.BooleanProperty( default=False ) hideReasons = ndb.BooleanProperty( default=False ) # Experimental option questionIds = ndb.StringProperty( repeated=True ) # Ordered
611
180
from flask import Flask from app.views import app
50
14
from wavenetlike.analyzers import DatasetAnalyzer from wavenetlike.datasetid import TorchAudioDataSetId def analyzer_example(): dataset = TorchAudioDataSetId("SPEECHCOMMANDS") data_analyzer = DatasetAnalyzer(dataset) data_analyzer.analyze_dataset() analysis_res = data_analyzer.get_analysis_result() print(analysis_res) if __name__ == "__main__": analyzer_example()
394
132
def get_ts_struct(ts): y=ts&0x3f ts=ts>>6 m=ts&0xf ts=ts>>4 d=ts&0x1f ts=ts>>5 hh=ts&0x1f ts=ts>>5 mm=ts&0x3f ts=ts>>6 ss=ts&0x3f ts=ts>>6 wd=ts&0x8 ts=ts>>3 yd=ts&0x1ff ts=ts>>9 ms=ts&0x3ff ts=ts>>10 pid=ts&0x3ff return y,m,d,hh,mm,ss,wd,yd,ms,pid
271
179