text
string
size
int64
token_count
int64
from tkinter import simpledialog from porcupine import actions, get_tab_manager, tabs def gotoline(): tab = get_tab_manager().select() # simpledialog isn't ttk yet, but it's not a huge problem imo lineno = simpledialog.askinteger( "Go to Line", "Type a line number and press Enter:") if lineno is not None: # not cancelled # there's no need to do a bounds check because tk ignores out-of-bounds # text indexes column = tab.textwidget.index('insert').split('.')[1] tab.textwidget.mark_set('insert', '%d.%s' % (lineno, column)) tab.textwidget.see('insert') tab.on_focus() def setup(): actions.add_command("Edit/Go to Line", gotoline, '<Control-l>', tabtypes=[tabs.FileTab])
778
239
import os import pickle import sys import warnings from collections import OrderedDict import biosppy.signals.tools as st import numpy as np import wfdb from biosppy.signals.ecg import correct_rpeaks, hamilton_segmenter from hrv.classical import frequency_domain, time_domain from scipy.signal import medfilt from tqdm import tqdm warnings.filterwarnings(action="ignore") base_dir = "dataset" fs = 100 # ECG sample frequency hr_min = 20 hr_max = 300 def feature_extraction(recording, signal, labels): data = [] for i in tqdm(range(len(labels)), desc=recording, file=sys.stdout): segment = signal[i * fs * 60:(i + 1) * fs * 60] segment, _, _ = st.filter_signal(segment, ftype='FIR', band='bandpass', order=int(0.3 * fs), frequency=[3, 45], sampling_rate=fs) # Finding R peaks rpeaks, = hamilton_segmenter(segment, sampling_rate=fs) rpeaks, = correct_rpeaks(segment, rpeaks, sampling_rate=fs, tol=0.1) # Extracting feature label = 0 if labels[i] == "N" else 1 if 40 <= len(rpeaks) <= 200: # Remove abnormal R peaks rri_tm, rri = rpeaks[1:] / float(fs), np.diff(rpeaks, axis=-1) / float(fs) rri = medfilt(rri, kernel_size=3) edr_tm, edr = rpeaks / float(fs), segment[rpeaks] # Remove physiologically impossible HR signal if np.all(np.logical_and(60 / rri >= hr_min, 60 / rri <= hr_max)): rri_time_features, rri_frequency_features = time_domain(rri * 1000), frequency_domain(rri, rri_tm) edr_frequency_features = frequency_domain(edr, edr_tm) # 6 + 6 + 6 + 1 = 19 data.append([ rri_time_features["rmssd"], rri_time_features["sdnn"], rri_time_features["nn50"], rri_time_features["pnn50"], rri_time_features["mrri"], rri_time_features["mhr"], rri_frequency_features["vlf"] / rri_frequency_features["total_power"], rri_frequency_features["lf"] / rri_frequency_features["total_power"], rri_frequency_features["hf"] / rri_frequency_features["total_power"], rri_frequency_features["lf_hf"], rri_frequency_features["lfnu"], rri_frequency_features["hfnu"], edr_frequency_features["vlf"] / edr_frequency_features["total_power"], edr_frequency_features["lf"] / edr_frequency_features["total_power"], edr_frequency_features["hf"] / edr_frequency_features["total_power"], edr_frequency_features["lf_hf"], edr_frequency_features["lfnu"], edr_frequency_features["hfnu"], label ]) else: data.append([np.nan] * 18 + [label]) else: data.append([np.nan] * 18 + [label]) data = np.array(data, dtype="float") return data if __name__ == "__main__": apnea_ecg = OrderedDict() # train data recordings = [ "a01", "a02", "a03", "a04", "a05", "a06", "a07", "a08", "a09", "a10", "a11", "a12", "a13", "a14", "a15", "a16", "a17", "a18", "a19", "a20", "b01", "b02", "b03", "b04", "b05", "c01", "c02", "c03", "c04", "c05", "c06", "c07", "c08", "c09", "c10" ] for recording in recordings: signal = wfdb.rdrecord(os.path.join(base_dir, recording), channels=[0]).p_signal[:, 0] labels = wfdb.rdann(os.path.join(base_dir, recording), extension="apn").symbol apnea_ecg[recording] = feature_extraction(recording, signal, labels) print() # test data recordings = [ "x01", "x02", "x03", "x04", "x05", "x06", "x07", "x08", "x09", "x10", "x11", "x12", "x13", "x14", "x15", "x16", "x17", "x18", "x19", "x20", "x21", "x22", "x23", "x24", "x25", "x26", "x27", "x28", "x29", "x30", "x31", "x32", "x33", "x34", "x35" ] answers = {} filename = os.path.join(base_dir, "event-2-answers") with open(filename, "r") as f: for answer in f.read().split("\n\n"): answers[answer[:3]] = list("".join(answer.split()[2::2])) for recording in recordings: signal = wfdb.rdrecord(os.path.join(base_dir, recording), channels=[0]).p_signal[:, 0] labels = answers[recording] apnea_ecg[recording] = feature_extraction(recording, signal, labels) with open(os.path.join(base_dir, "apnea-ecg.pkl"), "wb") as f: pickle.dump(apnea_ecg, f, protocol=2) print("ok")
4,634
1,810
""" Subclass of tiddlywebplugins.openid2 to support tiddlyweb_secondary_user cookie. """ import urlparse from tiddlyweb.web.util import server_host_url, make_cookie from tiddlywebplugins.openid2 import Challenger as OpenID FRAGMENT_PREFIX = 'auth:OpenID:' class Challenger(OpenID): def __init__(self): self.name = __name__ def _domain_path(self, environ): return "." + environ['tiddlyweb.config']['server_host']['host'] def _success(self, environ, start_response, info): """ After successful validation of an openid generate and send a cookie with the value of that openid. If this is a normal auth scenario make the name of the cookie the normal 'tiddlyweb_user'. If this is auth addition, where a fragment of 'auth:OpenID' is set, then name the cookie 'tiddlyweb_secondary_user'. """ usersign = info.getDisplayIdentifier() if info.endpoint.canonicalID: usersign = info.endpoint.canonicalID # canonicolize usersign to tiddlyweb form if usersign.startswith('http'): usersign = usersign.split('://', 1)[1] usersign = usersign.rstrip('/') redirect = environ['tiddlyweb.query'].get( 'tiddlyweb_redirect', ['/'])[0] uri = urlparse.urljoin(server_host_url(environ), redirect) cookie_name = 'tiddlyweb_user' cookie_age = environ['tiddlyweb.config'].get('cookie_age', None) try: fragment = uri.rsplit('#', 1)[1] except (ValueError, IndexError): fragment = None secondary_cookie_name = 'tiddlyweb_secondary_user' secondary_cookie_age = None secondary_cookie_only = False if fragment: openid = fragment[len(FRAGMENT_PREFIX):] uri = uri.replace(FRAGMENT_PREFIX + openid, FRAGMENT_PREFIX + usersign) secondary_cookie_only = True secret = environ['tiddlyweb.config']['secret'] cookie_header_string = make_cookie(cookie_name, usersign, mac_key=secret, path=self._cookie_path(environ), expires=cookie_age) secondary_cookie_header_string = make_cookie( secondary_cookie_name, usersign, mac_key=secret, path=self._cookie_path(environ), expires=cookie_age, domain=self._domain_path(environ)) headers = [('Location', uri.encode('utf-8')), ('Content-Type', 'text/plain'), ('Set-Cookie', secondary_cookie_header_string)] if not secondary_cookie_only: headers.append(('Set-Cookie', cookie_header_string)) start_response('303 See Other', headers) return [uri] def _render_form(self, environ, start_response, openid='', message='', form=''): redirect = environ['tiddlyweb.query'].get( 'tiddlyweb_redirect', ['/'])[0] start_response('200 OK', [( 'Content-Type', 'text/html')]) environ['tiddlyweb.title'] = 'OpenID Login' return [""" <div id='content'> <div class='message'>%s</div> <pre> <form action="" method="POST"> OpenID: <input name="openid" size="60" value="%s"/> <input type="hidden" name="tiddlyweb_redirect" value="%s" /> <input type="hidden" id="csrf_token" name="csrf_token" /> <input type="submit" value="submit" /> </form> <script type="text/javascript" src="%s/bags/tiddlyspace/tiddlers/TiddlySpaceCSRF"></script> <script type="text/javascript"> var csrfToken = window.getCSRFToken(), el = null; if (csrfToken) { el = document.getElementById('csrf_token'); el.value = csrfToken; } </script> </pre> </div>""" % (message, openid, redirect, environ['tiddlyweb.config']['server_prefix'])]
3,904
1,178
from datetime import datetime import freezegun # Initialize freezegun to avoid freezegun being reinitialized which is expensive initialize_freezegun = freezegun.freeze_time(datetime(2021, 1, 1)) initialize_freezegun.start()
226
69
''' Created on Apr 2, 2015 @author: tmahrt ''' import os from os.path import join import math import copy from pyacoustics.morph.morph_utils import common from pyacoustics.morph.morph_utils import plot_morphed_data from pyacoustics.utilities import utils from pyacoustics.utilities import sequences from pyacoustics.signals import audio_scripts from pyacoustics.utilities import my_math def intensityMorph(fromWavFN, toWavFN, fromWavTGFN, toWavTGFN, tierName, numSteps, coreChunkSize, plotFlag): fromDataTupleList = common.getIntervals(fromWavTGFN, tierName) toDataTupleList = common.getIntervals(toWavTGFN, tierName) outputName = os.path.splitext(fromWavFN)[0] + "_int_" + tierName _intensityMorph(fromWavFN, toWavFN, fromDataTupleList, toDataTupleList, numSteps, coreChunkSize, plotFlag, outputName) def _intensityMorph(fromWavFN, toWavFN, fromDataTupleList, toDataTupleList, numSteps, coreChunkSize, plotFlag, outputName=None): if outputName is None: outputName = os.path.splitext(fromWavFN)[0] + "_int" outputDir = join(os.path.split(fromWavFN)[0], "output") utils.makeDir(outputDir) # Determine the multiplication values to be used in normalization # - this extracts one value per chunk expectedLength = 0 normFactorList = [] truncatedToList = [] chunkSizeList = [] fromDataList = [] fromParams = audio_scripts.getParams(fromWavFN) toParams = audio_scripts.getParams(toWavFN) for fromTuple, toTuple in zip(fromDataTupleList, toDataTupleList): fromStart, fromEnd = fromTuple[:2] toStart, toEnd = toTuple[:2] expectedLength += (fromEnd - fromStart) * fromParams[2] fromDataList.extend(fromSubWav.rawDataList) normFactorListTmp, a = getRelativeNormalizedFactors(fromSubWav, toSubWav, coreChunkSize) tmpChunkList = [tmpChunkSize for value, tmpChunkSize in normFactorListTmp] chunkSizeList.append(sum(tmpChunkList)) normFactorList.extend(normFactorListTmp) truncatedToList.extend(a) interpolatedResults = [] normFactorGen = [sequences.interp(1.0, factor[0], numSteps) for factor in normFactorList] tmpChunkSizeList = [factor[1] for factor in normFactorList] for i in xrange(numSteps): outputFN = "%s_s%d_%d_%d.wav" % (outputName, coreChunkSize, numSteps - 1, i) tmpNormFactorList = [next(normFactorGen[j]) for j in xrange(len(normFactorGen))] # Skip the first value (same as the input value) if i == 0: continue tmpInputList = zip(tmpNormFactorList, tmpChunkSizeList) normalizationTuple = expandNormalizationFactors(tmpInputList) expandedNormFactorList = normalizationTuple[0] # It happened once that the expanded factor list was off by one value # -- I could not determine why, so this is just a cheap hack if len(expandedNormFactorList) == (expectedLength - 1): expandedNormFactorList.append(expandedNormFactorList[-1]) # print("Diff: ", expectedLength, len(expandedNormFactorList)) assert(expectedLength == len(expandedNormFactorList)) newWavObj = copy.deepcopy(fromWavObj) newRawDataList = [] # Apply the normalization and reinsert the data back # into the original file offset = 0 for fromTuple, chunkSize in zip(fromDataTupleList, chunkSizeList): fromStart, fromEnd = fromTuple[:2] fromSubWav = fromWavObj.extractSubsegment(fromStart, fromEnd) assert(len(fromSubWav.rawDataList) == len(expandedNormFactorList[offset:offset + chunkSize])) tmpList = [fromSubWav.rawDataList, expandedNormFactorList[offset:offset + chunkSize]] subRawDataList = [value * normFactor for value, normFactor in utils.safeZip(tmpList, enforceLength=True)] newRawDataList.extend(subRawDataList) offset += chunkSize newWavObj = audio.WavObj(newRawDataList, fromWavObj.samplingRate) newWavObj.save(join(outputDir, outputFN)) interpolatedResults.append(newWavObj.rawDataList) plotFN = "%s_s%d_%d.png" % (outputFN, coreChunkSize, numSteps) if plotFlag: plotMorphedData.plotIntensity(fromDataList, truncatedToList, interpolatedResults, expandedNormFactorList, os.path.join(outputDir, plotFN)) def getNormalizationFactor(lst, refLst=None): ''' ''' # Get the source values that we will be normalizing lst = list(set(lst)) if 0 in lst: lst.pop(lst.index(0)) actMaxV = float(max(lst)) actMinV = float(min(lst)) # Get the reference values if refLst is None: refMaxV = 32767.0 refMinV = -32767.0 else: refLst = list(set(refLst)) if 0 in refLst: refLst.pop(refLst.index(0)) refMaxV = float(max(refLst)) refMinV = float(min(refLst)) actualFactor = min(refMaxV / actMaxV, abs(refMinV) / abs(actMinV)) # print("Normalization factor: ", actualFactor) return actualFactor def getRelativeNormalizedFactors(fromDataList, toDataList, chunkSize): ''' Determines the factors to be used to normalize sourceWav from targetWav This can be used to relatively normalize the source based on the target on an iterative basis (small chunks are normalized rather than the entire wav. ''' # Sample proportionately from the targetWav # - if the two lists are the same length, there is no change # - if /target/ is shorter, it will be lengthened with some repeated values # - if /target/ is longer, it will be shortened with some values dropped tmpIndexList = sequences.interp(0, len(toDataList) - 1, fromDataList) newTargetRawDataList = [toDataList[int(round(i))] for i in tmpIndexList] assert(len(fromDataList) == len(newTargetRawDataList)) fromGen = sequences.subsequenceGenerator(fromDataList, chunkSize, sequences.sampleMiddle, sequences.DO_SAMPLE_GATED) toGen = sequences.subsequenceGenerator(newTargetRawDataList, chunkSize, sequences.sampleMiddle, sequences.DO_SAMPLE_GATED) normFactorList = [] i = 0 for fromTuple, toTuple in zip(fromGen, toGen): fromDataChunk = fromTuple[0] toDataChunk = toTuple[0] distToNextControlPoint = fromTuple[2] normFactor = getNormalizationFactor(fromDataChunk, toDataChunk) normFactorList.append((normFactor, distToNextControlPoint)) # i += 1 # if i >= 38: # print("hello") # print(len(sourceWav.rawDataList), allChunks) # assert(len(sourceWav.rawDataList) == allChunks) return normFactorList, newTargetRawDataList def expandNormalizationFactors(normFactorList): ''' Expands the normFactorList from being chunk-based to sample-based E.g. A wav with 1000 samples may be represented by a factorList of 5 chunks (5 factor values). This function will expand that to 1000. ''' i = 0 normFactorsFull = [] controlPoints = [] while i < len(normFactorList) - 1: startVal, chunkSize = normFactorList[i] endVal = normFactorList[i + 1][0] normFactorsFull.extend(my_math.linspace(startVal, endVal, chunkSize)) controlPoints.append(startVal) controlPoints.extend(my_math.linspace(startVal, startVal, chunkSize - 1)) i += 1 # We have no more data, so just repeat the final norm factor at the tail # of the file value, finalChunkSize = normFactorList[i] controlPoints.append(value) controlPoints.extend(my_math.linspace(startVal, startVal, finalChunkSize - 1)) normFactorsFull.extend(my_math.linspace(value, value, finalChunkSize)) print('Norm factors full: %d' % len(normFactorsFull)) return normFactorsFull, controlPoints
9,047
2,726
import numpy as np class MetricsController: def __init__(self,metrics=None,algorithm="generic",metrics_path="/home/heitor/recsys/data/metrics",k=10): self.algorithm=algorithm self.metrics_path=metrics_path self.k=k if metrics==None: self.fload_metrics() else: self.metrics={} for i in metrics: self.metrics[i]={} def append_data(self,uid,data): c=0 for i in self.metrics.keys(): self.metrics[i][uid]=data[c] c+=1 def get_metrics_mean(self): result=str() for c,i in enumerate(self.metrics.keys()): result+=("%s→%.6f"+ ('' if (c==len(self.metrics.keys())-1) else ',')) % (i,np.mean(self.metrics[i])) return result def print_metrics(self): #for c,i in enumerate(self.metrics.keys()): # print("%s→%.6f" % (i,np.mean(self.metrics[i])),end=((c==len(self.metrics.keys())-1)?'':',')) print(self.get_metrics_mean()) pass def __str__(self): return f"Metrics-{list(self.metrics.keys())}\nAlgorithm-{self.algorithm}\nMetricsPath-{self.metrics_path}\nRecSize-{self.k}\n"+self.get_metrics_mean() def fwrite_metrics(self): fname=self.metrics_path+"/"+self.algorithm+"_at_"+str(self.k) result=str() result=f"""{str(self.metrics)}""" # for c,i in enumerate(self.metrics.keys()): # result+=("%s\t%.6f" % (i,np.mean(self.metrics[i]))+('' if (c==len(self.metrics.keys())-1) else '\n')) f=open(fname,"w+") f.write(result) f.close() print("File "+fname+" written with success") def fload_metrics(self): fname=self.metrics_path+"/"+self.algorithm+"_at_"+str(self.k) f=open(fname,"r") self.metrics=eval(f.read()) f.close()
1,886
675
from typing import TYPE_CHECKING from sqlalchemy import Boolean, Column, Integer, String from sqlalchemy.orm import relationship from app.db.base_class import Base if TYPE_CHECKING: from .pet import Pet # noqa: F401 class User(Base): id = Column(Integer, primary_key=True, index=True, autoincrement=True, nullable=False) first_name = Column(String(20), index=True, nullable=False) last_name = Column(String(20), index=True, nullable=False) email = Column(String, unique=True, index=True, nullable=False) hashed_password = Column(String, nullable=False) is_active = Column(Boolean(), default=True) is_superuser = Column(Boolean(), default=False) pets = relationship("Pet", back_populates="owner")
736
236
from django.contrib import admin from django.utils.html import format_html from . import get_user_model @admin.register(get_user_model()) class UserAdmin(admin.ModelAdmin): list_display = ['ext_id', 'username', 'name', 'email', 'active', 'is_bot', 'is_admin', 'get_photo', 'last_seen'] list_filter = ('active', 'is_bot', 'is_admin') search_fields = ['username', 'name', 'email', 'ext_id'] readonly_fields = ['ext_id', 'username', 'name', 'email', 'active', 'is_bot', 'is_admin', 'get_photo', 'last_seen'] exclude = ['photo', 'photo_thumb'] def get_photo(self, obj): if obj.photo: return format_html( '<a href="{}" target="_blank"><img src="{}" style="width:50px;"></a>', obj.photo, obj.photo_thumb ) return None get_photo.short_description = 'Photo' get_photo.allow_tags = True def has_add_permission(self, _): return False
930
305
# -*- coding: utf-8 -*- # Generated by Django 1.11.15 on 2019-04-24 14:53 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Place', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=30)), ('lat', models.FloatField()), ('lon', models.FloatField()), ('identifier', models.CharField(max_length=30)), ], ), migrations.CreateModel( name='Route', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('date', models.DateField(auto_now_add=True)), ], ), migrations.CreateModel( name='RP', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('place_id', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='tourouteapp.Place')), ('route_id', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='tourouteapp.Route')), ], ), ]
1,476
440
#!/usr/bin/env python from flask import Flask, render_template, request, redirect, jsonify, url_for, flash from sqlalchemy import create_engine, asc from sqlalchemy.orm import sessionmaker from database_setup import Base, Category, Item, User from flask import session as login_session import random import string import collections import json import requests from flask import make_response from oauth2client.client import flow_from_clientsecrets from oauth2client.client import FlowExchangeError import httplib2 # from dict2xml import dict2xml from xml.etree.ElementTree import Element, SubElement, Comment, tostring import psycopg2 # from page_views import * app = Flask(__name__) CLIENT_ID = json.loads(open('/vagrant/catalog/client_secret.json', 'r').read())['web']['client_id'] APPLICATION_NAME = "Catalog Category Items Application" engine = create_engine('sqlite:///catalog.db') Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) session = DBSession() # User Helper Functions def createUser(login_session): newUser = User(name=login_session['username'], email=login_session[ 'email'], picture=login_session['picture']) session.add(newUser) session.commit() user = session.query(User).filter_by(email=login_session['email']).one() return user.id def getUserInfo(user_id): user = session.query(User).filter_by(id=user_id).one() return user def getUserID(email): try: user = session.query(User).filter_by(email=email).one() return user.id except: return None # login page @app.route('/login') def showLogin(): state = ''.join(random.choice(string.ascii_uppercase + string.digits) for x in xrange(32)) login_session['state'] = state # return "The current session state is %s" % login_session['state'] return render_template('user_login.html', STATE=state) # display home page / categories page # Categories @app.route('/') @app.route('/categories/') def showCategories(): categories = session.query(Category).order_by(asc(Category.name)) items = session.query(Item).all() if 'username' not in login_session: return render_template( 'public_categories.html', categories=categories, items=items ) else: return render_template('categories.html', categories=categories, items=items) # Show Category Items @app.route('/categories/<int:category_id>/') @app.route('/categories/<int:category_id>/items/') def showItems(category_id): category = session.query(Category).filter_by(id=category_id).one() creator = getUserInfo(category.user_id) items = session.query(Item).filter_by( category_id=category_id).all() if ('username' not in login_session or creator.id != login_session['user_id']): return render_template( 'public_items.html', items=items, category=category, creator=creator ) else: return render_template( 'items.html', items=items, category=category, creator=creator ) # Create a new category @app.route('/categories/new/', methods=['GET', 'POST']) def newCategory(): if 'username' not in login_session: return redirect('/login') if request.method == 'POST': if request.form['name']: newCategory = Category( name=request.form['name'], user_id=login_session['user_id']) session.add(newCategory) flash( 'New Category %s Successfully Created' % newCategory.name ) session.commit() return redirect(url_for('showCategories')) else: flash('Please Complete Name Field') return render_template('new_category.html') else: return render_template('new_category.html') # Edit a categories @app.route('/categories/<int:category_id>/edit/', methods=['GET', 'POST']) def editCategory(category_id): editedCategory = session.query(Category).filter_by(id=category_id).one() if 'username' not in login_session: return redirect('/login') if editedCategory.user_id != login_session['user_id']: return """<script>function myFunction() {alert('You are not authorized to edit this category. Please create your own category in order to edit.');}</script> <body onload='myFunction()''>""" if request.method == 'POST': if request.form['name']: editedCategory.name = request.form['name'] flash('Successfully edited your Category %s' % editedCategory.name) return redirect(url_for('showItems', category_id=category_id)) else: return redirect(url_for('editCategory', category_id=category_id)) else: return render_template( 'edit_category.html', category=editedCategory, category_id=category_id ) # Delete a category @app.route('/categories/<int:category_id>/delete/', methods=['GET', 'POST']) def deleteCategory(category_id): categoryToDelete = session.query( Category).filter_by(id=category_id).one() if 'username' not in login_session: return redirect('/login') if categoryToDelete.user_id != login_session['user_id']: return """<script>function myFunction() {alert('You do not have permission to delete this category. Please create your own category in order to delete.');}</script> <body onload='myFunction()''>""" if request.method == 'POST': session.delete(categoryToDelete) flash('%s Successfully Deleted' % categoryToDelete.name) session.commit() return redirect(url_for( 'showCategories', category_id=category_id) ) else: return render_template( 'delete_category.html', category=categoryToDelete ) # Item Services # Create a new item @app.route('/categories/<int:category_id>/items/new/', methods=['GET', 'POST']) def newItem(category_id): if 'username' not in login_session: return redirect('/login') category = session.query(Category).filter_by(id=category_id).one() if login_session['user_id'] != category.user_id: return """<script>function myFunction() {alert('You are not authorized to add items to this category. Please create your own category in order to add items.'); }</script><body onload='myFunction()''>""" if request.method == 'POST': if (request.form['name'] and request.form['description']): newItem = Item( name=request.form['name'], description=request.form['description'], category_id=category_id, user_id=category.user_id ) session.add(newItem) session.commit() flash('New %s Item Successfully Created' % (newItem.name)) return redirect(url_for('showItems', category_id=category_id)) else: flash("Please Complete Form") return redirect(url_for('newItem', category_id=category_id, category=category)) else: return render_template( 'new_item.html', category_id=category_id, category=category ) # Edit a item @app.route('/categories/<int:category_id>/items/<int:item_id>/edit', methods=['GET', 'POST']) def editItem(category_id, item_id): if 'username' not in login_session: return redirect('/login') editedItem = session.query(Item).filter_by(id=item_id).one() category = session.query(Category).filter_by(id=category_id).one() if login_session['user_id'] != category.user_id: return """<script>function myFunction() {alert('You are not authorized to edit items to this category. Please create your own category in order to edit items.'); }</script><body onload='myFunction()''>""" if request.method == 'POST': if (request.form['name'] and request.form['description']): editedItem.name = request.form['name'] editedItem.description = request.form['description'] editedItem.category_id = category.id session.add(editedItem) session.commit() flash('Item Successfully Edited') return redirect(url_for('showItems', category_id=category_id)) else: flash("Do Not Leave Any Blanks") return redirect(url_for('editItem', category_id=category_id, item_id=item_id, item=editedItem, category=category)) else: return render_template( 'edit_item.html', category_id=category_id, item_id=item_id, item=editedItem, category=category ) # Delete a item @app.route( '/categories/<int:category_id>/items/<int:item_id>/delete', methods=['GET', 'POST'] ) def deleteItem(category_id, item_id): if 'username' not in login_session: return redirect('/login') category = session.query(Category).filter_by(id=category_id).one() itemToDelete = session.query(Item).filter_by(id=item_id).one() if login_session['user_id'] != category.user_id: return """<script>function myFunction() {alert('You are not authorized to delete items to this category. Please create your own category in order to delete items.') ;}</script><body onload='myFunction()''>""" if request.method == 'POST': session.delete(itemToDelete) session.commit() flash('Item Successfully Deleted') return redirect(url_for('showItems', category_id=category_id)) else: return render_template( 'delete_item.html', item=itemToDelete, category=category ) # Disconnect based on provider @app.route('/logout') @app.route('/disconnect') def disconnect(): if 'provider' in login_session: if login_session['provider'] == 'google': gdisconnect() del login_session['gplus_id'] del login_session['access_token'] if login_session['provider'] == 'facebook': fbdisconnect() del login_session['facebook_id'] del login_session['username'] del login_session['email'] del login_session['picture'] del login_session['user_id'] del login_session['provider'] flash("You have successfully been logged out.") return redirect(url_for('showCategories')) else: flash("You were not logged in") return redirect(url_for('showCategories')) ##### JSON APIs to view Category Information @app.route('/categories/<int:category_id>/items/JSON') def categoryItemJSON(category_id): category = session.query(Category).filter_by(id=category_id).one() items = session.query(Item).filter_by( category_id=category_id).all() return jsonify(Items=[i.serialize for i in items]) @app.route('/categories/<int:category_id>/items/<int:item_id>/JSON') def itemJSON(category_id, item_id): item = session.query(Item).filter_by(id=item_id).one() return jsonify(Item=Item.serialize) @app.route('/categories/JSON') def categoryJSON(): categories = session.query(Category).all() return jsonify(categories=[r.serialize for r in categories]) ##### Social media routes ##### @app.route('/fbconnect', methods=['POST']) def fbconnect(): if request.args.get('state') != login_session['state']: response = make_response(json.dumps('Invalid state parameter.'), 401) response.headers['Content-Type'] = 'application/json' return response access_token = request.data print "access token received %s " % access_token app_id = json.loads(open('fb_client_secrets.json', 'r').read())[ 'web']['app_id'] app_secret = json.loads( open('fb_client_secret.json', 'r').read())['web']['app_secret'] url = 'https://graph.facebook.com/oauth/access_token?grant_type=fb_exchange_token&client_id=%s&client_secret=%s&fb_exchange_token=%s' % (app_id, app_secret, access_token) h = httplib2.Http() result = h.request(url, 'GET')[1] # Use token to get user info from API userinfo_url = "https://graph.facebook.com/v2.4/me" # strip expire tag from access token token = result.split("&")[0] url = 'https://graph.facebook.com/v2.4/me?%s&fields=name,id,email' % token h = httplib2.Http() result = h.request(url, 'GET')[1] # print "url sent for API access:%s"% url # print "API JSON result: %s" % result data = json.loads(result) login_session['provider'] = 'facebook' login_session['username'] = data["name"] login_session['email'] = data["email"] login_session['facebook_id'] = data["id"] # The token must be stored in the login_session # in order to properly logout, let's strip out # the information before the equals sign in our token stored_token = token.split("=")[1] login_session['access_token'] = stored_token # Get user picture url = 'https://graph.facebook.com/v2.4/me/picture?%s&redirect=0&height=200&width=200' % token h = httplib2.Http() result = h.request(url, 'GET')[1] data = json.loads(result) login_session['picture'] = data["data"]["url"] # see if user exists user_id = getUserID(login_session['email']) if not user_id: user_id = createUser(login_session) login_session['user_id'] = user_id output = '' output += '<h1>Welcome, ' output += login_session['username'] output += '!</h1>' output += '<img src="' output += login_session['picture'] output += ''' " style = "width: 300px; height: 300px; border-radius: 150px; -webkit-border-radius: 150px; -moz-border-radius: 150px;"> ''' flash("Now logged in as %s" % login_session['username']) return output @app.route('/fbdisconnect') def fbdisconnect(): facebook_id = login_session['facebook_id'] # The access token must me included to successfully logout access_token = login_session['access_token'] url = 'https://graph.facebook.com/%s/permissions?' \ 'access_token=%s' % (facebook_id, access_token) h = httplib2.Http() result = h.request(url, 'DELETE')[1] return "you have been logged out" @app.route('/gconnect', methods=['POST']) def gconnect(): # Validate state token if request.args.get('state') != login_session['state']: response = make_response(json.dumps('Invalid state parameter.'), 401) response.headers['Content-Type'] = 'application/json' return response # Obtain authorization code code = request.data try: # Upgrade the authorization code into a credentials object oauth_flow = flow_from_clientsecrets('client_secret.json', scope='') oauth_flow.redirect_uri = 'postmessage' credentials = oauth_flow.step2_exchange(code) except FlowExchangeError: response = make_response( json.dumps('Failed to upgrade the authorization code.'), 401) response.headers['Content-Type'] = 'application/json' return response # Check that the access token is valid. access_token = credentials.access_token url = ('https://www.googleapis.com/oauth2/v1/tokeninfo?access_token=%s' % access_token) h = httplib2.Http() result = json.loads(h.request(url, 'GET')[1]) # If there was an error in the access token info, abort. if result.get('error') is not None: response = make_response(json.dumps(result.get('error')), 500) response.headers['Content-Type'] = 'application/json' return response # Verify that the access token is used for the intended user. gplus_id = credentials.id_token['sub'] if result['user_id'] != gplus_id: response = make_response( json.dumps("Token's user ID doesn't match given user ID."), 401) response.headers['Content-Type'] = 'application/json' return response # Verify that the access token is valid for this app. if result['issued_to'] != CLIENT_ID: response = make_response( json.dumps("Token's client ID does not match app's."), 401) print "Token's client ID does not match app's." response.headers['Content-Type'] = 'application/json' return response stored_credentials = login_session.get('credentials') stored_gplus_id = login_session.get('gplus_id') if stored_credentials is not None and gplus_id == stored_gplus_id: response = make_response(json.dumps( 'Current user is already connected.'), 200) response.headers['Content-Type'] = 'application/json' return response # Store the access token in the session for later use. login_session['access_token'] = credentials.access_token login_session['gplus_id'] = gplus_id # Get user info userinfo_url = "https://www.googleapis.com/oauth2/v1/userinfo" params = {'access_token': credentials.access_token, 'alt': 'json'} answer = requests.get(userinfo_url, params=params) data = answer.json() login_session['username'] = data['name'] login_session['picture'] = data['picture'] login_session['email'] = data['email'] # ADD PROVIDER TO LOGIN SESSION login_session['provider'] = 'google' # see if user exists, if it doesn't make a new one user_id = getUserID(data["email"]) if not user_id: user_id = createUser(login_session) login_session['user_id'] = user_id output = '' output += '<h1>Welcome, ' output += login_session['username'] output += '!</h1>' output += '<img src="' output += login_session['picture'] output += ''' "style = "width: 300px; height: 300px; border-radius: 150px; -webkit-border-radius: 150px; -moz-border-radius: 150px;"> ''' flash("you are now logged in as %s" % login_session['username']) print "done!" return output # DISCONNECT - Revoke a current user's token and reset their login_session @app.route('/gdisconnect') def gdisconnect(): # Only disconnect a connected user. credentials = login_session.get('credentials') if credentials is None: response = make_response( json.dumps('Current user not connected.'), 401) response.headers['Content-Type'] = 'application/json' return response access_token = credentials.access_token url = 'https://accounts.google.com/' \ 'o/oauth2/revoke?token=%s' % access_token h = httplib2.Http() result = h.request(url, 'GET')[0] if result['status'] != '200': # For whatever reason, the given token was invalid. response = make_response( json.dumps('Failed to revoke token for given user.'), 400) response.headers['Content-Type'] = 'application/json' return response if __name__ == '__main__': app.secret_key = "lRYRXEimZGfbt3Q2TpD_6_Kj" app.debug = True app.run(host='0.0.0.0', port=8002)
19,691
5,694
class InvalidConfigurationError(Exception): pass class InvalidBotToken(Exception): pass class InvalidBotPlugin(Exception): pass
144
40
import os from setuptools import setup version = os.getenv('VERSION', '1.10.1') setup( name='tensorflow-autodetect', version=version, url='https://github.com/commaai/tensorflow-autodetect', author='comma.ai', author_email='', license='MIT', long_description='Auto-detect tensorflow or tensorflow-gpu package based on nvidia driver being installed', keywords='tensorflow tensorflow-gpu', install_requires=[ ('tensorflow-gpu' if os.path.exists('/proc/driver/nvidia/version') else 'tensorflow') + '==' + version, ], classifiers=[ 'Natural Language :: English', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ], )
723
218
from rest_framework import permissions from django.shortcuts import get_object_or_404 from django.contrib.auth.models import User from intask_api.projects.models import Project class IsProjectHeader(permissions.BasePermission): def has_object_permission(self, request, view, obj): return obj.header == request.user class CanUpdateProject(permissions.BasePermission): def has_object_permission(self, request, view, obj): return obj.header == request.user class CanDeleteProject(permissions.BasePermission): def has_object_permission(self, request, view, obj): return obj.header == request.user class IsParticipant(permissions.BasePermission): def has_object_permission(self, request, view, obj): return request.user in obj.users.all() class CanRetrieveProjectUser(permissions.BasePermission): def has_permission(self, request, view): project = get_object_or_404(Project, id=view.kwargs['project_id']) return request.user in project.users.all() class CanDeleteProjectUser(permissions.BasePermission): def has_permission(self, request, view): project = get_object_or_404(Project, id=view.kwargs['project_id']) user = get_object_or_404(User, id=view.kwargs['pk']) is_project_header = project.header == request.user is_current_user = request.user == user return is_project_header | is_current_user class CanAddProjectUser(permissions.BasePermission): def has_permission(self, request, view): project = get_object_or_404(Project, id=view.kwargs['project_id']) return request.user == project.header
1,639
470
'''3. Write a Python program to split a list into different variables. ''' universalList = [(1, 2, 3), ('w', 'e', 's')] lst1, lst2 = universalList print(lst1) print(lst2)
171
63
# Python3 from solution1 import checkParticipants as f qa = [ ([0, 1, 1, 5, 4, 8], [2]), ([0, 1, 2, 3, 4, 5], []), ([6], []), ([3, 3, 3, 3, 3, 3, 3, 3], [4, 5, 6, 7]), ([0, 0, 1, 5, 5, 4, 5, 4, 10, 8], [1, 2, 5, 6, 7, 9]) ] for *q, a in qa: for i, e in enumerate(q): print('input{0}: {1}'.format(i + 1, e)) ans = f(*q) if ans != a: print(' [failed]') print(' output:', ans) print(' expected:', a) else: print(' [ok]') print(' output:', ans) print()
576
289
from models.base_model import BaseModel from models.classifier import BertClassifier from models.sequence_labeling import BertSequenceLabeling from models.sequence_embedding import BertSequenceEmbedding
202
47
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # python3 -m pip install --force -U --user PlexAPI import json import time import logging import plexapi import plexapi.video import plexapi.myplex import plexapi.server import plexapi.library import plexapi.exceptions PLEX_URL = "" PLEX_TOKEN = "" WATCHED_HISTORY = "" LOG_FILE = "" BATCH_SIZE = 10000 PLEX_REQUESTS_SLEEP = 0 CHECK_USERS = [ ] LOG_FORMAT = \ "[%(name)s][%(process)05d][%(asctime)s][%(levelname)-8s][%(funcName)-15s]" \ " %(message)s" LOG_DATE_FORMAT = "%Y-%m-%dT%H:%M:%SZ" LOG_LEVEL = logging.INFO plexapi.server.TIMEOUT = 3600 plexapi.server.X_PLEX_CONTAINER_SIZE = 2500 _SHOW_GUID_RATING_KEY_MAPPING = {} _MOVIE_GUID_RATING_KEY_MAPPING = {} _EPISODE_GUID_RATING_KEY_MAPPING = {} logger = logging.getLogger("PlexWatchedHistoryImporter") def _get_config_str(key): return plexapi.CONFIG.get(key, default="", cast=str).strip("'").strip('"').strip() def _load_config(): global PLEX_URL, PLEX_TOKEN, WATCHED_HISTORY, CHECK_USERS, LOG_FILE, LOG_LEVEL if PLEX_URL == "": PLEX_URL = _get_config_str("sync.dst_url") if PLEX_TOKEN == "": PLEX_TOKEN = _get_config_str("sync.dst_token") if WATCHED_HISTORY == "": WATCHED_HISTORY = _get_config_str("sync.watched_history") if len(CHECK_USERS) == 0: config_check_users = _get_config_str("sync.check_users").split(",") CHECK_USERS = [user.strip() for user in config_check_users if user] if LOG_FILE == "": LOG_FILE = _get_config_str("sync.import_log_file") debug = plexapi.utils.cast(bool, _get_config_str("sync.debug").lower()) if debug: LOG_LEVEL = logging.DEBUG def _setup_logger(): logging.Formatter.converter = time.gmtime logging.raiseExceptions = False logger.setLevel(logging.DEBUG) logger.handlers = [] logger.propagate = False detailed_formatter = logging.Formatter(fmt=LOG_FORMAT, datefmt=LOG_DATE_FORMAT) file_handler = logging.FileHandler(filename=LOG_FILE, mode="a+") file_handler.setFormatter(detailed_formatter) file_handler.setLevel(LOG_LEVEL) logger.addHandler(file_handler) def _get_rating_keys(server, rating_key_guid_mapping, guid): if guid not in rating_key_guid_mapping: items = server.library.search(guid=guid) rating_key_guid_mapping[guid] = [item.ratingKey for item in items] return rating_key_guid_mapping[guid] def _set_movie_section_watched_history(server, movie_history): for movie_guid, movie_item_history in movie_history.items(): rating_keys = _get_rating_keys(server, _MOVIE_GUID_RATING_KEY_MAPPING, movie_guid) for rating_key in rating_keys: item = server.fetchItem(rating_key) if movie_item_history['watched'] and not item.isWatched: logger.debug(f"Watching Movie: {item.title}") item.markWatched() if movie_item_history['viewCount'] > item.viewCount: for _ in range(movie_item_history['viewCount'] - item.viewCount): logger.debug(f"Watching Movie: {item.title}") item.markWatched() if movie_item_history['viewOffset'] != 0: logger.debug(f"Updating Movie Timeline: {item.title}: {movie_item_history['viewOffset']}") item.updateTimeline(movie_item_history['viewOffset']) if movie_item_history['userRating'] != "": logger.debug(f"Rating Movie: {item.title}: {movie_item_history['userRating']}") item.rate(movie_item_history['userRating']) def _set_show_section_watched_history(server, show_history): for show_guid, show_item_history in show_history.items(): rating_keys = _get_rating_keys(server, _SHOW_GUID_RATING_KEY_MAPPING, show_guid) for rating_key in rating_keys: item = server.fetchItem(rating_key) if show_item_history['watched'] and not item.isWatched: logger.debug(f"Watching Show: {item.title}") item.markWatched() if show_item_history['userRating'] != "": logger.debug(f"Rating Show: {item.title}: {show_item_history['userRating']}") item.rate(show_item_history['userRating']) for episode_guid, episode_item_history in show_item_history['episodes'].items(): rating_keys = _get_rating_keys(server, _EPISODE_GUID_RATING_KEY_MAPPING, episode_guid) for rating_key in rating_keys: item = server.fetchItem(rating_key) if episode_item_history['watched'] and not item.isWatched: logger.debug(f"Watching Episode: {item.title}") item.markWatched() if episode_item_history['viewCount'] > item.viewCount: for _ in range(episode_item_history['viewCount'] - item.viewCount): logger.debug(f"Watching Episode: {item.title}") item.markWatched() if episode_item_history['viewOffset'] != 0: logger.debug(f"Updating Episode Timeline: {item.title}: {episode_item_history['viewOffset']}") item.updateTimeline(episode_item_history['viewOffset']) if episode_item_history['userRating'] != "": logger.debug(f"Rating Episode: {item.title}: {episode_item_history['userRating']}") item.rate(episode_item_history['userRating']) def _set_user_server_watched_history(server, watched_history): _set_movie_section_watched_history(server, watched_history['movie']) _set_show_section_watched_history(server, watched_history['show']) def main(): _load_config() _setup_logger() plex_server = plexapi.server.PlexServer(PLEX_URL, PLEX_TOKEN, timeout=300) plex_account = plex_server.myPlexAccount() with open(WATCHED_HISTORY, "r") as watched_history_file: watched_history = json.load(watched_history_file) logger.info(f"Starting Import") plex_users = plex_account.users() # Owner will be processed separately logger.info(f"Total Users: {len(plex_users) + 1}") if not (len(CHECK_USERS) > 0 and plex_account.username not in CHECK_USERS and plex_account.email not in CHECK_USERS): logger.info(f"Processing Owner: {plex_account.username}") user_history = watched_history[plex_account.username] _set_user_server_watched_history(plex_server, user_history) for user_index, user in enumerate(plex_users): if (len(CHECK_USERS) > 0 and user.username not in CHECK_USERS and user.email not in CHECK_USERS): continue if user.username not in watched_history: logger.warning(f"Missing User from Watched History: {user.username}") continue logger.info(f"Processing User: {user.username}") user_server_token = user.get_token(plex_server.machineIdentifier) try: user_server = plexapi.server.PlexServer(PLEX_URL, user_server_token, timeout=300) except plexapi.exceptions.Unauthorized: # This should only happen when no libraries are shared logger.warning(f"Skipped User with No Libraries Shared: {user.username}") continue user_history = watched_history[user.username] _set_user_server_watched_history(user_server, user_history) logger.info(f"Completed Import") if __name__ == "__main__": main()
7,563
2,506
import os import sys import time import exceptions import uuid import logging ''' @author: msune,lbergesio,omoya,CarolinaFernandez @organization: i2CAT, OFELIA FP7 PolicyEngine Rule class Encapsulates logic of a simple Rule ''' from django_thermostat.pypelib.Condition import Condition from django_thermostat.pypelib.persistence.PersistenceEngine import PersistenceEngine from django_thermostat.pypelib.utils.Logger import Logger class TerminalMatch(exceptions.Exception): value = None desc = None def __init__(self,rType,desc): if isinstance(rType['value'],bool): self.value = rType['value'] else: raise Exception("Unknown rule type") self.desc = desc def __str__(self): return "%s "%self.desc class Rule(): logger = Logger.getLogger() #Class Attributes _condition = None _description = None _errorMsg = None _uuid = None #uuid.uuid4().hex _defaultParser = "RegexParser" _defaultPersistence = "Django" #Types of rule POSITIVE_TERMINAL={'value':True,'terminal':True} POSITIVE_NONTERMINAL={'value':True,'terminal':False} NEGATIVE_TERMINAL={'value':False,'terminal':True} NEGATIVE_NONTERMINAL={'value':False,'terminal':False} _types = [POSITIVE_TERMINAL,POSITIVE_NONTERMINAL,NEGATIVE_TERMINAL, NEGATIVE_NONTERMINAL] #Rule type _type = None #Rule match Action _matchAction=None #Getters def getCondition(self): return self._condition def getDescription(self): return self._description def getType(self): return self._type def getErrorMsg(self): return self._errorMsg def getMatchAction(self): return self._matchAction def getUUID(self): return self._uuid #setters def setUUID(self,UUID): self._uuid = UUID #Constructor def __init__(self,condition,description,errorMsg,ruleType=POSITIVE_TERMINAL,action=None,uuid=None): if not isinstance(condition,Condition): raise Exception("Object must be an instance of Condition") if ruleType not in self._types: raise Exception("Unknown rule type") if action == None and (ruleType == self.NEGATIVE_NONTERMINAL or ruleType == self.POSITIVE_NONTERMINAL): raise Exception("You cannot create non-terminal actionless rules") self._condition = condition self._matchAction = action self._type = ruleType self._description = description self._errorMsg = errorMsg self._uuid = uuid def dump(self): #Debug dump toReturn = self._condition.dump() toReturn+="=> %s "%str(self._type['value']) if self._matchAction != None: toReturn += "(%s) "%str(self._matchAction) if self._type['terminal']: toReturn += "[TERM] " if self._description: toReturn+=" #"+self._description return toReturn #Resolver is passed at evaluation time to be able to dynamically redirect actions def evaluate(self,metaObj,resolver): try: Rule.logger.setLevel(logging.DEBUG) result = self._condition.evaluate(metaObj,resolver) Rule.logger.debug('Result was: %s rule: [%s]' % (str(result), self.dump())) except Exception as e: Rule.logger.error('Error on rule: %s',self.dump()) Rule.logger.error('Exception: %s', str(e)) Rule.logger.error('Rule will be skiped!') result = False if result: if self._matchAction != None: resolver.resolve(self._matchAction,metaObj) #If is terminal raise TerminalMatch if self._type['terminal']: raise TerminalMatch(self._type,self._errorMsg) #return whatever return def getConditionDump(self): return self.getCondition().dump()
3,556
1,246
def test_get_tags(te): r = te.get_tags() assert r == []
64
27
from birds.display_utils import geo_plot from birds.pann import load_pretrained_model, read_audio_fast, get_model_predictions_for_clip, BIRDS def test_prediction_works(): test_bird = 'comrav' model = load_pretrained_model() y = read_audio_fast(f'./data/audio/{test_bird}.mp3') predictions = get_model_predictions_for_clip(y, model) class_probs = predictions[BIRDS].sum().reset_index() class_probs.columns = ['ebird', 'p'] class_probs = class_probs.sort_values(by='p') top_ebird = class_probs.ebird.values[-1] assert top_ebird == test_bird def test_map(): html = geo_plot('norcar', 10, 10) with open('./temp/test_map.html', 'w') as f: f.write(html)
709
270
#!/bin/python import subprocess import os import sys import maps import time import logging def line(n): print('-'*n) logging.basicConfig(filename = "LetsHPC_Team_CodeRunner.log", level = logging.INFO) logger = logging.getLogger(__name__) ######################################################################################################## USAGE = """ Usage: run.py problem_name approach_name serial_executable parallel_executable runs log_directory output_directory input_directory base_directory 'problem_name' is the name of the problem assigned to you. 'approach_name' is the name of the appraoch assigned to you. 'serial_executable' must be the name of the compiled executable file for the serial code. 'parallel_executable' must be the name of the compiled executable file for the parallel code. 'runs' is the number of times to run the codes. Run at least thrice and ideally 10 times. 'log_directory' is the directory where you want to store the log files 'output_directory' is the directory where you want to store the output files 'input_directory' is the directory where you take the input from """ def foobar(l): if len(l) < 10: print USAGE return problem_name = l[1] approach_name = l[2] serial_executable = l[3] parallel_executable = l[4] runs = int(l[5]) compiler_to_use = l[-1] logger.info("-"*80) logger.info("Problem Name : %s" % (problem_name)) logger.info("Approach Name : %s" % (approach_name)) logger.info("Serial Executable : %s" % (serial_executable)) logger.info("Parallel Executable : %s" % (parallel_executable)) logger.info("Number of runs : %s" % (str(runs))) if problem_name not in maps.problem_list: print problem_name, 'not in', maps.problem_list logger.error("%s not in problem list" % (problem_name)) exit(0) if approach_name not in maps.approaches[problem_name]: print approach_name, 'not a valid approach for', problem_name print 'Choose from:' print maps.approaches[problem_name] logger.error("%s is not a valid approach" % (approach_name)) exit(0) log_directory = l[6] output_directory = l[7] input_directory = l[8] line(80) logger.info("Log Directory : %s" %(log_directory)) logger.info("Output Directory : %s" % (output_directory)) logger.info("Input Directory : %s" % (input_directory)) print 'Assuming that input has been created for:', problem_name subprocess.call('lscpu > ' + log_directory + "lscpu.txt", shell=True) subprocess.call('cat /proc/cpuinfo > ' + log_directory + "cpuinfo.txt", shell=True) for run in range(runs): os.chdir(l[9]) print 'Run:', str(run+1) print('Running Serial') logger.info("Started running the serial code for run_id = %d" %(run)) for n in maps.problem_size[problem_name]: print('Problem Size:', n) input_file = input_directory+problem_name+'_'+str(n)+'_input.txt' if compiler_to_use == 'openmp': logger.info("Running the Command : " + serial_executable + " " + str(n) + " " + str(0) # p=0 for serial code. + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs") subprocess.call(serial_executable + " " + str(n) + " " + str(0) # p=0 for serial code. + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs", shell=True) elif compiler_to_use == 'mpi': logger.info("Running the Command : mpirun -np 1 " + serial_executable + " " + str(n) + " " + str(0) # p=0 for serial code. + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs") subprocess.call("mpirun -np 1 " + serial_executable + " " + str(n) + " " + str(0) # p=0 for serial code. + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs", shell=True) line(80) print('Running Parallel') for p in maps.processor_range: print('Number of Processors:', p) logger.info("Running the parallel code with %d processors" % (p)) for n in maps.problem_size[problem_name]: os.chdir(l[9]) input_file = input_directory+problem_name+'_'+str(n)+'_input.txt' print('Problem Size:', n) if compiler_to_use == 'openmp': logger.info("Running the Command : " + parallel_executable + " " + str(n) + " " + str(p) + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs") subprocess.call(parallel_executable + " " + str(n) + " " + str(p) + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs", shell=True) elif compiler_to_use == 'mpi': logger.info("Running the Command : mpirun -np " + str(p) + " " + parallel_executable + " " + str(n) + " " + str(p) + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs") subprocess.call("mpirun -np " + str(p) + " " + parallel_executable + " " + str(n) + " " + str(p) + " " + input_file + " >> " + log_directory + problem_name + "_" + approach_name + ".logs", shell=True) line(80) print(os.getcwd()) ####################################################################### base = os.getcwd() all_files = os.listdir(base) inp = None while True: if 'codes_run_file' in all_files: inp = raw_input("Do you want to reuse the results of previous run? (y/n): ").lower() if inp == 'y': break elif inp == 'n': os.remove(base + '/codes_run_file') break else: print "Invalid input. Try again." else: break while True: compiler_to_use = raw_input("Which parallel framework would you be using? (openmp/mpi): ").lower() if compiler_to_use == 'mpi' or compiler_to_use == 'openmp': break else: print("Incorrect input. Try again.") while True: try: runs = int(raw_input("Enter the number of times you want the code to run (recommended: at least 10 runs): ")) if runs <= 0: # if not a positive int print message and ask for input again print("Input must be a positive integer, try again!") continue except ValueError as ve: print("That's not an int! Try again!") continue else: print('the number of runs is ' + str(runs)) break all_inputs = os.getcwd() + '/all_input/' base = os.getcwd() + '/all_codes/' starting_point = os.getcwd() all_codes = os.listdir(base) count = 0 try: os.remove(base + "progress.txt") except Exception as e: print "File already deleted" print(all_codes) code_to_run = None codes_already_run = None try: uber = open(os.getcwd() + "/codes_run_file", "r") codes_already_run = uber.readlines() uber.close() except Exception as e: command = "touch %s" % (starting_point + "/codes_run_file") subprocess.call(command, shell = True) if codes_already_run is None: code_to_run = all_codes[0] else: for each in all_codes: if each+"\n" not in codes_already_run: code_to_run = each break print "The following code will be run now", code_to_run if code_to_run is None: print "All the codes have already been executed."# + " You can run the collect data script now" sys.exit(1) for each_code in [code_to_run]: if each_code == "progress.txt" or "log" in each_code: continue subprocess.call("rm -rf " + base + each_code + "/output" , shell=True) subprocess.call("rm -rf " + base + each_code + "/logs" , shell=True) division = each_code.split("-") problem = division[2] approach = division[3] print "-"*80 print problem, approach all_files = os.listdir(base+each_code+"/") serial = None parallel = None for each_file in all_files: if 'clean' not in each_file.lower() and 'logs'!=each_file.lower() and 'output'!=each_file.lower(): if 'par' not in each_file.lower() and each_file!="ser": serial = each_file elif 'parallel' in each_file.lower(): parallel = each_file if compiler_to_use == 'mpi': compiler = "mpicc " elif compiler_to_use == 'openmp': compiler = "gcc " if ".cpp" in parallel: if compiler_to_use == "mpi": compiler = "mpiCC " elif compiler_to_use == "openmp": compiler = "g++ " print serial, parallel if 'logs' not in all_files: os.mkdir(base + each_code + "/logs") os.mkdir(base + each_code + "/output") if compiler_to_use == 'openmp': subprocess.call(compiler + base + each_code + "/" + parallel + " -fopenmp -lm -w -o " + base + each_code + "/parr", shell=True) subprocess.call(compiler + base + each_code + "/" + serial + " -fopenmp -lm -w -o " + base + each_code + "/ser", shell=True) elif compiler_to_use == 'mpi': subprocess.call(compiler + base + each_code + "/" + parallel + " -lm -w -o " + base + each_code + "/parr", shell=True) subprocess.call(compiler + base + each_code + "/" + serial + " -lm -w -o " + base + each_code + "/ser", shell=True) print serial,parallel #raw_input() foobar(['run.py', problem, approach, base + each_code + "/ser", base + each_code + "/parr", int(runs), base + each_code + "/logs/", \ base + each_code + "/output/", all_inputs, base + each_code + "/", compiler_to_use]) f = open(base + "progress.txt", "a") f.write(str(time.time()) + " " + str(count) + " " + str(each_code)+"\n") f.close() count +=1 print "Reached Here:", code_to_run, type(code_to_run) w2f = open(starting_point + "/codes_run_file", "a") string_to_write = code_to_run + "\n" w2f.write(string_to_write) w2f.close() print "Written To file"
12,306
3,539
#!/usr/bin/env python # -*- coding: utf-8 -*- ''' ***************************************** Author: zhlinh Email: zhlinhng@gmail.com Version: 0.0.1 Created Time: 2016-03-24 Last_modify: 2016-03-24 ****************************************** ''' ''' Given two strings s and t, determine if they are isomorphic. Two strings are isomorphic if the characters in s can be replaced to get t. All occurrences of a character must be replaced with another character while preserving the order of characters. No two characters may map to the same character but a character may map to itself. For example, Given "egg", "add", return true. Given "foo", "bar", return false. Given "paper", "title", return true. Note: You may assume both s and t have the same length. ''' class Solution(object): def isIsomorphic(self, s, t): """ :type s: str :type t: str :rtype: bool """ if len(s) != len(t): return False m1 = [0] * 256 m2 = [0] * 256 for i in range(len(s)): if m1[ord(s[i])] != m2[ord(t[i])]: return False m1[ord(s[i])] = i + 1 m2[ord(t[i])] = i + 1 return True
1,236
407
import pymongo ## local 테스트 ## from dotenv import load_dotenv import os import pytz import datetime import itertools from uuid import uuid4 from collections import defaultdict # ObjectId로 쿼리할때 필요 from bson.objectid import ObjectId """ RDBMS Mongo DB Database Database Table Collection Row Document Index Index DB server Mongod DB client mongo """ class BaseWorker: def __init__(self): ## local 테스트 ## dot_env_path = os.path.dirname(os.path.abspath(__file__)) load_dotenv(dotenv_path=f"{dot_env_path}/../cred/.mongopasswd",verbose=True) ip = os.getenv("ip") port = os.getenv("port") user = os.getenv("user") passwd = os.getenv("passwd") self.client = pymongo.MongoClient(f"mongodb://{user}:{passwd}@{ip}:{port}/") class newWorker(BaseWorker): def __init__(self): super().__init__() self.db = self.client["stock"] self.collection = self.db["en"] def insert_regacy(self, results, subject): for result in results: result["subject"] = subject self.collection.insert_one(result) def migration_for_mistyping(self): self.collection.update_many({ "subejct": { "$exists": True } }, { "$rename": { 'subejct': 'subject'} }) class regacyWorker(BaseWorker): def __init__(self, dbname: str, collection: str): super().__init__() self.db = self.client[dbname] self.collection = self.db[collection] def get_all_data(self, subejct): res = self.collection.find() filterd_res = list() for r in res: if("url" in r["sentiment"][0]): filterd_res.append(r) print(subject, len(filterd_res)) return filterd_res if(__name__ == "__main__"): new_db_worker = newWorker() new_db_worker.migration_for_mistyping()
1,860
629
import numpy as np import cv2 as cv import os import sys class ObjectDetector: """ Object Detector is the class model for using YOLOv2 and gathering results """ def __init__(self): self.network_loading() def network_loading(self): # Loading in the trained darknet models labels self.LABELS = open(os.getcwd() + "\coco.names").read().strip().split("\n") self.readingNetwork = cv.dnn.readNetFromDarknet(os.getcwd() + "\yolov3.cfg", os.getcwd() + "\yolov3.weights") def read_image(self, image_name): # Reading in a specific image from the files that exist in the input folder working_image = cv.imread(image_name) self.labelNames = self.readingNetwork.getLayerNames() self.labelNames = [self.labelNames[i[0] - 1] for i in self.readingNetwork.getUnconnectedOutLayers()] imageInputBlob = cv.dnn.blobFromImage(working_image, 1 / 255.0, (416, 416), swapRB=True, crop=False) self.readingNetwork.setInput(imageInputBlob) layerOutputs = self.readingNetwork.forward(self.labelNames) return self.processReading(layerOutputs) def processReading(self, processingResults): # Takes in the results from a reading and proceses them to check for valid objects classIDs = [] for objects in processingResults: # loop over each of the detections for detection in objects: scores = detection[5:] classID = np.argmax(scores) confidence = scores[classID] if confidence > 0.9: # Appending the names of all the objects to be sorted later classIDs.append(self.LABELS[classID]) # Just returning class names as it's only thing relevant to OSR later return classIDs
1,659
611
# -*- coding: utf-8 from django.apps import AppConfig class ClosuredagConfig(AppConfig): name = 'closuredag'
115
41
class Stack(): def __init__(self): self.stack = [] def push(self, value): self.stack.append(value) def pop(self): return self.stack.pop() def __len__(self): return len(self.stack) def print(self): return print(self.stack) if __name__ == '__main__': str_ = input() stack = Stack() for i in str_: stack.push(i) for j in range(len(stack)): print(stack.pop(), end='') print()
499
177
from utils.db import connection, print_version import pandas as pd def add_table(csv_file, table_name, engine): df = pd.read_csv(csv_file) df = df.drop('Unnamed: 0') df.to_sql(name=table_name, con=engine, index=False, if_exists='replace') table = 'data/tables/postcode_coordinates.csv' add_table(table, 'Postcode_coordinates', connection) cur = connection.cursor() cur.execute('''SELECT * FROM Postcode_coordinates''') data = cur.fetchmany(5) print(data)
484
167
import runpy runpy._run_module_as_main('SeeMee')
48
20
from .name2idx import parameters as C def f_params(): x = [0]*C.len_f_params x[C.Kon_IL13Rec] = 0.00341992 x[C.Rec_phosphorylation] = 999.631 x[C.pRec_intern] = 0.15254 x[C.pRec_degradation] = 0.172928 x[C.Rec_intern] = 0.103346 x[C.Rec_recycle] = 0.00135598 x[C.JAK2_phosphorylation] = 0.157057 x[C.pJAK2_dephosphorylation] = 6.21906E-4 x[C.STAT5_phosphorylation] = 0.0382596 x[C.pSTAT5_dephosphorylation] = 3.43392E-4 x[C.SOCS3mRNA_production] = 0.00215826 x[C.DecoyR_binding] = 1.24391E-4 x[C.JAK2_p_inhibition] = 0.0168268 x[C.SOCS3_translation] = 11.9086 x[C.SOCS3_accumulation] = 3.70803 x[C.SOCS3_degradation] = 0.0429186 x[C.CD274mRNA_production] = 8.21752E-5 return x
755
454
import os import datetime import click from flask.cli import with_appcontext from werkzeug.security import generate_password_hash def init_app(app): app.cli.add_command(init_db_command) @click.command('init-db') @with_appcontext def init_db_command(): """Clear the existing data and create new tables.""" init_db() click.echo('Initialized the database.') def init_db(): from . import models models.db.create_all() user = models.User(username='admin', password=generate_password_hash('admin')) models.db.session.add(user) models.db.session.commit()
593
186
with open('./logs/test.log', 'r') as f1: data = f1.readlines() formatted = [x.replace('.pdf', '') for x in data] with open('./logs/test2.log', 'r') as f1: f1.writelines(formatted)
191
79
import argparse import json from pyhmy import ( get_all_validator_addresses, get_validator_information ) def get_block_by_num(block_num, endpoint): params = [ str(hex(block_num)), False, ] payload = { "id": "1", "jsonrpc": "2.0", "method": "hmy_getBlockByNumber", "params": params } headers = { 'Content-Type': 'application/json' } timeout = 5 try: resp = requests.request('POST', endpoint, headers=headers, data=json.dumps(payload), timeout=timeout, allow_redirects=True) return json.loads(resp.content) except Exception as e: v_print(f'{e.__class__}: {e}') return None if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument("--start", required=True, type=int, help="First block") parser.add_argument("--end", required=True, type=int, help="Last block") parser.add_argument("--endpoint", default="http://localhost:9500", help="Endpoint to query") parser.add_argument("--verbose", action='store_true', help="Verbose print for debug") args = parser.parse_args() if args.verbose: def v_print(*args, **kwargs): print(*args, **kwargs) else: def v_print(*args, **kwargs): return block_timestamps = [] block_tx = [] block_stx = [] for block_num in range(args.start, args.end): v_print(f'Block {block_num}/{args.end}', end="\r") reply = get_block_by_num(block_num, args.endpoint) try: block_timestamps.append(int(reply['result']['timestamp'], 0)) block_tx.append(len(reply['result']['transactions'])) block_stx.append(len(reply['result']['stakingTransactions'])) except Exception as e: v_print(f'{e.__class__}: {e}') pass block_times = [y - x for x, y in zip(block_timestamps, block_timestamps[1:])] avg = sum(block_times) / len(block_times) print(f'Average Block Time: {avg}') unique_times = Counter(block_times) print(f'Unique block times: {unique_times.most_common()}') # offset = [0].extend(block_times)
2,206
717
import re valid = ['hcl', 'iyr', 'pid', 'ecl', 'hgt','eyr', 'byr' ] def check_passport_part1(buffer): passport = map(lambda x: x.split(':')[0], buffer.split(' ')) for v in valid: if v not in passport: return False return True def split_passport(buffer): dictionary = {} for fields in buffer.split(' '): if fields != '': f = fields.split(':') dictionary[f[0]] = f[1] return dictionary def check_value(field, value): if field == 'byr': return int(value) >= 1920 and int(value) <= 2002 if field == 'iyr': return int(value) >= 2010 and int(value) <= 2020 if field == 'eyr': return int(value) >= 2020 and int(value) <= 2030 if field == 'hgt': if 'cm' in value: h = value[:value.index('cm')] return int(h) >= 150 and int(h) <= 193 if 'in' in value: h = value[:value.index('in')] return int(h) >= 59 and int(h) <= 76 return False if field == 'hcl': return re.match("#[0-9a-f]{6}", value) != None if field == 'ecl': colors = ['amb', 'blu', 'brn', 'gry', 'grn', 'hzl', 'oth'] return colors.count(value) == 1 if field == 'pid': return re.match("\d{9}", value) != None def check_passport_part2(passport): for v in valid: if not passport.has_key(v): return False if not check_value(v, passport.get(v)): return False return True def part2(): total = 0 with open("input.txt", "r") as f: buffer = '' for line in f: if line != '\n': buffer = line.rstrip() + ' ' + buffer else: passport = split_passport(buffer) if check_passport_part2(passport): total = total + 1 buffer = '' print (total) def part1(): total = 0 with open("input.txt", "r") as f: buffer = '' for line in f: if line != '\n': buffer = line.rstrip() + ' ' + buffer else: if check_passport_part1(buffer): total = total + 1 buffer = '' print (total) #part1() part2()
2,271
754
from django.db import models from django.contrib.auth import get_user_model from django.contrib.postgres.fields import ArrayField from utils.model_utils import default_1d_array_of_strings class SeoModel(models.Model): seo_title = models.CharField("SEO title", max_length=100, blank=True, null=True) seo_description = models.TextField("SEO description", max_length=400, blank=True, null=True) seo_keywords = models.CharField("SEO keywords", max_length=200, blank=True, null=True) class Meta: abstract = True
535
171
''' symmetrical-synthesis Copyright (c) 2020-present NAVER Corp. MIT license ''' import os import time import glob import cv2 import random import numpy as np import tensorflow as tf import random try: import data_util except ImportError: from dataset import data_util tf.app.flags.DEFINE_boolean('random_resize', False, 'True or False') tf.app.flags.DEFINE_boolean('past_dataset', False, 'True or False') tf.app.flags.DEFINE_string('google_path', None, '') tf.app.flags.DEFINE_integer('min_train3', 2, '') tf.app.flags.DEFINE_string('match_info', None, '') tf.app.flags.DEFINE_float('match_prob', 0.0, '') tf.app.flags.DEFINE_boolean('mnist_mode', False, '') FLAGS = tf.app.flags.FLAGS ''' image_path = '/where/your/images/*.jpg' ''' def load_image(im_fn, input_size=224): org_image = cv2.imread(im_fn, cv2.IMREAD_IGNORE_ORIENTATION | cv2.IMREAD_COLOR)[:,:,::-1] # rgb converted ''' if FLAGS.random_resize: resize_table = [0.5, 1.0, 1.5, 2.0] selected_scale = np.random.choice(resize_table, 1)[0] shrinked_hr_size = int(hr_size / selected_scale) h, w, _ = high_image.shape if h <= shrinked_hr_size or w <= shrinked_hr_size: high_image = cv2.resize(high_image, (hr_size, hr_size)) else: h_edge = h - shrinked_hr_size w_edge = w - shrinked_hr_size h_start = np.random.randint(low=0, high=h_edge, size=1)[0] w_start = np.random.randint(low=0, high=w_edge, size=1)[0] high_image_crop = high_image[h_start:h_start+hr_size, w_start:w_start+hr_size, :] high_image = cv2.resize(high_image_crop, (hr_size, hr_size)) ''' h, w, _ = org_image.shape min_len = np.min([h, w]) # center crop margin, we follow the method, which was introduced in DELF paper. if FLAGS.mnist_mode: crop_image = org_image.copy() else: try: cc_margin = np.random.randint(low=1, high=int(min_len * 0.05), size=1)[0] crop_image = org_image[cc_margin:-cc_margin, cc_margin:-cc_margin, :].copy() except: crop_image = org_image.copy() new_input_size = int(input_size * 1.125) crop_image = cv2.resize(crop_image, (new_input_size, new_input_size), interpolation=cv2.INTER_AREA) # random crop range h_edge = new_input_size - input_size#32#256 - input_size # input_size is 224 w_edge = new_input_size - input_size#256 - input_size h_start = np.random.randint(low=0, high=h_edge, size=1)[0] w_start = np.random.randint(low=0, high=w_edge, size=1)[0] return_image = crop_image[h_start:h_start+input_size, w_start:w_start+input_size,:] # flip lr if random.randint(0, 1): return_image = return_image[:,::-1,:] #print('return', return_image.shape) return return_image #high_image, low_image def get_images_dict(image_folder): ''' image_folder = '/data/IR/DB/sid_images' folder structure sid_images - sid0 - image00.png, image01.png, ... - sid1 - ... - sid2 - ... ''' if FLAGS.match_info is not None: match_dict = {} f_match = open(FLAGS.match_info, 'r') match_lines = f_match.readlines() cnt = 0 for match_line in match_lines: ver1_cls, ver2_cls, prob = match_line.split() prob = float(prob) if prob >= FLAGS.match_prob: match_dict[ver2_cls] = 1 possible_image_type = ['jpg', 'JPG', 'png', 'JPEG', 'jpeg'] sid_list = glob.glob(os.path.join(image_folder, '*')) images_dict = {} images_list = [] images_cnt = 0 sid_idx = 0 for sid_folder in sid_list: ext_folder = sid_folder #ext_folder = os.path.join(sid_folder, 'exterior') images_path = [image_path for image_paths in [glob.glob(os.path.join(ext_folder, '*.%s' % ext)) for ext in possible_image_type] for image_path in image_paths] n_instance = 2 if len(images_path) < n_instance: continue for image_path in images_path: images_list.append([image_path, sid_idx]) images_dict[sid_idx] = images_path images_cnt += len(images_path) sid_idx += 1 #print(images_dict) stat_db = {} stat_db['num_sid'] = len(images_dict) stat_db['images_cnt'] = images_cnt return images_dict, stat_db, images_list def get_record(image_folder, input_size, batch_size): images_dict, stat_db, images_list = get_images_dict(image_folder) print('place total sids: %d, total images: %d' % (stat_db['num_sid'], stat_db['images_cnt'])) if FLAGS.google_path is not None: images_dict_google, stat_db_google, images_list_google = get_images_dict(FLAGS.google_path) print('google total sids: %d, total images: %d' % (stat_db_google['num_sid'], stat_db_google['images_cnt'])) #time.sleep(3) n_instance = 2 b_replace = False real_batch_size = batch_size // n_instance while True: try: gt_labels = np.random.choice(len(images_dict), real_batch_size, replace=b_replace) anchor_images = [] pos_images = [] for n in range(n_instance - 1): pos_images.append([]) for label in gt_labels: tmp_image_list = images_dict[label] image_index = np.random.choice(len(tmp_image_list), n_instance, replace=False) anchor_image = load_image(tmp_image_list[image_index[0]], input_size) anchor_images.append(anchor_image) for n, ind in enumerate(image_index[1:]): pos_image = load_image(tmp_image_list[ind], input_size) pos_images[n].append(pos_image) #print(len(gt_labels)) if n_instance == 2: pos_images = pos_images[0] elif n_instance == 1: pos_images = pos_images else: pos_images = np.concatenate(pos_images, axis=0) yield anchor_images, pos_images, gt_labels #im_fn, gt_label except Exception as e: print(e) continue def generator(image_folder, input_size=224, batch_size=32): for anchor_images, pos_images, gt_labels in get_record(image_folder, input_size, batch_size): yield anchor_images, pos_images, gt_labels def get_generator(image_folder, **kwargs): return generator(image_folder, **kwargs) ## image_path = '/where/is/your/images/' def get_batch(image_path, num_workers, **kwargs): try: generator = get_generator(image_path, **kwargs) enqueuer = data_util.GeneratorEnqueuer(generator, use_multiprocessing=True) enqueuer.start(max_queue_size=24, workers=num_workers) generator_ouptut = None while True: while enqueuer.is_running(): if not enqueuer.queue.empty(): generator_output = enqueuer.queue.get() break else: time.sleep(0.001) yield generator_output generator_output = None finally: if enqueuer is not None: enqueuer.stop() if __name__ == '__main__': image_path = '/data/IR/DB/data_refinement/place_exterior' num_workers = 4 batch_size = 128 input_size = 224 data_generator = get_batch(image_path=image_path, num_workers=num_workers, batch_size=batch_size, input_size=224) _ = 0 while True: _ += 1 #break start_time = time.time() data = next(data_generator) anchor_images = np.asarray(data[0]) pos_images = np.asarray(data[1]) gts = np.asarray(data[2]) print('%d done!!! %f' % (_, time.time() - start_time), anchor_images.shape, pos_images.shape, gts.shape) #for sub_idx, (loaded_image, gt) in enumerate(zip(loaded_images, gts)): # save_path = '/data/IR/DB/naver_place/test/%03d_%03d_gt_%d_image.jpg' % (_, sub_idx, gt) # cv2.imwrite(save_path, loaded_image[:,:,::-1])
8,314
2,872
# -*- coding: utf-8 -*- """ Created on Sat Jun 17 20:55:53 2019 @author: Parikshith.H """ import sqlite3 conn=sqlite3.connect('music.sqlite') cur=conn.cursor() cur.execute('DROP TABLE IF EXISTS Tracks') cur.execute('CREATE TABLE Tracks(title TEXT,plays INTEGER)') cur.execute('''INSERT INTO Tracks(title,plays) VALUES ('Thunder2',100)''') cur.execute('''INSERT INTO Tracks VALUES ('Thunder3',100)''') cur.execute('INSERT INTO Tracks(title,plays) VALUES (?,?)',('Thunderstuck',200)) cur.execute('INSERT INTO Tracks(title,plays) VALUES (?,?)',('Dangerous',20)) cur.execute('INSERT INTO Tracks(title,plays) VALUES (?,?)',('Myway',150)) cur.execute('INSERT INTO Tracks(title,plays) VALUES (?,?)',('Newway',30)) cur.execute('SELECT * FROM Tracks') for row in cur: print(row) print('****************************') cur.execute('''UPDATE Tracks SET plays=50 WHERE title='Myway' ''') cur.execute('SELECT * FROM Tracks') for row in cur: print(row) print('****************************') cur.execute('''DELETE FROM Tracks WHERE plays<100 ''') cur.execute('SELECT * FROM Tracks') for row in cur: print(row) cur.close() conn.close() # ============================================================================= # #output: # ('Thunder2', 100) # ('Thunder3', 100) # ('Thunderstuck', 200) # ('Dangerous', 20) # ('Myway', 150) # ('Newway', 30) # **************************** # ('Thunder2', 100) # ('Thunder3', 100) # ('Thunderstuck', 200) # ('Dangerous', 20) # ('Myway', 50) # ('Newway', 30) # **************************** # ('Thunder2', 100) # ('Thunder3', 100) # ('Thunderstuck', 200) # =============================================================================
1,678
619
#!/usr/bin/python from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.externalpkg import extmod def main(): module = AnsibleModule(argument_spec={}) module.exit_json(extmod_path=extmod.path()) if __name__ == '__main__': main()
269
89
from ..grammar import * from .Metalanguage import Metalanguage class EBNFTerminal(Terminal): def __init__(self, subject, left_bound='"', right_bound='"'): super().__init__(subject, left_bound=left_bound, right_bound=right_bound) class EBNFDefinitionList(DefinitionList): def __init__(self, definitions, alternation='|'): super().__init__(definitions, alternation=alternation) class EBNFNonTerminal(NonTerminal): def __init__(self, subject, left_bound='', right_bound=''): super().__init__(subject, left_bound=left_bound, right_bound=right_bound) class EBNFConcat(Sequence): def __init__(self, terms, separator=', '): super().__init__(terms, separator=separator) class EBNFRule(Rule): def __init__(self, left, right, production='=', terminator=';'): super().__init__(left, right, production=production, terminator=terminator) class EBNFRepetition(Repetition): def __init__(self, subject, left_bound='{', right_bound='}'): super().__init__(subject, left_bound=left_bound, right_bound=right_bound) class EBNFSpecialSequence(Bracket): def __init__(self, subject, left_bound='?', right_bound='?'): super().__init__(subject, left_bound=left_bound, right_bound=right_bound) class EBNF(Metalanguage): def __init__(self, ruleset: Ruleset, normalise=False): super().__init__(ruleset, syntax_dict={ # Core Sequence: Concat, DefinitionList: EBNFDefinitionList, Rule: EBNFRule, Terminal: EBNFTerminal, NonTerminal: EBNFNonTerminal, # Auxiliary Optional: Optional, Group: Group, Repetition: EBNFRepetition, # Additional Except: Except }, normalise=normalise)
1,869
564
from sqlalchemy.orm import Session from db.crud.jobs import create_new_job, retrieve_job from schemas.jobs import JobCreate from tests.utils.users import create_random_owner from tests.utils.jobs import create_sample_job def test_retrieve_job_by_id(db_session: Session): """Test retrieving job from db""" owner = create_random_owner(session=db_session) job = create_sample_job(owner, db_session) retrieved_job = retrieve_job(job_id=job.id, session=db_session) assert retrieved_job.id == job.id assert retrieved_job.title == job.title
560
180
from __future__ import unicode_literals import pytest import spacy import os try: xrange except NameError: xrange = range @pytest.fixture() def token(doc): return doc[0] @pytest.mark.models def test_load_resources_and_process_text(): from spacy.en import English nlp = English() doc = nlp(u'Hello, world. Here are two sentences.') @pytest.mark.models def test_get_tokens_and_sentences(doc): token = doc[0] sentence = next(doc.sents) assert token is sentence[0] assert sentence.text == 'Hello, world.' @pytest.mark.models def test_use_integer_ids_for_any_strings(nlp, token): hello_id = nlp.vocab.strings['Hello'] hello_str = nlp.vocab.strings[hello_id] assert token.orth == hello_id == 3125 assert token.orth_ == hello_str == 'Hello' def test_get_and_set_string_views_and_flags(nlp, token): assert token.shape_ == 'Xxxxx' for lexeme in nlp.vocab: if lexeme.is_alpha: lexeme.shape_ = 'W' elif lexeme.is_digit: lexeme.shape_ = 'D' elif lexeme.is_punct: lexeme.shape_ = 'P' else: lexeme.shape_ = 'M' assert token.shape_ == 'W' def test_export_to_numpy_arrays(nlp, doc): from spacy.attrs import ORTH, LIKE_URL, IS_OOV attr_ids = [ORTH, LIKE_URL, IS_OOV] doc_array = doc.to_array(attr_ids) assert doc_array.shape == (len(doc), len(attr_ids)) assert doc[0].orth == doc_array[0, 0] assert doc[1].orth == doc_array[1, 0] assert doc[0].like_url == doc_array[0, 1] assert list(doc_array[:, 1]) == [t.like_url for t in doc] @pytest.mark.models def test_word_vectors(nlp): doc = nlp("Apples and oranges are similar. Boots and hippos aren't.") apples = doc[0] oranges = doc[2] boots = doc[6] hippos = doc[8] assert apples.similarity(oranges) > boots.similarity(hippos) @pytest.mark.models def test_part_of_speech_tags(nlp): from spacy.parts_of_speech import ADV def is_adverb(token): return token.pos == spacy.parts_of_speech.ADV # These are data-specific, so no constants are provided. You have to look # up the IDs from the StringStore. NNS = nlp.vocab.strings['NNS'] NNPS = nlp.vocab.strings['NNPS'] def is_plural_noun(token): return token.tag == NNS or token.tag == NNPS def print_coarse_pos(token): print(token.pos_) def print_fine_pos(token): print(token.tag_) @pytest.mark.models def test_syntactic_dependencies(): def dependency_labels_to_root(token): '''Walk up the syntactic tree, collecting the arc labels.''' dep_labels = [] while token.head is not token: dep_labels.append(token.dep) token = token.head return dep_labels @pytest.mark.models def test_named_entities(): def iter_products(docs): for doc in docs: for ent in doc.ents: if ent.label_ == 'PRODUCT': yield ent def word_is_in_entity(word): return word.ent_type != 0 def count_parent_verb_by_person(docs): counts = defaultdict(defaultdict(int)) for doc in docs: for ent in doc.ents: if ent.label_ == 'PERSON' and ent.root.head.pos == VERB: counts[ent.orth_][ent.root.head.lemma_] += 1 return counts def test_calculate_inline_mark_up_on_original_string(): def put_spans_around_tokens(doc, get_classes): '''Given some function to compute class names, put each token in a span element, with the appropriate classes computed. All whitespace is preserved, outside of the spans. (Yes, I know HTML won't display it. But the point is no information is lost, so you can calculate what you need, e.g. <br /> tags, <p> tags, etc.) ''' output = [] template = '<span classes="{classes}">{word}</span>{space}' for token in doc: if token.is_space: output.append(token.orth_) else: output.append( template.format( classes=' '.join(get_classes(token)), word=token.orth_, space=token.whitespace_)) string = ''.join(output) string = string.replace('\n', '') string = string.replace('\t', ' ') return string @pytest.mark.models def test_efficient_binary_serialization(doc): from spacy.tokens.doc import Doc byte_string = doc.to_bytes() open('moby_dick.bin', 'wb').write(byte_string) nlp = spacy.en.English() for byte_string in Doc.read_bytes(open('moby_dick.bin', 'rb')): doc = Doc(nlp.vocab) doc.from_bytes(byte_string) @pytest.mark.models def test_multithreading(nlp): texts = [u'One document.', u'...', u'Lots of documents'] # .pipe streams input, and produces streaming output iter_texts = (texts[i % 3] for i in xrange(100000000)) for i, doc in enumerate(nlp.pipe(iter_texts, batch_size=50, n_threads=4)): assert doc.is_parsed if i == 100: break
5,135
1,755
import os from copy import deepcopy import matplotlib.pyplot as plt import numpy as np from bridge_sim import model, sim, temperature, traffic, plot, util from bridge_sim.model import Config, Point, Bridge from bridge_sim.plot.util import equal_lims from bridge_sim.sim.responses import without from bridge_sim.util import print_i, print_w from bridge_sim.internal.plot import axis_cmap_r def plot_year_effects(config: Config, x: float, z: float, num_years: int): """Plot all effects over a single year and 100 years at a point.""" install_day = 37 year = 2018 weather = temperature.load("holly-springs-18") _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), 60 * 10 ) ( ll_responses, ps_responses, temp_responses, shrinkage_responses, creep_responses, ) = np.repeat(None, 5) start_day, end_day = None, None def set_responses(n): nonlocal weather, start_day, end_day weather["temp"] = temperature.resize(weather["temp"], year=year) weather = temperature.repeat(config, "holly-springs-18", weather, n) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) start_day, end_day = install_day, 365 * n nonlocal ll_responses, ps_responses, temp_responses, shrinkage_responses, creep_responses ( ll_responses, ps_responses, temp_responses, shrinkage_responses, creep_responses, ) = sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=model.RT.YTrans, with_creep=True, weather=weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, ret_all=True, ) # from sklearn.decomposition import FastICA, PCA # ica = FastICA(n_components=3) # try_ = ica.fit_transform((ll_responses + temp_responses + creep_responses + shrinkage_responses).T) # plt.plot(try_) # plt.show() plt.landscape() lw = 2 def legend(): leg = plt.legend( facecolor="white", loc="upper right", framealpha=1, fancybox=False, borderaxespad=0, ) for legobj in leg.legendHandles: legobj.set_linewidth(lw) plt.subplot(1, 2, 1) set_responses(1) xax = np.interp( np.arange(len(traffic_array)), [0, len(traffic_array) - 1], [start_day, end_day] ) plt.plot(xax, ll_responses[0] * 1e3, c="green", label="traffic", lw=lw) plt.plot(xax, temp_responses[0] * 1e3, c="red", label="temperature") plt.plot(xax, shrinkage_responses[0] * 1e3, c="blue", label="shrinkage", lw=lw) plt.plot(xax, creep_responses[0] * 1e3, c="black", label="creep", lw=lw) legend() plt.ylabel("Y translation (mm)") plt.xlabel("Time (days)") plt.subplot(1, 2, 2) end_day = 365 * num_years set_responses(num_years) xax = ( np.interp( np.arange(len(traffic_array)), [0, len(traffic_array) - 1], [start_day, end_day], ) / 365 ) plt.plot(xax, ll_responses[0] * 1e3, c="green", label="traffic", lw=lw) plt.plot(xax, temp_responses[0] * 1e3, c="red", label="temperature") plt.plot(xax, shrinkage_responses[0] * 1e3, c="blue", label="shrinkage", lw=lw) plt.plot(xax, creep_responses[0] * 1e3, c="black", label="creep", lw=lw) legend() plt.ylabel("Y translation (mm)") plt.xlabel("Time (years)") equal_lims("y", 1, 2) plt.suptitle(f"Y translation at X = {x} m, Z = {z} m") plt.tight_layout(rect=[0, 0.03, 1, 0.95]) plt.savefig(config.get_image_path("classify/ps", f"year-effect-{x}-{z}.png")) def plot_sensor_placement(config: Config, num_years: int): all_points = [ model.Point(x=x, z=z) for x in np.linspace(config.bridge.x_min, config.bridge.x_max, 300) for z in np.linspace(config.bridge.z_min, config.bridge.z_max, 100) ] response_type = model.ResponseType.YTrans install_day = 37 year = 2018 weather = temperature.load("holly-springs-18") config.sensor_freq = 1 _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), 10 ) weather["temp"] = temperature.resize(weather["temp"], year=year) weather = temperature.repeat(config, "holly-springs-18", weather, num_years) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) start_day, end_day = install_day, 365 * num_years for pier in [9]: pier_centre = model.Point( x=config.bridge.supports[pier].x, z=config.bridge.supports[pier].z, ) points = [p for p in all_points if pier_centre.distance(p) < 7] ps = model.PierSettlement(pier=pier, settlement=5 / 1e3) ( _0, _1, temp_responses, shrinkage_responses, creep_responses, ) = sim.responses.to( config=config, points=points, traffic_array=traffic_array, response_type=response_type, with_creep=True, weather=weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, ret_all=True, ) ps_responses = sim.responses.to_pier_settlement( config=config, points=points, responses_array=_0, response_type=response_type, pier_settlement=[(ps, ps)], ).T[-1] ps_responses += sim.responses.to_creep( config=config, points=points, responses_array=_0, response_type=response_type, pier_settlement=[(ps, ps)], install_pier_settlement=[ps], install_day=install_day, start_day=start_day, end_day=end_day, ).T[-1] long_term_responses = ( temp_responses.T[-1] + shrinkage_responses.T[-1] + creep_responses.T[-1] ) ############ # Plotting # ############ plt.landscape() plt.subplot(3, 1, 1) responses = sim.model.Responses( response_type=response_type, responses=list(zip(abs(long_term_responses) * 1e3, points)), ) plot.contour_responses(config, responses, levels=30, interp=(200, 60)) plot.top_view_bridge(config.bridge, piers=True) plt.subplot(3, 1, 2) responses = sim.model.Responses( response_type=response_type, responses=list(zip(abs(ps_responses) * 1e3, points)), ) plot.contour_responses(config, responses, levels=30, interp=(200, 60)) plot.top_view_bridge(config.bridge, piers=True) plt.subplot(3, 1, 3) responses = sim.model.Responses( response_type=response_type, responses=list( zip((abs(ps_responses) - abs(long_term_responses)) * 1e3, points) ), ) plot.contour_responses(config, responses, levels=30, interp=(200, 60)) plot.top_view_bridge(config.bridge, piers=True) plt.savefig(config.get_image_path("classify/ps", "placement.pdf")) def plot_removal(config: Config, x: float, z: float): response_type = model.RT.YTrans weather = temperature.load("holly-springs-18") weather["temp"] = temperature.resize(weather["temp"], year=2018) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) install_day = 37 start_day, end_day = install_day, install_day + 365 _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), time=60 ) responses = ( sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, weather=weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, # ret_all=True, )[0] * 1e3 ) def legend(): return plt.legend( facecolor="white", loc="upper right", framealpha=1, fancybox=False, borderaxespad=0, ) plt.landscape() plt.subplot(2, 2, 1) xax = np.interp( np.arange(len(weather)), [0, len(weather) - 1], [start_day, end_day] ) plt.plot(xax, weather["temp"], c="red") plt.ylabel("Temperature °C") plt.xlabel("Days since T_0") plt.title("Temperature in 2018") plt.subplot(2, 2, 2) xax = np.interp( np.arange(len(responses)), [0, len(responses) - 1], [start_day, end_day] ) plt.plot(xax, responses) plt.ylabel("Y translation (mm)") plt.xlabel("Days since T_0") plt.title("Y translation in 2018") plt.subplot(2, 2, 3) num_samples = 365 * 24 temps = util.apply(weather["temp"], np.arange(num_samples)) rs = util.apply(responses, np.arange(num_samples)) lr, _ = temperature.regress_and_errors(temps, rs) lr_x = np.linspace(min(temps), max(temps), 100) y = lr.predict(lr_x.reshape((-1, 1))) plt.plot(lr_x, y, lw=2, c="red", label="linear fit") plt.scatter(temps, rs, s=2, alpha=0.5, label="hourly samples") leg = legend() leg.legendHandles[1]._sizes = [30] plt.ylabel("Y translation (mm)") plt.xlabel("Temperature °C") plt.title("Linear model from 2018 data") ############# # 2019 data # ############# weather_2019 = temperature.load("holly-springs") weather_2019["temp"] = temperature.resize(weather_2019["temp"], year=2019) start_date, end_date = ( weather_2019["datetime"].iloc[0].strftime(temperature.f_string), weather_2019["datetime"].iloc[-1].strftime(temperature.f_string), ) start_day, end_day = install_day + 365, install_day + (2 * 365) responses_2019 = ( sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, weather=weather_2019, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, )[0] * 1e3 ) plt.subplot(2, 2, 4) xax_responses = np.interp( np.arange(len(responses_2019)), [0, len(responses_2019) - 1], [start_day, end_day], ) plt.plot(xax_responses, responses_2019, label="2019 responses") temps_2019 = util.apply(weather_2019["temp"], xax_responses) y = lr.predict(temps_2019.reshape((-1, 1))) plt.plot(xax_responses, y, label="prediction") plt.ylabel("Y translation (mm)") plt.xlabel("Days since T_0") plt.title("Y translation in 2019") for legobj in legend().legendHandles: legobj.set_linewidth(2.0) plt.suptitle(f"Predicting long-term effect at X = {x} m, Z = {z} m") plt.tight_layout(rect=[0, 0.03, 1, 0.95]) plt.savefig(config.get_image_path("classify/ps", "regress.pdf")) def plot_removal_2(config: Config, x: float, z: float): response_type = model.RT.YTrans weather_2018 = temperature.load("holly-springs-18") weather_2018["temp"] = temperature.resize(weather_2018["temp"], year=2018) start_date, end_date = ( weather_2018["datetime"].iloc[0].strftime(temperature.f_string), weather_2018["datetime"].iloc[-1].strftime(temperature.f_string), ) install_day = 37 start_day, end_day = install_day, install_day + 365 _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), time=60 ) responses_2018 = ( sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, weather=weather_2018, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, # ret_all=True, )[0] * 1e3 ) num_samples = 365 * 24 temps = util.apply(weather_2018["temp"], np.arange(num_samples)) rs = util.apply(responses_2018, np.arange(num_samples)) lr, err = temperature.regress_and_errors(temps, rs) def legend(): plt.legend( facecolor="white", loc="lower left", framealpha=1, fancybox=False, borderaxespad=0, labelspacing=0.02, ) ############################## # Iterate through each year. # ############################## plt.landscape() weather_2019 = temperature.load("holly-springs") weather_2019["temp"] = temperature.resize(weather_2019["temp"], year=2019) start_date, end_date = ( weather_2019["datetime"].iloc[0].strftime(temperature.f_string), weather_2019["datetime"].iloc[-1].strftime(temperature.f_string), ) for y_i, year in enumerate([2019, 2024, 2039]): plt.subplot(3, 1, y_i + 1) start_day = install_day + ((year - 2018) * 365) end_day = start_day + 365 responses_2019 = ( sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, weather=weather_2019, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, )[0] * 1e3 ) # Plot actual values. xax = np.interp( np.arange(len(responses_2019)), [0, len(responses_2019) - 1], [0, 364] ) plt.plot(xax, responses_2019, label="responses in year", lw=2) # Daily prediction. xax_responses = np.arange(365) temps_2019 = util.apply(weather_2019["temp"], xax_responses) y_daily = lr.predict(temps_2019.reshape((-1, 1))) y_2_week = [ np.mean(y_daily[max(0, i - 14) : min(i + 14, len(y_daily))]) for i in range(len(y_daily)) ] for percentile, alpha in [(100, 20), (75, 40), (50, 60), (25, 100)]: err = np.percentile(err, percentile) p = percentile / 100 plt.fill_between( xax_responses, y_2_week + (err * p), y_2_week - (err * p), color="orange", alpha=alpha / 100, label=f"{percentile}% of regression error", ) plt.plot(xax_responses, y_daily, color="black", lw=2, label="daily prediction") plt.plot( xax_responses, y_2_week, color="red", lw=2, label="2 week sliding window" ) plt.ylabel("Y. trans (mm)") plt.title(f"Year {year}") if y_i == 0: legend() if y_i == 2: plt.xlabel("Days in year") else: plt.tick_params("x", bottom=False, labelbottom=False) equal_lims("y", 3, 1) plt.suptitle(f"Predicting long-term effects at X = {x} m, Z = {z} m") plt.tight_layout(rect=[0, 0.03, 1, 0.95]) plt.savefig(config.get_image_path("classify/ps", "regress-2.pdf")) def plot_removal_3(config: Config, x: float, z: float): # First calculate the linear model. response_type = model.RT.YTrans weather_2018 = temperature.load("holly-springs-18") weather_2018["temp"] = temperature.resize(weather_2018["temp"], year=2018) start_date, end_date = ( weather_2018["datetime"].iloc[0].strftime(temperature.f_string), weather_2018["datetime"].iloc[-1].strftime(temperature.f_string), ) install_day = 37 start_day, end_day = install_day, install_day + 365 _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), time=60 ) responses_2018 = ( sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, weather=weather_2018, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, )[0] * 1e3 ) num_samples = 365 * 24 temps = util.apply(weather_2018["temp"], np.arange(num_samples)) rs = util.apply(responses_2018, np.arange(num_samples)) lr, _ = temperature.regress_and_errors(temps, rs) # Calculate long-term weather. NUM_YEARS = 5 PIER = 5 long_weather = deepcopy(weather_2018) long_weather["temp"] = temperature.resize(long_weather["temp"], year=2019) print_i(f"Repeating {NUM_YEARS} of weather data") long_weather = temperature.repeat( config, "holly-springs-18", long_weather, NUM_YEARS ) print_i(f"Repeated {NUM_YEARS} of weather data") start_date, end_date = ( long_weather["datetime"].iloc[0].strftime(temperature.f_string), long_weather["datetime"].iloc[-1].strftime(temperature.f_string), ) start_day = install_day + 365 end_day = start_day + 365 * NUM_YEARS MAX_PS = 20 THRESHES = np.arange(0, MAX_PS, 1) acc_mat = np.zeros((MAX_PS, len(THRESHES))) fp_mat = np.zeros(acc_mat.shape) fn_mat = np.zeros(acc_mat.shape) tp_mat = np.zeros(acc_mat.shape) tn_mat = np.zeros(acc_mat.shape) for p_i, ps in enumerate(range(MAX_PS)): print_i(f"Using pier settlement = {ps} mm") long_responses = sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, pier_settlement=[ ( model.PierSettlement(pier=PIER, settlement=0.00001), model.PierSettlement(pier=PIER, settlement=ps / 1e3), ) ], install_pier_settlement=[], weather=long_weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, ret_all=False, ignore_pier_creep=True, ) healthy_responses = sim.responses.to( config=config, points=[model.Point(x=x, z=z)], traffic_array=traffic_array, response_type=response_type, with_creep=True, pier_settlement=[], install_pier_settlement=None, weather=long_weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, ret_all=False, ignore_pier_creep=True, ) plt.plot(healthy_responses[0] * 1e3, label="healthy") plt.plot(long_responses[0] * 1e3, label="pier settlement") plt.legend() plt.savefig(config.get_image_path("hello", f"q3-{p_i}.png")) plt.close() for t_i, thresh in enumerate(THRESHES): thresh *= -1 print(thresh) print(max(healthy_responses[0])) print(min(healthy_responses[0])) print(max(long_responses[0])) print(min(long_responses[0])) fp = len([x for x in healthy_responses[0] * 1e3 if x <= thresh]) tp = len([x for x in long_responses[0] * 1e3 if x <= thresh]) tn = len([x for x in healthy_responses[0] * 1e3 if x > thresh]) fn = len([x for x in long_responses[0] * 1e3 if x > thresh]) acc_mat[p_i][t_i] = (tp + tn) / (tp + tn + fp + fn) fp_mat[p_i][t_i] = fp tp_mat[p_i][t_i] = tp fn_mat[p_i][t_i] = fn tn_mat[p_i][t_i] = tn ################## # Save matrices. # ################## plt.imshow(acc_mat, cmap=axis_cmap_r) plt.savefig(config.get_image_path("hello", f"mat.png")) plt.close() plt.imshow(fp_mat, cmap=axis_cmap_r) plt.savefig(config.get_image_path("hello", f"mat-fp.png")) plt.close() plt.imshow(fn_mat, cmap=axis_cmap_r) plt.savefig(config.get_image_path("hello", f"mat-fn.png")) plt.close() plt.imshow(tp_mat, cmap=axis_cmap_r) plt.savefig(config.get_image_path("hello", f"mat-tp.png")) plt.close() plt.imshow(tn_mat, cmap=axis_cmap_r) plt.savefig(config.get_image_path("hello", f"mat-tn.png")) plt.close() def support_with_points(bridge: Bridge, delta_x: float): for support in bridge.supports: if support.x < bridge.length / 2: s_x = support.x - ((support.length / 2) + delta_x) else: s_x = support.x + ((support.length / 2) + delta_x) support.point = Point(x=s_x, z=support.z) for support_2 in bridge.supports: if support_2.z == support.z and np.isclose( support_2.x, bridge.length - support.x ): support.opposite_support = support_2 print_w(f"Support sensor at X = {support.point.x}, Z = {support.point.z}") if not hasattr(support, "opposite_support"): raise ValueError("No opposite support") return bridge.supports def plot_min_diff(config: Config, num_years: int, delta_x: float = 0.5): plt.landscape() log_path = config.get_image_path("classify/q1", "min-thresh.txt") if os.path.exists(log_path): os.remove(log_path) install_day = 37 start_day, end_day = install_day, 365 * num_years year = 2018 weather = temperature.load("holly-springs-18") _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), 60 * 10 ) weather["temp"] = temperature.resize(weather["temp"], year=year) # weather = temperature.repeat(config, "holly-springs-18", weather, num_years) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) # For each support load the responses to traffic and assign to "Support". for s_i, support in enumerate(support_with_points(config.bridge, delta_x=delta_x)): support.responses = ( sim.responses.to_traffic_array( config=config, points=[support.point], traffic_array=traffic_array, response_type=model.RT.YTrans, # with_creep=True, # weather=weather, # start_date=start_date, # end_date=end_date, # install_day=install_day, # start_day=start_day, # end_day=end_day, )[0] * 1e3 ) # Determine max difference for each sensor pair. for s_i, support in enumerate(config.bridge.supports): min1, max1 = min(support.responses), max(support.responses) min2, max2 = ( min(support.opposite_support.responses), max(support.opposite_support.responses), ) delta_1, delta_2 = abs(min1 - max2), abs(min2 - max1) # max_delta = max(abs(support.responses - support.opposite_support.responses)) support.max_delta = max(delta_1, delta_2) to_write = f"Max delta {support.max_delta} for support {s_i}, sensor at X = {support.point.x}, Z = {support.point.z}" with open(log_path, "a") as f: f.write(to_write) # Bridge supports. plot.top_view_bridge(config.bridge, lanes=True, piers=True, units="m") for s_i, support in enumerate(config.bridge.supports): if s_i % 4 == 0: support.max_delta = max( support.max_delta, config.bridge.supports[s_i + 3].max_delta ) elif s_i % 4 == 1: support.max_delta = max( support.max_delta, config.bridge.supports[s_i + 1].max_delta ) elif s_i % 4 == 2: support.max_delta = max( support.max_delta, config.bridge.supports[s_i - 1].max_delta ) elif s_i % 4 == 3: support.max_delta = max( support.max_delta, config.bridge.supports[s_i - 3].max_delta ) plt.scatter([support.point.x], [support.point.z], c="red") plt.annotate( f"{np.around(support.max_delta, 2)} mm", xy=(support.point.x - 3, support.point.z + 2), color="b", size="large", ) plt.title("Maximum difference between symmetric sensors") plt.tight_layout() plt.savefig(config.get_image_path("classify/q1", "min-thresh.pdf")) def plot_contour_q2(config: Config, num_years: int, delta_x: float = 0.5): # Select points: over the deck and the sensors! points = [ Point(x=x, z=z) for x in np.linspace(config.bridge.x_min, config.bridge.x_max, 100) for z in np.linspace(config.bridge.z_min, config.bridge.z_max, 30) ] sensor_points = [ s.point for s in support_with_points(config.bridge, delta_x=delta_x) ] points += sensor_points install_day = 37 start_day, end_day = install_day, 365 * num_years year = 2018 weather = temperature.load("holly-springs-18") # Responses aren't much from traffic, and we are getting the maximum from 4 # sensors, so short traffic data doesn't really matter. _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), 10 ) weather["temp"] = temperature.resize(weather["temp"], year=year) # weather = temperature.repeat(config, "holly-springs-18", weather, num_years) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) # Generate the data! responses = ( sim.responses.to( config=config, points=points, traffic_array=traffic_array, response_type=model.RT.YTrans, with_creep=True, weather=weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, ) * 1e3 ) # Convert to Responses, determining maximum response per point. max_responses = [min(rs) for rs in responses] sensor_responses = max_responses[-len(sensor_points) :] responses = sim.model.Responses( response_type=model.RT.YTrans, responses=[(r, p) for r, p in zip(max_responses, points)], units="mm", ).without(without.edges(config, 2)) # Adjust maximum responses per sensor so they are symmetric! for s_i, support in enumerate(support_with_points(config.bridge, delta_x=delta_x)): support.max_response = sensor_responses[s_i] for support in support_with_points(config.bridge, delta_x=delta_x): support.max_response = min( support.max_response, support.opposite_support.max_response ) for s_i, support in enumerate(support_with_points(config.bridge, delta_x=delta_x)): if s_i % 4 == 0: support.max_response = max( support.max_response, config.bridge.supports[s_i + 3].max_response ) elif s_i % 4 == 1: support.max_response = max( support.max_response, config.bridge.supports[s_i + 1].max_response ) elif s_i % 4 == 2: support.max_response = max( support.max_response, config.bridge.supports[s_i - 1].max_response ) elif s_i % 4 == 3: support.max_response = max( support.max_response, config.bridge.supports[s_i - 3].max_response ) plt.landscape() plot.contour_responses(config, responses, interp=(200, 60), levels=20) plot.top_view_bridge(config.bridge, lanes=True, piers=True, units="m") for s_i, support in enumerate(support_with_points(config.bridge, delta_x=delta_x)): plt.scatter([support.point.x], [support.point.z], c="black") plt.annotate( f"{np.around(support.max_response, 2)}", xy=(support.point.x - 3, support.point.z + 2), color="black", size="large", ) plt.title( f"Minimum Y translation over {num_years} years \n from traffic, temperature, shrinkage & creep" ) plt.tight_layout() plt.savefig(config.get_image_path("classify/q2", "q2-contour.pdf")) plt.close() def plot_min_ps_1(config: Config, num_years: int, delta_x: float = 0.5): THRESH = 2 # Pier settlement from question 1. plt.landscape() log_path = config.get_image_path("classify/q1b", "min-ps.txt") if os.path.exists(log_path): # Start with fresh logfile. os.remove(log_path) install_day = 37 start_day, end_day = install_day, 365 * num_years year = 2018 weather = temperature.load("holly-springs-18") _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), 60 * 10 ) weather["temp"] = temperature.resize(weather["temp"], year=year) # weather = temperature.repeat(config, "holly-springs-18", weather, num_years) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) # For each support.. for s_i, support in enumerate(support_with_points(config.bridge, delta_x=delta_x)): # ..increase pier settlement until threshold triggered. for settlement in np.arange(0, 10, 0.1): responses = ( sim.responses.to( config=config, points=[support.point, support.opposite_support.point], traffic_array=traffic_array, response_type=model.RT.YTrans, with_creep=True, weather=weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, pier_settlement=[ ( model.PierSettlement(pier=s_i, settlement=0), model.PierSettlement(pier=s_i, settlement=settlement / 1e3), ) ], skip_weather_interp=True, ) * 1e3 ) delta = max(abs(responses[0] - responses[1])) to_write = f"Max delta {delta} for settlement {settlement} mm for support {s_i}, sensor at X = {support.point.x}, Z = {support.point.z}" print_w(to_write) # Because of "abs", "delta" will be positive. if delta > THRESH: break # Write the minimum settlement value for this support to a file. with open(log_path, "a") as f: f.write(to_write) # Annotate the support with the minimum settlement value. plt.scatter([support.point.x], [support.point.z], c="red") plt.annotate( f"{np.around(settlement, 2)} mm", xy=(support.point.x - 3, support.point.z + 2), color="b", size="large", ) # Plot the results. plot.top_view_bridge(config.bridge, lanes=True, piers=True, units="m") plt.title("Minimum pier settlement detected (Question 1B)") plt.tight_layout() plt.savefig(config.get_image_path("classify/q1b", "q1b-min-ps.pdf")) plt.close() def plot_min_ps_2(config: Config, num_years: int, delta_x: float = 0.5): THRESH = 6 # Pier settlement from question 1. plt.landscape() log_path = config.get_image_path("classify/q2b", "2b-min-ps.txt") if os.path.exists(log_path): # Start with fresh logfile. os.remove(log_path) install_day = 37 start_day, end_day = install_day, 365 * num_years year = 2018 weather = temperature.load("holly-springs-18") _0, _1, traffic_array = traffic.load_traffic( config, traffic.normal_traffic(config), 60 * 10 ) weather["temp"] = temperature.resize(weather["temp"], year=year) # weather = temperature.repeat(config, "holly-springs-18", weather, num_years) start_date, end_date = ( weather["datetime"].iloc[0].strftime(temperature.f_string), weather["datetime"].iloc[-1].strftime(temperature.f_string), ) for s_i, support in enumerate(support_with_points(config.bridge, delta_x=delta_x)): # Increase pier settlement until threshold triggered. for settlement in np.arange(0, 10, 0.1): responses = ( sim.responses.to( config=config, points=[support.point], traffic_array=traffic_array, response_type=model.RT.YTrans, with_creep=True, weather=weather, start_date=start_date, end_date=end_date, install_day=install_day, start_day=start_day, end_day=end_day, pier_settlement=[ ( model.PierSettlement(pier=s_i, settlement=0), model.PierSettlement(pier=s_i, settlement=settlement / 1e3), ) ], skip_weather_interp=True, ) * 1e3 ) # Determine the minimum response for this level of settlement. max_r = min(responses[0]) to_write = f"Min {max_r} for settlement {settlement} mm for support {s_i}, sensor at X = {support.point.x}, Z = {support.point.z}" print_w(to_write) if max_r < -THRESH: break # Write the minimum response and settlement for this support to a file. with open(log_path, "a") as f: f.write(to_write) plt.scatter([support.point.x], [support.point.z], c="red") plt.annotate( f"{np.around(settlement, 2)} mm", xy=(support.point.x - 3, support.point.z + 2), color="b", size="large", ) plot.top_view_bridge(config.bridge, lanes=True, piers=True, units="m") plt.title("Minimum pier settlement detected (Question 2B)") plt.tight_layout() plt.savefig(config.get_image_path("classify/q2b", "q2b-min-ps.pdf"))
35,774
12,370
#%% import json import requests from io import StringIO import pandas as pd # %% with open("../db_pass", "r") as f: token = json.load(f)['token'] # %% data = { 'token': token, 'content': 'record', 'format': 'csv', 'type': 'flat', 'csvDelimiter': '', 'rawOrLabel': 'raw', 'rawOrLabelHeaders': 'raw', 'exportCheckboxLabel': 'false', 'exportSurveyFields': 'false', 'exportDataAccessGroups': 'false', 'returnFormat': 'csv', 'fields': 'patient_id,age,bmi,covid_test_date,date_of_test,weight,height,admission_date,final_date,death,sex' } r = requests.post('http://192.168.45.244/api/',data=data) print('HTTP Status: ' + str(r.status_code)) data = StringIO(r.text) # %% df = pd.read_csv(data) df = df[df["height"].apply(lambda x: not pd.isna(x))] df = df.dropna(axis=1, how='all') df["bmi"] = df["bmi"].apply(lambda x: round(x, 1)) df.to_csv("metadata.csv", index=False) print(df) # %%
936
371
""" Sequence statistics: Count, length, bounding boxes size. """ import os from glob import glob import pickle from tqdm import tqdm def extract_stats(cache_path): # Load sequences from file with open(cache_path, "rb") as fp: # Unpickling seq_list = pickle.load(fp) if len(seq_list) == 0: return 0, 0., 0. # For each sequence len_sum, size_sum = 0., 0. for seq in seq_list: len_sum += len(seq) size_sum += seq.size_avg return len(seq_list), len_sum / len(seq_list), size_sum / len(seq_list) def main(in_dir, out_path=None, postfix='_dsfd_seq.pkl'): out_path = os.path.join(in_dir, 'sequence_stats.txt') if out_path is None else out_path # Validation if not os.path.isdir(in_dir): raise RuntimeError('Input directory not exist: ' + in_dir) # Parse file paths input_query = os.path.join(in_dir, '*' + postfix) file_paths = sorted(glob(input_query)) # For each file in the input directory with the specified postfix pbar = tqdm(file_paths, unit='files') count_sum, len_sum, size_sum = 0., 0., 0. vid_count = 0 for i, file_path in enumerate(pbar): curr_count, curr_mean_len, curr_mean_size = extract_stats(file_path) if curr_count == 0: continue count_sum += curr_count len_sum += curr_mean_len size_sum += curr_mean_size vid_count += 1 pbar.set_description('mean_count = %.1f, mean_len = %.1f, mean_size = %.1f, valid_vids = %d / %d' % (count_sum / vid_count, len_sum / vid_count, size_sum / vid_count, vid_count, i + 1)) # Write result to file if out_path is not None: with open(out_path, "w") as f: f.write('mean_count = %.1f\n' % (count_sum / vid_count)) f.write('mean_len = %.1f\n' % (len_sum / vid_count)) f.write('mean_size = %.1f\n' % (size_sum / vid_count)) f.write('valid videos = %d / %d\n' % (vid_count, len(file_paths))) if __name__ == "__main__": # Parse program arguments import argparse parser = argparse.ArgumentParser('detections2sequences') parser.add_argument('input', metavar='DIR', help='input directory') parser.add_argument('-o', '--output', default=None, metavar='PATH', help='output directory') parser.add_argument('-p', '--postfix', metavar='POSTFIX', default='_dsfd_seq.pkl', help='the files postfix to search the input directory for') args = parser.parse_args() main(args.input, args.output, args.postfix)
2,622
901
# Generated by Django 3.0.7 on 2020-07-04 04:05 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('app', '0003_auto_20200704_0400'), ] operations = [ migrations.AlterModelOptions( name='person', options={'verbose_name_plural': 'People'}, ), ]
352
134
import datetime from tkapi.util import queries from tkapi.zaak import Zaak, ZaakSoort from tkapi.dossier import Dossier, DossierWetsvoorstel from tkapi.document import Document from .core import TKApiTestCase class TestDossier(TKApiTestCase): def test_get_dossiers(self): dossiers = self.api.get_dossiers(filter=None, max_items=10) self.assertEqual(10, len(dossiers)) def test_get_dossier_by_nummer(self): nummer = 34435 filter = Dossier.create_filter() filter.filter_nummer(nummer) dossiers = self.api.get_dossiers(filter=filter) self.assertEqual(len(dossiers), 1) dossiers[0].print_json() def test_dossier_filter(self): self.check_dossier_filter('2016Z16486', 34537) self.check_dossier_filter('2016Z24906', 34640) def check_dossier_filter(self, zaak_nr, expected_dossier_nummer): dossier_filter = Dossier.create_filter() dossier_filter.filter_zaak(zaak_nr) dossiers = self.api.get_dossiers(filter=dossier_filter) # for dossier in dossiers: # dossier.print_json() self.assertEqual(len(dossiers), 1) # print(dossiers[0].nummer) self.assertEqual(dossiers[0].nummer, expected_dossier_nummer) class TestDossiersForZaken(TKApiTestCase): start_datetime = datetime.datetime(year=2016, month=1, day=1) end_datetime = datetime.datetime(year=2016, month=1, day=14) def test_get_dossiers(self): zaak_filter = Zaak.create_filter() zaak_filter.filter_date_range( TestDossiersForZaken.start_datetime, TestDossiersForZaken.end_datetime ) zaak_filter.filter_soort(ZaakSoort.WETGEVING) zaken = self.api.get_zaken(zaak_filter) print('Wetgeving zaken found: ' + str(len(zaken))) dossier_filter = Dossier.create_filter() zaak_nummers = [zaak.nummer for zaak in zaken] print(zaak_nummers) dossier_filter.filter_zaken(zaak_nummers) dossiers = self.api.get_dossiers(filter=dossier_filter) dossier_zaak_nummers = set() for dossier in dossiers: print('dossier.nummer: ', str(dossier.nummer)) for zaak in dossier.zaken: dossier_zaak_nummers.add(zaak.nummer) print('dossier_zaak_nummers', dossier_zaak_nummers) for zaak in zaken: if zaak.nummer not in dossier_zaak_nummers: print(zaak.nummer) # zaak.print_json() # self.assertTrue(zaak_nr in dossier_zaak_nummers) # print(zaken) for zaak_nummer in zaak_nummers: self.assertTrue(zaak_nummer in dossier_zaak_nummers) class TestDossierAfgesloten(TKApiTestCase): start_datetime = datetime.datetime(year=2015, month=1, day=1) end_datetime = datetime.datetime.now() def test_filter_afgesloten(self): dossier_filter = Dossier.create_filter() dossier_filter.filter_afgesloten(True) dossiers = self.api.get_dossiers(filter=dossier_filter) # There are currently no afgesloten dossiers, this will hopefully change in the future self.assertEqual(len(dossiers), 0) class TestDossierFilter(TKApiTestCase): def test_filter_dossier_nummer(self): nummer = 33885 dossier = queries.get_dossier(nummer) self.assertEqual(nummer, dossier.nummer) def test_filter_dossier_nummer_toevoeging(self): nummer = 35300 toevoeging = 'XVI' dossier = queries.get_dossier(nummer, toevoeging=toevoeging) self.assertEqual(nummer, dossier.nummer) self.assertEqual(toevoeging, dossier.toevoeging) def test_get_document_actors(self): # nummer = 35234 nummer = 33885 dossier = queries.get_dossier(nummer) for zaak in dossier.zaken: print('==========') print(zaak.soort, zaak.onderwerp, zaak.volgnummer) for actor in zaak.actors: print(actor.naam, actor.persoon.achternaam if actor.persoon else None, actor.fractie, actor.commissie) for doc in zaak.documenten: print(doc.soort, doc.onderwerp, doc.titel, doc.volgnummer) for actor in doc.actors: print(actor.naam) class TestWetsvoorstelDossier(TKApiTestCase): def test_get_wetsvoorstellen_dossiers(self): max_items = 200 wetsvoorstellen = self.api.get_items(DossierWetsvoorstel, max_items=max_items) self.assertEqual(max_items, len(wetsvoorstellen)) def test_get_begroting_dossiers(self): filter = Zaak.create_filter() filter.filter_date_range(datetime.date(year=2019, month=6, day=1), datetime.date.today()) filter.filter_soort(ZaakSoort.BEGROTING, is_or=True) zaken = self.api.get_zaken(filter=filter) for zaak in zaken: dossier_id = str(zaak.dossier.nummer) print(dossier_id) def test_get_dossiers_via_documenten(self): pd_filter = Document.create_filter() # NOTE: this date filter does not seem to work in combination with the soort filter. # start_datetime = datetime.datetime(year=2016, month=1, day=1) # end_datetime = datetime.datetime(year=2016, month=2, day=1) # pd_filter.filter_date_range(start_datetime, end_datetime) pd_filter.filter_soort('Voorstel van wet', is_or=True) pd_filter.filter_soort('Voorstel van wet (initiatiefvoorstel)', is_or=True) pds = self.api.get_documenten(pd_filter) dossier_nrs = [] pds_no_dossier_nr = [] for pd in pds[:10]: print(pd.dossier_nummers) if pd.dossier_nummers: dossier_nrs += pd.dossier_nummers else: pds_no_dossier_nr.append(pd) for pd in pds_no_dossier_nr: print(pd.dossier_nummers) print(pd.onderwerp) dossier_nrs = sorted(set(dossier_nrs)) print(dossier_nrs) for dossier_nr in dossier_nrs: print(dossier_nr) print(len(dossier_nrs)) # def test_get_dossiers(self): # zaak_filter = Zaak.create_filter() # start_datetime = datetime.datetime(year=2005, month=1, day=1) # end_datetime = datetime.datetime.now() # zaak_filter.filter_date_range(start_datetime, end_datetime) # zaak_filter.filter_soort('Wetgeving') # zaken = self.api.get_zaken(zaak_filter) # print('Wetgeving zaken found: ' + str(len(zaken))) # zaak_nummers = [zaak.nummer for zaak in zaken] # print(zaak_nummers) # dossiers = [] # nrs_batch = set() # for zaak_nr in zaak_nummers: # nrs_batch.add(zaak_nr) # if len(nrs_batch) < 10: # continue # dossier_filter = Dossier.create_filter() # dossier_filter.filter_zaken(nrs_batch) # nrs_batch = set() # dossiers_for_zaak = self.api.get_dossiers(filter=dossier_filter) # if dossiers_for_zaak: # dossiers += dossiers_for_zaak # print('Dossier found for zaak: ' + str(zaak_nr)) # else: # print('WARNING: No dossier found for zaak: ' + str(zaak_nr)) # dossier_nummers = [] # for dossier in dossiers: # print('\n=======') # print(dossier.nummer) # print(dossier.afgesloten) # print(dossier.organisatie) # print(dossier.titel) # dossier_nummers.append(dossier.nummer) # # dossier.print_json() # dossier_nrs = sorted(set(dossier_nummers)) # print(dossier_nrs) # print(len(dossier_nrs))
7,735
2,842
""" This library contains several functions designed to help with the illustration of hexagonal grids Functions: plot_hexagaons : plots a specified data vector over a 2-D hexagon grid. create_alpha_mask : creates an alpha shape (a concave hull), which is required for plotting contours; without it, the contour function extrapolates outside of the model area. plot_scattered_contour : plots contour lines over an irregular grid, such as a hexagonal one. plot_hexagons_3d : plots a 2-dimensional hexagon grid with specified z-dimensions """ def plot_hexagons (data, hexagon_grid_cores, hexagon_radius, hexagon_orientation = 0, colormap = 'steel', color = None, vmin = None, vmax = None, vincr = None, xlabel = None, ylabel = None, clabel = None, hide_colorbar = False, **kwargs): """ Call to plot a specified vector (positions relative to node IDs) in a hexagonal grid @params: data - Required : vector of values for hexagonal plot, positions corresponding to cell IDs (counting from zero) hexagon_grid_cores - Required : tessellated polygons over area of interest hexagon_radius - Required : radius of hexagons used for tessellation hexagon_orientation - Optional : orientation of hexagon in clock-wise degrees [0 = flat top] colormap - Optional : specify a colormap as string vmin - Optional : externally specified min value for colorbar vmax - Optional : externally specified max value for colorbar vincr - Optional : specified value increment for colorbar xlabel - Optional : string for xlabel ylabel - Optional : string for ylabel clabel - Optional : string for colorbar label **kwargs - Optional : keyword arguments for matplotlib.patches.RegularPolygon """ import matplotlib import numpy as np import math #-------------------------------------------------------------------------- # Prepare data for plotting #-------------------------------------------------------------------------- # If not specified, define range of values if vmin == None or vmax == None: vmin = np.min(data) vmax = np.max(data) vrange = vmax-vmin if vincr == None: vincr = vrange/100 # Snap value range to integers vmin = int(vmin/vincr)*vincr # minimum value for colorbar vmax = (int(vmax/vincr)+1)*vincr # maximum value for colorbar if color is None: # Retrieve colormap if colormap == 'steel': # Create colormap 'steel' from matplotlib.colors import LinearSegmentedColormap cmap_steel = [(0.007843137,0.305882353,0.443137255), (0.301960784,0.592156863,0.784313725),(0.623529412,0.776470588,0.882352941)] cm = LinearSegmentedColormap.from_list('steel', cmap_steel, N=100) cmaps = cm else: cmaps = colormap # Correct orientation orientation = math.radians(-hexagon_orientation+30) # Hexagon radius only goes to normal of sides edgepoint_distance = hexagon_radius/np.cos(np.deg2rad(30)) # Retrieve colormap information if color is None: cmap = matplotlib.cm.get_cmap(cmaps) #-------------------------------------------------------------------------- # Start plotting #-------------------------------------------------------------------------- # Create empty figure ax1 = matplotlib.pyplot.gca() # Plot hexagons for hex in range(len(hexagon_grid_cores[:,0])): # Retrieve color value if color is None: rgba = cmap((data[hex]-vmin)/(vrange)) rgba = matplotlib.colors.rgb2hex(rgba) else: rgba = color # Add the patch ax1.add_patch( matplotlib.patches.RegularPolygon( (hexagon_grid_cores[hex,0], hexagon_grid_cores[hex,1]), # x and y 6, # edges edgepoint_distance, orientation=orientation, facecolor = rgba, **kwargs) ) # Determine meaningful colorbar steps if color is None: colorbar_increment = vincr colorbar_min = int(vmin/colorbar_increment)*colorbar_increment # minimum value for colorbar colorbar_max = (int(vmax/colorbar_increment)+1)*colorbar_increment # maximum value for colorbar colorbar_increment_numbers = int((colorbar_max-colorbar_min)/colorbar_increment+1) colorbar_steps = [] for num in range(colorbar_increment_numbers): colorbar_steps = colorbar_steps + [colorbar_min+num*colorbar_increment] # Recompute the ax.dataLim ax1.relim() # Update ax.viewLim using the new dataLim ax1.autoscale_view() # Create colorbar if hide_colorbar == False and color is None: norm = matplotlib.colors.Normalize(vmin=vmin,vmax=vmax) sm = matplotlib.pyplot.cm.ScalarMappable(cmap=cmap, norm=norm) sm.set_array([]) cbar = matplotlib.pyplot.colorbar(sm) # Label plot if xlabel != None: matplotlib.pyplot.xlabel(xlabel) if ylabel != None: matplotlib.pyplot.ylabel(ylabel) if clabel != None and not hide_colorbar and color is None: cbar.set_label(clabel, rotation=270, labelpad=20) def create_alpha_mask(points, distance_limit, resolution_x = 1000, resolution_y = 1000, visualization = True): """ Creates interpolation grid, then masks over the alpha shape spanned up by points and defined by distance_limit. @params: points - Required : points spanning up alpha shape distance_limit - Required : distance threshold for removing Delaunay simplices resolution_x - Optional : resolution for grid in x, default is 1000 resolution_y - Optional : resolution for grid in y, default is 1000 visualization - Optional : boolean for visualizing result, default is False Returns: grid_mask : An array containing 1 for cells inside, and 0 for cells outside """ import numpy as np from scipy.spatial import Delaunay from matplotlib.collections import LineCollection import matplotlib.path as mplPath #---------------------------------------------------------------------- # Create Grid #---------------------------------------------------------------------- # Create meshgrid xi = np.transpose(np.linspace(min(points[:,0]), max(points[:,0]), resolution_x)) yi = np.transpose(np.linspace(min(points[:,1]), max(points[:,1]), resolution_y)) X, Y = np.meshgrid(xi, yi) # Reshape into vector gridpoints_x = np.reshape(X, resolution_x*resolution_y) gridpoints_y = np.reshape(Y, resolution_x*resolution_y) # Combine into gridpoints array gridpoints = np.transpose(np.asarray((gridpoints_x, gridpoints_y))) #---------------------------------------------------------------------- # Create Alpha Shape #---------------------------------------------------------------------- # Start Delaunay triangulation tri = Delaunay(points) # Auxiliary function for plotting, if required if visualization == True: import matplotlib.pyplot as plt edges = set() edge_points = [] def add_edge(i, j): """Add a line between the i-th and j-th points, if not in the list already""" if (i, j) in edges or (j, i) in edges: # already added return edges.add( (i, j) ) edge_points.append(points[ [i, j] ]) # Remove simplices outside of distance_limit simplex_flag = np.zeros(len(tri.simplices[:,0])) # Flags bad simplices counter = 0 for ia, ib, ic in tri.vertices: # ia, ib, ic = indices of corner points of the triangle if np.sqrt((points[ia,0]-points[ib,0])**2+(points[ia,1]-points[ib,1])**2) < distance_limit and \ np.sqrt((points[ia,0]-points[ic,0])**2+(points[ia,1]-points[ic,1])**2) < distance_limit and \ np.sqrt((points[ib,0]-points[ic,0])**2+(points[ib,1]-points[ic,1])**2) < distance_limit: # do nothing simplex_flag[counter] = 0 else: # simplex has at least one side larger than threshold, flag it simplex_flag[counter] = 1 counter += 1 tri.simplices = tri.simplices[simplex_flag == 0,:] # Remove bad simplices tri.vertices = tri.vertices[simplex_flag == 0,:] # Remove bad simplices # Visualize, if requested if visualization == True: # Mark all remaining simplices for ia, ib, ic in tri.vertices: add_edge(ia, ib) add_edge(ib, ic) add_edge(ic, ia) # Draw them lines = LineCollection(edge_points) plt.figure() plt.gca().add_collection(lines) plt.plot(points[:,0], points[:,1], 'o') #---------------------------------------------------------------------- # Mask over Alpha Shape #---------------------------------------------------------------------- # Prepare point flag flag_gridpoints = np.zeros(len(gridpoints[:,0]), dtype = np.int) # Evaluate gridpoints for sim in range(len(tri.simplices[:,0])): # Print progress bar cv = sim mv = len(tri.simplices[:,0])-1 print('\r%s |%s| %s%% %s' % ('Masking: ', '\033[33m'+'█' * int(50 * cv // mv) + '-' * (50 - int(50 * cv // mv))+'\033[0m', ("{0:." + str(1) + "f}").format(100 * (cv / float(mv))), ' Complete'), end = '\r') # Create simplex path bbPath = mplPath.Path(np.array([points[tri.simplices[sim,0],:], points[tri.simplices[sim,1],:], points[tri.simplices[sim,2],:], points[tri.simplices[sim,0],:]])) # Flag points that are inside this simplex for gridpts in range(len(gridpoints[:,0])): if flag_gridpoints[gridpts] == 0: # only process points not already allocated if bbPath.contains_point((gridpoints[gridpts,0],gridpoints[gridpts,1])) == True: flag_gridpoints[gridpts] = 1 # Plot, if required if visualization == True: plt.scatter(gridpoints[flag_gridpoints == 1,0], gridpoints[flag_gridpoints == 1,1],color = 'g') plt.scatter(gridpoints[flag_gridpoints == 0,0], gridpoints[flag_gridpoints == 0,1],color = 'r') # Reshape flag_gridpoints into a 2D array global grid_mask grid_mask = np.reshape(flag_gridpoints,(resolution_y,resolution_x)) # Return result return grid_mask def plot_scattered_contour(x, y, data, resolution_x=1000, resolution_y=1000, grid_mask = None, vmin = None, vmax = None, vincr = None, suppress_clabel = False, **kwargs): """ Call to plot contour of scattered data @params: x - Required : x-coordinate y - Required : y-coordinate data - Required : data for the contours resolution_x - Optional : resolution of auxiliary grid in x resolution_y - Optional : resolution of auxiliary grid in y grid_mask - Optional : mask array of dimension [resolution_y,resolution_x] vmin - Optional : min value for contour vmax - Optional : max value for contour vincr - Optional : increment for contour suppress_clabel - Optional : Flag wether contours should be labeld, False by default **kwargs - Optional : keyword arguments for matplotlib.patches.RegularPolygon """ import numpy as np import matplotlib import scipy #-------------------------------------------------------------------------- # Integrity checks #-------------------------------------------------------------------------- # Check if grid_mask matches meshgrid dimensions if len(grid_mask) != 1: if len(grid_mask[:,0]) != resolution_y or len(grid_mask[0,:]) != resolution_x: raise Exception('Grid mask dimensions must match resolution in x and y!') # Check if one of the cells has dried; this algorithm can't handle that yet if vmin < -1000: print('\033[31m'+'WARNING:'+'\033[0m'+' Dried cells detected. Contour not printed.') return # Extract vmin and vmax, if not specified if vmin == None or vmax == None: vmin = np.min(data) vmax = np.max(data) # Set vincr, if not specified if vincr == None: vincr = (vmax-vmin)/10 # Snap value range to integers vmin = int(vmin/vincr)*vincr # minimum value for colorbar vmax = (int(vmax/vincr)+1)*vincr # maximum value for colorbar #-------------------------------------------------------------------------- # Prepare data for plotting #-------------------------------------------------------------------------- # Convert source material into required format source = np.transpose(np.asarray([x,y])) # Create and convert target material xi = np.transpose(np.linspace(min(x), max(x), resolution_x)) yi = np.transpose(np.linspace(min(y), max(y), resolution_y)) X, Y = np.meshgrid(xi, yi) target = np.transpose(np.asarray([X,Y])) # Interpolate and transpose Z = scipy.interpolate.griddata(source, data, target) Z = np.transpose(Z) # Mask values, if grid_mask was specified if len(grid_mask) != 1: Z[grid_mask == 0] = float('NaN') # Define function for masking levels = np.arange(vmin,vmax,vincr) #-------------------------------------------------------------------------- # Plot that shit #-------------------------------------------------------------------------- CS = matplotlib.pyplot.contour(xi,yi,Z,levels=levels,**kwargs) if not suppress_clabel: matplotlib.pyplot.clabel(CS, inline=1, inline_spacing = 0) return def plot_hexagons_3d(grid, zdim, hexagon_radius, hexagon_orientation = 0, xlabel = 'x', ylabel = 'y', zlabel = 'z', clabel = 'depth', depth_colormap = 'steel', alpha = 1, **kwargs): """ Call to tessellate a given polygon with hexagons @params: grid - Required : x-y-coordinates of center of hexagons, array of form [nx2] zdim - Required : bottom and top elevation of hexagon cells, array of form [nx2] hexagon_radius - Required : radius of hexagons used for tessellation hexagon_orientation - Required : orientation of hexagon in clock-wise degrees [0 = flat top] xlabel - Optional : label for x-axis ylabel - Optional : label for y-axis zlabel - Optional : label for z-axis clabel - Optional : label for colorbar depth_colormap - Optional : string of colormap, if requested alpha - Optional : alpha value for transparency of polygons, default is 1 **kwargs - Optional : keyword arguments for Poly3DCollection """ # PLOT 3D import numpy as np import matplotlib.pyplot as plt from mpl_toolkits.mplot3d import Axes3D from mpl_toolkits.mplot3d.art3d import Poly3DCollection, Line3DCollection import math if depth_colormap == 'steel': # Create colormap 'steel' from matplotlib.colors import LinearSegmentedColormap cmap_steel = [(0.007843137,0.305882353,0.443137255), (0.301960784,0.592156863,0.784313725),(0.623529412,0.776470588,0.882352941)] cm = LinearSegmentedColormap.from_list('steel', cmap_steel, N=100) cmaps = cm else: cmaps = depth_colormap # Initialize figure fig = plt.figure() ax = fig.add_subplot(111, projection='3d') # Hexagon radius only goes to normal of sides edgepoint_distance = hexagon_radius/np.cos(np.deg2rad(30)) # Determine depth range, if colorbar is requested vmin = np.min(zdim[:,1]-zdim[:,0]) vmax = np.max(zdim[:,1]-zdim[:,0]) c_range = vmax-vmin # Plot hexagons for hex in range(len(grid[:,0])): # Reset coordinate variables x = [] y = [] # Read top and bottom elevation zbot = zdim[hex,0] ztop = zdim[hex,1] # Pre-allocate memory for coordinate matrix Z = np.zeros((12,3)) # Determine cell color, if requested if depth_colormap != 'None': import matplotlib # Retrieve colormap information cmap = matplotlib.cm.get_cmap(cmaps) rgba = cmap((ztop-zbot-vmin)/c_range) #cmap((zbot-vmin)/(vmax-vmin)) rgba = list(rgba) rgba[3] = alpha # rgba = matplotlib.colors.rgb2hex(rgba) # Plot grid counter = 0 for angle in range(0-hexagon_orientation, 420-hexagon_orientation, 60): # Coordinates of edge point x = np.append(x,grid[hex,0]+math.cos(math.radians(angle)) * edgepoint_distance) y = np.append(y,grid[hex,1]+math.sin(math.radians(angle)) * edgepoint_distance) # Write into coordinate matrix if counter < 6: Z[counter,0] = grid[hex,0]+math.cos(math.radians(angle)) * edgepoint_distance Z[counter,1] = grid[hex,1]+math.sin(math.radians(angle)) * edgepoint_distance Z[counter,2] = ztop Z[6+counter,0] = grid[hex,0]+math.cos(math.radians(angle)) * edgepoint_distance Z[6+counter,1] = grid[hex,1]+math.sin(math.radians(angle)) * edgepoint_distance Z[6+counter,2] = zbot counter += 1 # Vertices of hexagon sides verts = [[Z[0],Z[1],Z[7],Z[6]], [Z[1],Z[2],Z[8],Z[7]], [Z[2],Z[3],Z[9],Z[8]], [Z[3],Z[4],Z[10],Z[9]], [Z[4],Z[5],Z[11],Z[10]], [Z[5],Z[0],Z[6],Z[11]]] if depth_colormap != 'None': # Plot hexagon side face = Poly3DCollection(verts, **kwargs) face.set_facecolor(rgba) ax.add_collection3d(face) else: face = Poly3DCollection(verts, **kwargs) face.set_facecolor(rgba) ax.add_collection3d(face) # Vertices of hexagon top verts = [[Z[0],Z[1],Z[2],Z[3],Z[4],Z[5]]] # Plot hexagon top if depth_colormap != 'None': # Plot hexagon side face = Poly3DCollection(verts, **kwargs) face.set_facecolor(rgba) ax.add_collection3d(face) else: face = Poly3DCollection(verts, **kwargs) face.set_facecolor(rgba) ax.add_collection3d(face) # Vertices of hexagon bot verts = [[Z[6],Z[7],Z[8],Z[9],Z[10],Z[11]]] # Plot hexagon bot if depth_colormap != 'None': # Plot hexagon side face = Poly3DCollection(verts, **kwargs) face.set_facecolor(rgba) ax.add_collection3d(face) else: face = Poly3DCollection(verts, **kwargs) face.set_facecolor(rgba) ax.add_collection3d(face) # Determine meaningful colorbar steps, if colorbar was requested if depth_colormap != 'None': colorbar_increment = 0.1 colorbar_min = int(vmin/colorbar_increment)*colorbar_increment # minimum value for colorbar colorbar_max = (int(vmax/colorbar_increment)+1)*colorbar_increment # maximum value for colorbar colorbar_increment_numbers = int((colorbar_max-colorbar_min)/colorbar_increment+1) colorbar_steps = [] for num in range(colorbar_increment_numbers): colorbar_steps = colorbar_steps + [colorbar_min+num*colorbar_increment] # Create colorbar norm = matplotlib.colors.Normalize(vmin=vmin,vmax=vmax) sm = matplotlib.pyplot.cm.ScalarMappable(cmap=cmap, norm=norm) sm.set_array([]) cbar = matplotlib.pyplot.colorbar(sm) cbar.set_label(clabel, rotation=270, labelpad=20) # Label axes ax.set_xlabel(xlabel) ax.set_ylabel(ylabel) ax.set_zlabel(zlabel) # Equal aspect scaling doesn't work yet, manual workaround # Designate array of edges xyzlims = np.zeros((3,2)) xyzlims[0,0] = np.min(grid[:,0]) xyzlims[0,1] = np.max(grid[:,0]) xyzlims[1,0] = np.min(grid[:,1]) xyzlims[1,1] = np.max(grid[:,1]) xyzlims[2,0] = np.min(zdim) xyzlims[2,1] = np.max(zdim) # Determine maximal range maxrange = np.max([xyzlims[0,1]-xyzlims[0,0],xyzlims[1,1]-xyzlims[1,0],xyzlims[2,1]-xyzlims[2,0]]) # Determine difference to maximal range xdif = maxrange - (xyzlims[0,1]-xyzlims[0,0]) ydif = maxrange - (xyzlims[1,1]-xyzlims[1,0]) zdif = maxrange - (xyzlims[2,1]-xyzlims[2,0]) # Set axis limits -> equal aspect ax.set_xlim3d(xyzlims[0,0]-xdif/2,xyzlims[0,1]+xdif/2) ax.set_ylim3d(xyzlims[1,0]-ydif/2,xyzlims[1,1]+ydif/2) ax.set_zlim3d(xyzlims[2,0]-zdif/2,xyzlims[2,1]+zdif/2) # Show result plt.show() def vulture_plot(incr = 1, elev = 40., fps = 50): """ Creates a short animated .gif providing a flight around the 3-D model, requiring an open, compatible 3D figure @params: incr - Optional : degree increment for rotation frames; defines temporal resolution of .gif (default = 1) elev - Optional : elevation angle for camera (default = 40) fps - Optional : frames per second for resulting .gif; defines speed of .gif display (default 50) """ # Import libraries import imageio import os import matplotlib.pyplot as plt # Retrieve axis ax = plt.gca() # Rotate, save and compile vulture plot images = [] for cv in range(0,360,incr): # Rotate image ax.view_init(elev=40., azim=cv) plt.show() # Save it as temporary file plt.savefig("dummy.png") # Append it to saved movie images.append(imageio.imread("dummy.png")) # Remove temporary file os.remove("dummy.png") # Print progress bar mv = 359 # max value print('\r%s |%s| %s%% %s' % ('Printing: ', '\033[33m'+'█' * int(50 * cv // mv) + '-' * (50 - int(50 * cv // mv))+'\033[0m', ("{0:." + str(1) + "f}").format(100 * (cv / float(mv))), ' Complete'), end = '\r') # Compile .gif imageio.mimsave('output_quick.gif', images,fps=fps) def visualize_genealogy(genealogy,weights = None, rejuvenation = None,colormap = 'jet'): """ Creates an inline figure visualizing the particle genealogy over one resampling step. @params: genealogy - Required : vector describing genealogy of resampled particles, referring to indices weights - Optional : weight of particles prior to resampling rejuvenation - Optional : vector of booleans describing whether particles were rejuvenated colormap - Optional : colormap string for visualization """ import numpy as np from IPython import get_ipython import matplotlib import matplotlib.pyplot as plt # Determine number of particles n_particles = len(genealogy) # Assign optional variables, if not provided if weights is None == True: weights = np.ones(n_particles) # if rejuvenation is None == True: # rejuvenation = np.ones((n_particles),dtype = np.bool) # Switch to inline printing get_ipython().run_line_magic('matplotlib', 'inline') # Create dummy features for the legend full_line = plt.Line2D([], [], color='black',label='inherited') dashed_line = plt.Line2D([], [], linestyle = '--', color='black',label='rejuvenated') particle = plt.Line2D([], [], linestyle = 'None', marker ='.', color='black',label='particle') # Plot legend plt.legend(handles=[dashed_line,full_line,particle],bbox_to_anchor=(0., -0.05, 1., .102), loc=3, ncol=3, mode="expand", borderaxespad=0.) # Determine colormap for particles cmap = matplotlib.cm.get_cmap(colormap) # Extract particle colors rgba = [None] * n_particles for n in range(n_particles): rgba[n] = matplotlib.colors.rgb2hex(cmap(n/(n_particles-1))) # Create plot for n in range(n_particles): plt.plot([genealogy[n],n],[1,2],'--',c=rgba[genealogy[n]]) # Draw genealogy of current particle # if rejuvenation[n] == False: # plt.plot([genealogy[n],n],[1,2],c=rgba[genealogy[n]]) # else: # plt.plot([genealogy[n],n],[1,2],c='w') # plt.plot([genealogy[n],n],[1,2],'--',c=rgba[genealogy[n]]) # Scatter previous and current particle index if weights[n] == 0: # Particle weight is zero - print as greyscale plt.scatter(n,1,s = weights[n]/np.max(weights)*55+5,c='xkcd:medium grey') else: plt.scatter(n,1,s = weights[n]/np.max(weights)*55+5,c=rgba[n]) plt.scatter(n,2,s=20,c=rgba[n]) # Deactivate axes plt.axis('off') # Show, and revert to automatic printing plt.show() get_ipython().run_line_magic('matplotlib', 'qt5')
27,649
8,917
''' self_concepts_test This module serves as the unit test for self_concepts ''' import argparse, sys sys.path.append('../../source/python') from self_concepts import Concept from self_concepts import Property from self_concepts import Relationship from self_concepts import Ontology from self_concepts import Blackboard from self_concepts import Agent from self_concepts import SelfException # Helper functions in support of concise and verbose reporting def parseArguments(): '''Collect and return the test's arguments.''' parser = argparse.ArgumentParser(description='Test ') parser.add_argument('-c', '--concise', action='store_true', help='test self_concept with concise results') return parser.parse_args() def reportHeader(message): '''Print a report header.''' if arguments.concise != True: print(message) else: print('#', end='') def reportSection(message): '''Print a section header.''' if arguments.concise != True: print(' ' + message) else: print('*', end='') def reportDetail(message): '''Print a report detail.''' if arguments.concise != True: print(' ' + message) else: print('.', end='') def reportDetailFailure(message): '''Print a report failure.''' if arguments.concise != True: print('!!!!!!! ' + message) else: print('!') exit() def reportConceptName(concept: 'Concept'): '''Print the name of the concept.''' reportDetail(' Function applied to ' + concept.__class__.__name__ + ' (' + concept.name + ')') # Various functions, classes, and instances used for testing class AnotherConcept(Concept): pass CONCEPT_NAME_1 = 'A well-formed concept' CONCEPT_NAME_2 = 'A well-formed concept' CONCEPT_NAME_3 = 'Another well-formed concept' CONCEPT_NAME_4 = 'A well-formed concept' c1 = Concept(CONCEPT_NAME_1) c2 = Concept(CONCEPT_NAME_2) c3 = AnotherConcept(CONCEPT_NAME_3) c4 = Concept(CONCEPT_NAME_4) class AnotherProperty(Property): pass class YetAnotherProperty(AnotherProperty): pass PROPERTY_NAME_1 = 'A well-formed property' PROPERTY_NAME_2 = 'A well-formed property' PROPERTY_NAME_3 = 'Another well-formed property' PROPERTY_NAME_4 = 'A well-formed property' PROPERTY_VALUE_1 = 42 PROPERTY_VALUE_2 = 'A value' PROPERTY_VALUE_3 = c1 PROPERTY_VALUE_4 = 'A value' p1 = Property(PROPERTY_NAME_1, PROPERTY_VALUE_1) p2 = Property(PROPERTY_NAME_2, PROPERTY_VALUE_2) p3 = AnotherProperty(PROPERTY_NAME_3, PROPERTY_VALUE_3) p4 = Property(PROPERTY_NAME_4, PROPERTY_VALUE_4) class AnotherRelationship(Relationship): pass RELATIONSHIP_NAME_1 = 'A well-formed relationship' RELATIONSHIP_NAME_2 = 'A well-formed relationship' RELATIONSHIP_NAME_3 = 'Another well-formed relationship' RELATIONSHIP_NAME_4 = 'A well-formed relationship' r1 = Relationship(RELATIONSHIP_NAME_1, c1, c2) r2 = Relationship(RELATIONSHIP_NAME_2, c2, c3) r3 = AnotherRelationship(RELATIONSHIP_NAME_3, c3, c1) r4 = Relationship(RELATIONSHIP_NAME_4, c1, c4) ONTOLOGY_NAME_1 = 'A well-formed ontology' o1 = Ontology(ONTOLOGY_NAME_1) BLACKBOARD_NAME_1 = 'A well-formed blackboard' b1 = Blackboard(BLACKBOARD_NAME_1) class AnotherAgent(Agent): def activity(self, parameters: 'Concept' = None): super().activity(parameters) if parameters == None: reportDetail(' Activity (' + self.name + ')') else: reportDetail(' Activity (' + self.name + ') with parameters (' + parameters.name + ')') def start(self, parameters: 'Concept' = None): super().start(parameters) if parameters == None: reportDetail(' Start (' + self.name + ')') else: reportDetail(' Start (' + self.name + ') with parameters (' + parameters.name + ')') def stop(self, parameters: 'Concept' = None): super().stop(parameters) if parameters == None: reportDetail(' Stop (' + self.name + ')') else: reportDetail(' Stop (' + self.name + ') with parameters (' + parameters.name + ')') def pause(self, parameters: 'Concept' = None): super().pause(parameters) if parameters == None: reportDetail(' Pause (' + self.name + ')') else: reportDetail(' Pause (' + self.name + ') with parameters (' + parameters.name + ')') def isAlive(self) -> bool: state = super().isAlive() reportDetail(' isAlive (' + self.name + ')') return True def status(self) -> Concept: state = super().status() reportDetail(' Status (' + self.name + ')') return Concept('Status') def signal(self, source: 'Concept', message: 'Concept', parameters: 'Concept' = None): super().signal(source, message, parameters) reportDetail(' Signal to ' + self.__class__.__name__ + ' (' + self.name + ') by ' + source.__class__.__name__ + ' (' + source.name + ') regarding ' + message.__class__.__name__ + ' (' + message.name + ')') def connect(self, channel: 'Relationship', parameters: 'Concept' = None): super().connect(channel, parameters) if parameters == None: reportDetail(' Connect (' + self.name + ') to a channel (' + channel.name + ')') else: reportDetail(' Connect (' + self.name + ') with parameters (' + parameters.name + ') to a channel (' + channel.name + ')') AGENT_NAME_1 = 'A well-formed agent' AGENT_NAME_2 = 'Another well-formed agent' AGENT_NAME_3 = 'Yet another well-formed agent' a1 = AnotherAgent(AGENT_NAME_1) a2 = AnotherAgent(AGENT_NAME_2) a3 = AnotherAgent(AGENT_NAME_3) # Concept unit test def testConcept(): reportHeader('Concept') reportSection('attributes') if c1.name == CONCEPT_NAME_1: reportDetail('Correctly set and retrived name') else: reportDetailFailure('Name was not set or retrived') try: s = c1.properties reportDetailFailure('Properties were directly accessed') except SelfException: reportDetail('Correctly denied direct access to properties') try: c1.properties = set() reportDetailFailure('Properties were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to properties') reportSection('addProperty') c1.addProperty(p1) if c1.propertyExists(p1): reportDetail('Correctly added property') else: reportFailure('Property was not added') try: c1.addProperty(p1) reportDetailFailure('Property already exists') except SelfException: reportDetail('Correctly denied adding property that already exists') try: c1.addProperty('An ill-formed property') reportDetailFailure('Property is ill-formed') except SelfException: reportDetail('Correctly denied adding ill-formed property') reportSection('removeProperty') c1.removeProperty(p1) if not c1.propertyExists(p1): reportDetail('Correctly removed property') else: reportFailure('Property was not removed') try: c1.removeProperty(p2) reportDetailFailure('Property exists') except SelfException: reportDetail('Correctly denied removing property that does not exist') try: c1.removeProperty('An ill-formed property') reportDetailFailure('Property is ill-formed') except SelfException: reportDetail('Correctly denied removing ill-formed property') reportSection('removeAllProperties') c1.addProperty(p1) c1.addProperty(p2) c1.removeAllProperties() if c1.numberOfProperties() == 0: reportDetail('Correctly removed all properties') else: reportDetailFailure('Properties were not removed') reportSection('propertyExists') c1.addProperty(p1) if c1.propertyExists(p1): reportDetail('Correctly checked that property exists') else: reportDetailFailure('Property does not exist') if not c1.propertyExists(p2): reportDetail('Correctly checked that property does not exist') else: reportDetailFailure('Property exists') try: c1.propertyExists('An ill-formed property') reportDetailFailure('Property is ill-formed') except SelfException: reportDetail('Correctly denied checking existence of ill-formed property') reportSection('numberOfProperties') c1.addProperty(p2) if c1.numberOfProperties() == 2: reportDetail('Correctly reported number of properties') else: reportDetailFailure('Number of properties is wrong') reportSection('iterateOverProperties') c1.iterateOverProperties(reportConceptName) reportDetail('Correctly iterated over properties') c1.iterateOverProperties(reportConceptName, PROPERTY_NAME_1) reportDetail('Correctly iterated over properties with given name') c1.iterateOverProperties(reportConceptName, None, AnotherProperty) reportDetail('Correctly iterated over properties with given property class') c1.iterateOverProperties(reportConceptName, PROPERTY_NAME_2, Property) reportDetail('Correctly iterated over properties with given name and property class') try: c1.iterateOverProperties(reportConceptName, None, SelfException) reportDetailFailure('Property class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed property class') try: c1.iterateOverProperties(reportConceptName, None, 'An ill-formed property class') reportDetailFailure('Property class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed property class') # Property unit test def testProperty(): reportHeader('Property') reportSection('attributes') if p3.name == PROPERTY_NAME_3: reportDetail('Correctly set and retrived name') else: reportDetailFailure('Name was not set or retrived') if p3.value == c1: reportDetail('Correctly set and retrieved value') else: reportDetailFailure('Value was not set or retrieved') # Relationship unit test def testRelationship(): reportHeader('Relationship') reportSection('constructor') try: r0 = Relationship('A well-formed relationship', c1, c2) reportDetail('Correctly constructed relationship') except SelfException: reportDetailFailure('Relationship was not constructed') try: r0 = Relationship('A well-formed relationship', Concept, Concept) reportDetail('Correctly constructed relationship') except SelfException: reportDetailFailure('Relationship was not constructed') try: r0 = Relationship('An ill-formed relationship', 'An ill-formed edge', c2) reportDetailFailure('Edge is ill-formed') except SelfException: reportDetail('Correctly denied constructing relationship with ill-formed edge') try: r0 = Relationship('An ill-formed relationship', c1, 'An ill-formed edge') reportDetailFailure('Edge is ill-formed') except SelfException: reportDetail('Correctly denied constructing relationship with ill-formed edge') reportSection('attributes') r1.name = RELATIONSHIP_NAME_1; if r1.name == RELATIONSHIP_NAME_1: reportDetail('Correctly set and retrived name') else: reportDetailFailure('Name was not set or retrieved') r1.edge1 = c1 if r1.edge1 == c1: reportDetail('Correctly set and retrieved edge') else: reportDetailFailure('Edge was not set or retrieved') try: r1.edge1 = 'An ill-formed edge' reportDetailFailure('Edge is ill-formed') except SelfException: reportDetail('Correctly denied assigning ill-formed edge') try: r1.edge2 = 'An ill-formed edge' reportDetailFailure('Edge is ill-formed') except SelfException: reportDetail('Correctly denied assigning ill-formed edge') try: s = r1.edge1Properties reportDetailFailure('Edge properties were directly accessed') except SelfException: reportDetail('Correctly denied direct access to edge properties') try: r1.edge1Properties = set() reportDetailFailure('Edge properties were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to edge properties') try: s = r1.edge2Properties reportDetailFailure('Edge properties were directly accessed') except SelfException: reportDetail('Correctly denied direct access to edge properties') try: r1.edge2Properties = set() reportDetailFailure('Edge properties were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to edge properties') reportSection('addEdgeProperty') r1.addEdgeProperty(Relationship.EDGE1, p1) if r1.edgePropertyExists(Relationship.EDGE1, p1): reportDetail('Correctly added edge property') else: reportFailure('Edge property was not added') try: r1.addEdgeProperty(Relationship.EDGE1, p1) reportDetailFailure('Edge property already exists') except SelfException: reportDetail('Correctly denied adding edge property that already exists') try: r1.addEdgeProperty(Relationship.EDGE1, 'An ill-formed property') reportDetailFailure('Edge property is ill-formed') except SelfException: reportDetail('Correctly denied adding ill-formed edge property') r1.addEdgeProperty(Relationship.EDGE2, p1) if r1.edgePropertyExists(Relationship.EDGE2, p1): reportDetail('Correctly added edge property') else: reportFailure('Edge property was not added') try: r1.addEdgeProperty(Relationship.EDGE2, p1) reportDetailFailure('Edge property already exists') except SelfException: reportDetail('Correctly denied adding edge property that already exists') try: r1.addEdgeProperty(Relationship.EDGE2, 'An ill-formed property') reportDetailFailure('Edge property is ill-formed') except SelfException: reportDetail('Correctly denied adding ill-formed edge property') reportSection('removeEdgeProperty') r1.removeEdgeProperty(Relationship.EDGE1, p1) if not r1.edgePropertyExists(Relationship.EDGE1, p1): reportDetail('Correctly removed edge property') else: reportFailure('Edge property was not removed') try: r1.removeEdgeProperty(Relationship.EDGE1, p2) reportDetailProperty('Edge property exists') except SelfException: reportDetail('Correctly denied removing edge property that does not exist') try: r1.removeEdgeProperty(Relationship.EDGE1, 'An ill-formed property') reportDetailFailure('Edge property is ill-formed') except SelfException: reportDetail('Correctly denied removing ill-formed edge property') r1.removeEdgeProperty(Relationship.EDGE2, p1) if not r1.edgePropertyExists(Relationship.EDGE2, p1): reportDetail('Correctly removed edge property') else: reportFailure('Edge property was not removed') try: r1.removeEdgeProperty(Relationship.EDGE2, p2) reportDetailFailure('Edge property exists') except SelfException: reportDetail('Correctly denied removing edge property that does not exist') try: r1.removeEdgeProperty(Relationship.EDGE2, 'An ill-formed property') reportDetailFailure('Edge property is ill-formed') except SelfException: reportDetail('Correctly denied removing ill-formed edge property') reportSection('removeAllEdgeProperties') r1.addEdgeProperty(Relationship.EDGE1, p1) r1.addEdgeProperty(Relationship.EDGE1, p2) r1.removeAllEdgeProperties(Relationship.EDGE1) if r1.numberOfEdgeProperties(Relationship.EDGE1) == 0: reportDetail('Correctly removed all edge properties') else: reportDetailFailure('Edge properties were not removed') r1.addEdgeProperty(Relationship.EDGE2, p1) r1.addEdgeProperty(Relationship.EDGE2, p2) r1.removeAllEdgeProperties(Relationship.EDGE2) if r1.numberOfEdgeProperties(Relationship.EDGE2) == 0: reportDetail('Correctly removed all edge properties') else: reportDetailFailure('Edge properties were not removed') reportSection('edgePropertyExists') r1.addEdgeProperty(Relationship.EDGE1, p1) r1.addEdgeProperty(Relationship.EDGE2, p1) if r1.edgePropertyExists(Relationship.EDGE1, p1): reportDetail('Correctly checked that edge property exists') else: reportDetailFailure('Edge property does not exist') if not r1.edgePropertyExists(Relationship.EDGE1, p2): reportDetail('Correctly checked that edge property does not exist') else: reportDetailFailure('Edge property exists') try: r1.edgePropertyExists(Relationship.EDGE1, 'An ill-formed property') reportDetailFailure('Edge property is ill-formed') except SelfException: reportDetail('Correctly denied checking existence of ill-formed edge property') if r1.edgePropertyExists(Relationship.EDGE2, p1): reportDetail('Correctly checked that edge property exists') else: reportDetailFailure('Edge property does not exist') if not r1.edgePropertyExists(Relationship.EDGE2, p2): reportDetail('Correctly checked that edge property does not exist') else: reportDetailFailure('Edge property exists') try: r1.edgePropertyExists(Relationship.EDGE2, 'An ill-formed property') reportDetailFailure('Edge property is ill-formed') except SelfException: reportDetail('Correctly denied checking existence of ill-formed edge property') reportSection('numberOfEdgeProperties') r1.addEdgeProperty(Relationship.EDGE1, p2) r1.addEdgeProperty(Relationship.EDGE2, p2) if r1.numberOfEdgeProperties(Relationship.EDGE1) == 2: reportDetail('Correctly reported number of edge properties') else: reportDetailFailure('Number of edge properties is wrong') if r1.numberOfEdgeProperties(Relationship.EDGE2) == 2: reportDetail('Correctly reported number of edge properties') else: reportDetailFailure('Number of edge properties is wrong') reportSection('iterateOverEdgeProperties') r1.iterateOverEdgeProperties(Relationship.EDGE1, reportConceptName) reportDetail('Correctly iterated over edge properties') r1.iterateOverEdgeProperties(Relationship.EDGE1, reportConceptName, PROPERTY_NAME_1) reportDetail('Correctly iterated over edge properties with given name') r1.iterateOverEdgeProperties(Relationship.EDGE1, reportConceptName, None, AnotherProperty) reportDetail('Correctly iterated over edge properties with given property class') r1.iterateOverEdgeProperties(Relationship.EDGE1, reportConceptName, PROPERTY_NAME_2, Property) reportDetail('Correctly iterated over edge properties with given name and property class') try: r1.iterateOverEdgeProperties(Relationship.EDGE1, reportConceptName, None, SelfException) reportDetailFailure('Property class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed property class') try: r1.iterateOverEdgeProperties(Relationship.EDGE1, reportConceptName, None, 'An ill-formed property class') reportDetailFailure('Edge property class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed edge property class') r1.iterateOverEdgeProperties(Relationship.EDGE2, reportConceptName) reportDetail('Correctly iterated over edge properties') r1.iterateOverEdgeProperties(Relationship.EDGE2, reportConceptName, PROPERTY_NAME_1) reportDetail('Correctly iterated over edge properties with given name') r1.iterateOverEdgeProperties(Relationship.EDGE2, reportConceptName, None, AnotherProperty) reportDetail('Correctly iterated over edge properties with given property class') r1.iterateOverEdgeProperties(Relationship.EDGE2, reportConceptName, PROPERTY_NAME_2, Property) reportDetail('Correctly iterated over edge properties with given name and property class') try: r1.iterateOverEdgeProperties(Relationship.EDGE2, reportConceptName, None, SelfException) reportDetailFailure('Property class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed property class') try: r1.iterateOverEdgeProperties(Relationship.EDGE2, reportConceptName, None, 'An ill-formed property class') reportDetailFailure('Edge property class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed edge property class') # Ontology unit test def testOntology(): reportHeader('Ontology') reportSection('attributes') if o1.name == ONTOLOGY_NAME_1: reportDetail('Correctly set and retrived name') else: reportDetailFailure('Name was not set or retrived') try: s = o1.concepts reportDetailFailure('Concepts were directly accessed') except SelfException: reportDetail('Correctly denied direct access to concepts') try: o1.concepts = set() reportDetailFailure('Concepts were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to concepts') try: s = o1.relationships reportDetailFailure('Relationships were directly accessed') except SelfException: reportDetail('Correctly denied direct access to relationships') try: o1.relationships = set() reportDetailFailure('Relationships were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to relationships') reportSection('addConcept') o1.addConcept(c1) if o1.conceptExists(c1): reportDetail('Correctly added concept') else: reportFailure('Concept was not added') try: o1.addConcept(c1) reportDetailFailure('Concept already exists') except SelfException: reportDetail('Correctly denied adding concept that already exists') try: o1.addConcept('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied adding ill-formed concept') reportSection('removeConcept') o1.removeConcept(c1) if not o1.conceptExists(c1): reportDetail('Correctly removed concept') else: reportFailure('Concept was not removed') try: o1.removeConcept(c2) reportDetailFailure('Concept exists') except SelfException: reportDetail('Correctly denied removing concept that does not exist') try: o1.removeConcept('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied removing an ill-formed concept') o1.addConcept(c1) o1.addConcept(c2) o1.addRelationship(r1) try: o1.removeConcept(c1) reportDetailFailure('Concept is bound') except SelfException: reportDetail('Correctly denied removing concept that is bound') reportSection('removeAllConcepts') o1.removeRelationship(r1) o1.removeAllConcepts() if o1.numberOfConcepts() == 0: reportDetail('Correctly removed all concepts') else: reportDetailFailure('Concepts were not removed') o1.addConcept(c1) o1.addConcept(c2) o1.addRelationship(r1) try: o1.removeAllConcepts() reportDetailFailure('Concepts are bound') except SelfException: reportDetail('Correctly denied removing concepts that are bound') o1.removeRelationship(r1) o1.removeConcept(c2) o1.removeConcept(c1) reportSection('conceptExists') o1.addConcept(c1) if o1.conceptExists(c1): reportDetail('Correctly checked that concept exists') else: reportDetailFailure('Concept does not exist') if not o1.conceptExists(c2): reportDetail('Correctly checked that concept does not exist') else: reportDetailFailure('Concept exists') try: o1.conceptExists('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied checking existence of ill-formed concept') reportSection('numberOfConcepts') o1.addConcept(c2) if o1.numberOfConcepts() == 2: reportDetail('Correctly reported number of concepts') else: reportDetailFailure('Number of concepts is wrong') reportSection('iterateOverConcepts') o1.addConcept(c3) o1.iterateOverConcepts(reportConceptName) reportDetail('Correctly iterated over concepts') o1.iterateOverConcepts(reportConceptName, CONCEPT_NAME_1) reportDetail('Correctly iterated over concepts with given name') o1.iterateOverConcepts(reportConceptName, None, AnotherConcept) reportDetail('Correctly iterated over concepts with given concept class') o1.iterateOverConcepts(reportConceptName, CONCEPT_NAME_2, Concept) reportDetail('Correctly iterated over concepts with given name and concept class') try: o1.iterateOverConcepts(reportConceptName, None, SelfException) reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') try: o1.iterateOverConcepts(reportConceptName, None, 'An ill-formed concept class') reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') reportSection('addRelationship') o1.addRelationship(r1) o1.addRelationship(r2) o1.addRelationship(r3) if o1.numberOfRelationships() == 3: reportDetail('Correctly added relationship') else: reportDetailFailure('Relationship was not added') try: o1.addRelationship(r1) reportDetailFailure('Relationship already exists') except SelfException: reportDetail('Correctly denied addding relationship that already exists') try: o1.addRelationship('An ill-formed relationship') reportDetailFailure('Relationship is ill-formed') except SelfException: reportDetail('Correctly denied adding ill-formed relationship') try: o1.addRelationship(r4) reportDetailFalure('Relationship is not closed') except SelfException: reportDetail('Correctly denied adding relationship that is not closed') reportSection('removeRelationship') o1.removeRelationship(r3) if not o1.relationshipExists(r3): reportDetail('Correctly remove relationship') else: reportDetailFailure('Relationship was not removed') try: o1.removeRelationship(r3) reportDetailFailure('Relationship exists') except SelfException: reportDetail('Corectly denied removing relationship that does not exist') try: o1.removeRelationship('An ill-formed relationship') reportDetailFailure('Relationship is ill-formed') except SelfException: reportDetail('Correctly denied removing ill-formed relationship') reportSection('removeAllRelationships') o1.removeAllRelationships() if o1.numberOfRelationships() == 0: reportDetail('Correctly removed all relationships') else: reportDetailFailure('Relationships were not removed') reportSection('relationshipExists') o1.addRelationship(r1) if o1.relationshipExists(r1): reportDetail('Correctly checked that relationship exists') else: reportDetailFailure('Relationship does not exist') if not o1.relationshipExists(r3): reportDetail('Correctly checked that relationship does not exist') else: reportDetailFailure('Relationship exists') try: o1.relationshipExists('An ill-formed relationship') reportDetailFailure('Relationship is ill-formed') except SelfException: reportDetail('Correctly denied checking existance of ill-formed relationship') reportSection('numberOfRelationship') o1.addRelationship(r2) if o1.numberOfRelationships() == 2: reportDetail('Correctly reported number of relationships') else: reportDetailFailure('Number of relationships is wrong') reportSection('iterateOverRelationships') o1.addRelationship(r3) o1.iterateOverRelationships(reportConceptName) reportDetail('Correctly iterated over relationships') o1.iterateOverRelationships(reportConceptName, RELATIONSHIP_NAME_1) reportDetail('Correctly iterated over relationships with given name') o1.iterateOverRelationships(reportConceptName, None, AnotherRelationship) reportDetail('Correctly iterated over relationships with given relationship class') o1.iterateOverRelationships(reportConceptName, RELATIONSHIP_NAME_2, Relationship) reportDetail('Correctly iterated over relationshps with given name and concept class') try: o1.iterateOverRelationships(reportConceptName, None, SelfException) reportDetailFailure('Relationship class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed relationship class') try: o1.iterateOverRelationships(reportConceptName, None, 'An ill-formed relationship class') reportDetailFailure('Relationship class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed relationship class') reportSection('conceptIsBound') if o1.conceptIsBound(c1): reportDetail('Correctly checked that concept is bound') else: reportDetailFailure('Concept is not bound') if not o1.conceptIsBound(c4): reportDetail('Correctly checked that concept is not bound') else: reportDetailFailure('Concept is bound') try: o1.conceptIsBound('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied checking if an ill-formed concept is bound') reportSection('numberOfUnboundConcepts') o1.addConcept(c4) if o1.numberOfUnboundConcepts() == 1: reportDetail('Correctly reported number of unbound concepts') else: reportDetailFailure('Number of unbound concepts is wrong') reportSection('numberOfBoundConcepts') if o1.numberOfBoundConcepts() == 3: reportDetail('Correctly reported number of bound concepts') else: reportDetailFailure('Number of bound concepts is wrong') reportSection('iterateOverUnboundConcepts') o1.iterateOverUnboundConcepts(reportConceptName) reportDetail('Correctly iterated over unbound concepts') o1.iterateOverUnboundConcepts(reportConceptName, CONCEPT_NAME_1) reportDetail('Correctly iterated over unbound concepts with given name') o1.iterateOverUnboundConcepts(reportConceptName, None, AnotherConcept) reportDetail('Correctly iterated over unbound concepts with given concept class') o1.iterateOverUnboundConcepts(reportConceptName, CONCEPT_NAME_2, Concept) reportDetail('Correctly iterated over unbound concepts with given name and concept class') try: o1.iterateOverUnboundConcepts(reportConceptName, None, SelfException) reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') try: o1.iterateOverUnboundConcepts(reportConceptName, None, 'An ill-formed concept class') reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') reportSection('iterateOverBoundConcepts') o1.iterateOverBoundConcepts(reportConceptName) reportDetail('Correctly iterated over bound concepts') o1.iterateOverBoundConcepts(reportConceptName, CONCEPT_NAME_1) reportDetail('Correctly iterated over bound concepts with given name') o1.iterateOverBoundConcepts(reportConceptName, None, AnotherConcept) reportDetail('Correctly iterated over bound concepts with given concept class') o1.iterateOverBoundConcepts(reportConceptName, CONCEPT_NAME_2, Concept) reportDetail('Correctly iterated over bound concepts with given name and concept class') try: o1.iterateOverBoundConcepts(reportConceptName, None, SelfException) reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') try: o1.iterateOverBoundConcepts(reportConceptName, None, 'An ill-formed concept class') reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') # Blackboard unit test def testBlackboard(): reportHeader('Blackboard') reportSection('attributes') if b1.name == BLACKBOARD_NAME_1: reportDetail('Correctly set and retrieved name') else: reportDetailFailure('Name was not set or retrieved') try: s = b1.concepts reportDetailFailure('Concepts were directly accessed') except SelfException: reportDetail('Correctly denied direct access to concepts') try: b1.concepts = set() reportDetailFailure('Concepts were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to concepts') try: s = b1.conceptClasses reportDetailFailure('Concepts classes were directly accessed') except SelfException: reportDetail('Correctly denied direct access to concept classes') try: b1.conceptClasses = set() reportDetailFailure('Concept classes were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to concept classes') try: s = b1.publications reportDetailFailure('Publications were directly accessed') except SelfException: reportDetail('Correctly denied direct access to publications') try: b1.publications = set() reportDetailFailure('Publications were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to publications') try: s = b1.conceptSubscriptions reportDetailFailure('Subscriptions were directly accessed') except SelfException: reportDetail('Correctly denied direct access to subsubscriptions') try: b1.conceptSubscriptions = set() reportDetailFailure('Subscriptions were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to subscriptions') try: s = b1.classSubscriptions reportDetailFailure('Class subscriptions were directly accessed') except SelfException: reportDetail('Correctly denied direct access to class subscriptions') try: b1.classSubscriptions = set() reportDetailFailure('Class subscriptions were directly assigned') except SelfException: reportDetail('Correctly denied direct assignment to class subscriptions') reportSection('publishConcept') b1.publishConcept(a1, c1) if b1.conceptExists(c1): reportDetail('Correctly published concept') else: reportDetailFailure('Concept was not published') b1.subscribeToConceptClass(a2, AnotherConcept) b1.publishConcept(a1, c3) if len(b1.subscribers(c3)) == 1: reportDetail('Correctly subscribed to concept class instance') else: reportDetailFailure('Subscription failed') try: b1.publishConcept(a1, c1) reportDetailFailure('Concept already exists') except SelfException: reportDetail('Correctly denied adding concept that already exists') try: b1.publishConcept('An ill-formed agent', c1) reportDetailFailure('Agent is ill-formed') except SelfException: reportDetail('Correctly denied publishing ill-formed agent') try: b1.publishConcept(a1, 'An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied publishing ill-formed concept') reportSection('unpublishConcept') b1.unpublishConcept(c1) b1.unpublishConcept(c3) if not b1.conceptExists(c3): reportDetail('Correctly unpublished concept') else: reportDetailFailure('Concept was not unpublished') b1.publishConcept(a1, c1) b1.publishConcept(a2, c2) b1.publishConcept(a1, c3) b1.unpublishConcept() if b1.numberOfConcepts() == 0: reportDetail('Correctly unpublished all concepts') else: reportDetailFailure('Concepts were not unpublished') try: b1.unpublishConcept(c3) reportDetailFailure('Concept exists') except SelfException: reportDetail('Correctly denied unpublishing concept that does not exist') try: b1.unpublishConcept('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied unpublishing ill-formed concept') reportSection('publisher') b1.publishConcept(a1, c1) if b1.publisher(c1) == a1: reportDetail('Correctly returned publisher') else: reportDetailFailure('Publisher was not returned') try: b1.publisher(c2) reportDetailFailure('Concept does not exist') except SelfException: reportDetail('Correctly denied returning publisher of concept that does not exist') try: b1.publisher('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied returning publisher of ill-formed concept') reportSection('signalPublisher') b1.signalPublisher(Concept('A well-formed source'), Concept('A well-formed message'), c1) reportDetail('Correctly signaled publisher') b1.signalPublisher(Concept('A well-formed source'), Concept('A well-formed message')) reportDetail('Correctly signaled publishers') try: b1.signalPublisher(Concept('A well-formed source'), Concept('A well-formed message'), c2) reportDetailFailure('Concept does not exist') except SelfException: reportDetail('Correctly denied signaling a publisher of concept that does not exist') try: b1.signalPublisher(Concept('A well-formed source'), Concept('A well-formed message'), 'An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied signaling publisher of ill-formed concept') try: b1.signalPublisher('An ill-formed source', Concept('A well-formed message'), c1) reportDetail('Source is ill-formed') except SelfException: reportDetail('Correctly denied signaling publisher of ill-formed source') try: b1.signalPublisher(Concept('A well-formed source'), 'An ill-formed message', c1) reportDetailFailure('Message is ill-formed') except SelfException: reportDetail('Correctly denied signaling publisher of ill-formed message') reportSection('conceptExists') if b1.conceptExists(c1): reportDetail('Correctly checked that concept exists') else: reportDetailFailure('Concept does not exist') if not b1.conceptExists(c2): reportDetail('Correctly checked that concept does not exist') else: reportDetailFailure('Concept exists') try: b1.conceptExists('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied checking of ill-formed concept') reportSection('numberOfConcepts') b1.publishConcept(a2, c3) if b1.numberOfConcepts() == 2: reportDetail('Correctly reported number of concepts') else: reportDetailFailure('Number of concepts is wrong') reportSection('iterateOverConcepts') b1.iterateOverConcepts(reportConceptName) reportDetail('Correctly iterated over concepts') b1.iterateOverConcepts(reportConceptName, CONCEPT_NAME_1) reportDetail('Correctly iterated over concepts with given name') b1.iterateOverConcepts(reportConceptName, None, AnotherConcept) reportDetail('Correctly iterated over concepts with given concept class') b1.iterateOverConcepts(reportConceptName, CONCEPT_NAME_2, Concept) reportDetail('Correctly iterated over concepts with given name and concept class') try: b1.iterateOverConcepts(reportConceptName, None, SelfException) reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') try: b1.iterateOverConcepts(reportConceptName, None, 'An ill-formed concept class') reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied iterating over ill-formed concept class') reportSection('subscribeToConcept') b1.subscribeToConcept(a3, c3) if len(b1.subscribers(c3)) == 2: reportDetail('Correctly subscribed to concept') else: reportDetailFailure('Concept was not subscribed') try: b1.subscribeToConcept(a3, c3) reportDetailFailure('Concept is already subscribed') except SelfException: reportDetail('Correctly denied subscribing to concept more than once') try: b1.subscribeToConcept(a3, c4) reportDetailFailure('Concept exists') except SelfException: reportDetail('Correctly denied subscribing to concept that does not exist') try: b1.subscribeToConcept('An ill-formed agent', c3) reportDetailFailure('Agent is ill-formed') except SelfException: reportDetail('Correctly denied subscribing by ill-formed agent') try: b1.subscribeToConcept(a2, 'An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied subscribing to ill-formed concept') reportSection('unsubscribeFromConcept') b1.unsubscribeFromConcept() if len(b1.subscribers()) == 0: reportDetail('Correctly unsubscribed by from all concepts by all agents') else: reportDetailFailure('Concepts were not unsubscribed') b1.subscribeToConcept(a1, c1) b1.subscribeToConcept(a1, c3) b1.subscribeToConcept(a2, c1) b1.subscribeToConcept(a2, c3) b1.unsubscribeFromConcept(a1) if (len(b1.subscribers(c1)) == 1 and len(b1.subscribers(c3)) == 1): reportDetail('Correctly unsubscribed from all concepts by agent') else: reportDetailFailure('Concepts were not unsubscribed') b1.subscribeToConcept(a1, c1) b1.unsubscribeFromConcept(None, c1) if (len(b1.subscribers(c1)) == 0 and len(b1.subscribers(c3)) == 1): reportDetail('Correctly unsubscribied from concept by all agents') else: reportDetailFailure('Concepts were not unsubscribed') b1.unsubscribeFromConcept(a2, c3) if len(b1.subscribers(c3)) == 0: reportDetail('Correctly unsubscribed from concept by agent') else: reportDetailFailure('Concept was not unsubscribed') try: b1.unsubscribeFromConcept(None, c2) reportDetailFailure('Concept does not exist') except SelfException: reportDetail('Correctly denied unsubscribing from concept that does not exist') try: b1.unsubscribeFromConcept('An ill-formed agent', c1) reportDetailFailure('Agent is ill-formed') except SelfException: reportDetail('Correctly denied unsubscibing from ill-formed agent') try: b1.unsubscribeFromConcept(a1, 'An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied unsubscrbing from ill-formed concept') reportDetail('subscribers') b1.subscribeToConcept(a1, c1) b1.subscribeToConcept(a2, c1) if len(b1.subscribers(c1)) == 2: reportDetail('Correctly return subscribers') else: reportDetailFailure('Subscribers were not returned') if len(b1.subscribers(c3)) == 0: reportDetail('Correctly returned subscribers') else: reportDetailFailure('Subscribers were not returned') if len(b1.subscribers()) == 2: reportDetail('Correctly returned subscribers') else: reportDetailFailure('Subscribers were not returned') try: b1.subscribers(c2) reportDetailFailure('Concept exists') except SelfException: reportDetail('Correctly denied returning subscribers from concept that does not exist') try: b1.subscribers('An ill-formed concept') reportDetailFailure('Concept is ill-formed') except: reportDetail('Correctly denied returning subscribers from ill-formed concept') reportDetail('signalSubscribers') b1.signalSubscribers(Concept('A well-formed source'), Concept('A well-formed message'), c1) reportDetail('Correctly signaled subscribers') b1.signalSubscribers(Concept('A well-formed source'), Concept('A well-formed message')) reportDetail('Correctly signaled subscribers') try: b1.signalSubscribers(Concept('A well-formed source'), Concept('A well-formed message'), c2) reportDetailFailure('Concept does not exist') except SelfException: reportDetail('Correctly denied signaling subscribers of concept that does not exist') try: b1.signalSubscribers(Concept('A well-formed source'), Concept('A well-formed message'), 'An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied signaling subscribers of ill-formed concept') try: b1.signalSubscribers('An ill-formed source', Concept('A well-formed message'), c1) reportDetail('Source is ill-formed') except SelfException: reportDetail('Correctly denied signaling subscribers of ill-formed source') try: b1.signalSubscribers(Concept('A well-formed source'), 'An ill-formed message', c1) reportDetailFailure('Message is ill-formed') except SelfException: reportDetail('Correctly denied signaling subscribers of ill-formed message') reportSection('subscribeToConceptClass') b1.unsubscribeFromConceptClass() b1.subscribeToConceptClass(a1, Concept) b1.subscribeToConceptClass(a2, Concept) b1.subscribeToConceptClass(a3, AnotherConcept) if len(b1.classSubscribers()) == 3: reportDetail('Correctly subscribed to concept class') else: reportDetailFailure('Concept class was not subscribed') if len(b1.classSubscribers(Concept)) == 2: reportDetail('Correctly subscribed to concept class') else: reportDetail('Concept class was not subscribed') if len(b1.classSubscribers(AnotherConcept)) == 1: reportDetail('Correctly subscribed to concept class') else: reportDetailFailure('CConcept class was not subscribed') try: b1.subscribeToConceptClass(a1, Concept) reportDetailFailure('Concept class is already subscribed') except SelfException: reportDetail('Correctly denied subscribing to concept class more than once') try: b1.subscribeToConceptClass('An ill-formed agent', c3) reportDetailFailure('Agent is ill-formed') except SelfException: reportDetail('Correctly denied subscribing by ill-formed agent') try: b1.subscribeToConceptClass(a2, 'An ill-formed concept') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied subscribing to ill-formed concept') reportSection('unsubscribeFromConceptClass') b1.unsubscribeFromConceptClass() if len(b1.classSubscribers()) == 0: reportDetail('Correctly unsubscribed by from all concept classes by all agents') else: reportDetailFailure('Concept classes were not unsubscribed') b1.subscribeToConceptClass(a1, Concept) b1.subscribeToConceptClass(a1, AnotherConcept) b1.subscribeToConceptClass(a2, Concept) b1.subscribeToConceptClass(a3, AnotherConcept) b1.unsubscribeFromConceptClass(a1) if (len(b1.classSubscribers(Concept)) == 1 and len(b1.classSubscribers(AnotherConcept)) == 1): reportDetail('Correctly unsubcribed from all concept classes by agent') else: reportDetailFailure('Concept classes were not unsubscribed') b1.subscribeToConceptClass(a1, Concept) b1.unsubscribeFromConceptClass(None, Concept) if len(b1.classSubscribers(AnotherConcept)) == 1: reportDetail('Correctly unsubscribied from concept class by all agents') else: reportDetailFailure('Concept class was not unsubscribed') b1.unsubscribeFromConceptClass(a3, AnotherConcept) if len(b1.classSubscribers()) == 0: reportDetail('Correctly unsubscribed from concept class by agent') else: reportDetailFailure('Concept class was not unsubscribed') try: b1.unsubscribeFromConceptClass(None, c2) reportDetailFailure('Concept class does not exist') except SelfException: reportDetail('Correctly denied unsubscribing from concept class that does not exist') try: b1.unsubscribeFromConceptClass('An ill-formed agent', c1) reportDetailFailure('Agent is ill-formed') except SelfException: reportDetail('Correctly denied unsubscibing from ill-formed agent') try: b1.unsubscribeFromConceptClass(a1, 'An ill-formed concept class') reportDetailFailure('Concept is ill-formed') except SelfException: reportDetail('Correctly denied unsubscrbing from ill-formed concept class') reportSection('classSubscribers') b1.subscribeToConceptClass(a1, Concept) b1.subscribeToConceptClass(a2, Concept) if len(b1.classSubscribers(Concept)) == 2: reportDetail('Correctly return subscribers') else: reportDetailFailure('Subscribers were not returned') if len(b1.classSubscribers()) == 2: reportDetail('Correctly returned subscribers') else: reportDetailFailure('Subscribers were not returned') try: b1.classSubscribers(AnotherConcept) reportDetailFailure('Concept class exists') except SelfException: reportDetail('Correctly denied returning subscribers from concept class that does not exist') try: b1.classSubscribers('An ill-formed concept class') reportDetailFailure('Concept class is ill-formed') except: reportDetail('Correctly denied returning subscribers from ill-formed concept class') reportSection('signalConceptClassSubscribers') b1.subscribeToConceptClass(a3, AnotherConcept) b1.signalClassSubscribers(Concept('A well-formed source'), Concept('A well-formed message'), Concept) reportDetail('Correctly signaled subscribers') b1.signalClassSubscribers(Concept('A well-formed source'), Concept('A well-formed message')) reportDetail('Correctly signaled subscribers') try: b1.signalClassSubscribers(Concept('A well-formed source'), Concept('A well-formed message'), SelfException) reportDetailFailure('Concept class does not exist') except SelfException: reportDetail('Correctly denied signaling subscribers of concept class that does not exist') try: b1.signalClassSubscribers(Concept('A well-formed source'), Concept('A well-formed message'), 'An ill-formed concept class') reportDetailFailure('Concept class is ill-formed') except SelfException: reportDetail('Correctly denied signaling subscribers of ill-formed concept class') try: b1.signalClassSubscribers('An ill-formed source', Concept('A well-formed message'), Concept) reportDetail('Source is ill-formed') except SelfException: reportDetail('Correctly denied signaling subscribers of ill-formed source') try: b1.signalClassSubscribers(Concept('A well-formed source'), 'An ill-formed message', Concept) reportDetailFailure('Message is ill-formed') except SelfException: reportDetail('Correctly denied signaling subscribers of ill-formed message') # Agent unit test def testAgent(): reportHeader('Agent') reportSection('activity') a1.activity() reportDetail('Correctly carried out the activity') a1.activity(Concept('A well-formed parameter')) reportDetail('Correctly carried out the activity') try: a1.activity('An ill-formed parameter') reportDetailFailure('Parameters are ill-formed') except SelfException: reportDetail('Correctly denied carrying out activity with ill-formed parameters') reportSection('start') a1.start() reportDetail('Correctly started the agent activity') a1.start(Concept('A well-formed parameter')) reportDetail('Correctly started the agent activity') try: a1.start('An ill-formed parameter') reportDetailFailure('Parameters are ill-formed') except SelfException: reportDetail('Correctly denied starting activity with ill-formed parameters') reportSection('stop') a1.stop() reportDetail('Correctly stopped the agent activity') a1.stop(Concept('A well-formed parameter')) reportDetail('Correctly stopped the agent activity') try: a1.start('An ill-formed parameter') reportDetailFailure('Parameters are ill-formed') except SelfException: reportDetail('Correctly denied starting activity with ill-formed parameters') reportSection('pause') a1.pause() reportDetail('Correctly paused the agent activity') a1.pause(Concept('A well-formed parameter')) reportDetail('Correctly paused the agent activity') try: a1.start('An ill-formed parameter') reportDetailFailure('Parameters are ill-formed') except SelfException: reportDetail('Correctly denied starting activity with ill-formed parameters') reportSection('isAlive') if a1.isAlive(): reportDetail('Correctly checked that agent is alive') else: reportDetailFailure('Agent is not alive') reportSection('status') if a1.status().name == 'Status': reportDetail('Correctly checked agent status') else: reportDetailFailure('Agent status is wrong') reportSection('signal') a1.signal(Concept('A well-defined source'), Concept('A well-defined message')) reportDetail('Correctly signaled the agent') a1.signal(Concept('A well-defined source'), Concept('A well-defined message'), Concept('A well-defined parameter')) reportDetail('Correctly signaled the agent') try: a1.signal('An ill-defined source', Concept('A well-defined message'), Concept('A well-defined parameter')) reportDetailFailure('Source is ill-defined') except SelfException: reportDetail('Correctly denied connecting with ill-defined source') try: a1.signal(Concept('A well-defined source'), 'An ill-defined message', Concept('A well-defined parameter')) reportDetailFailure('Message is ill-defined') except SelfException: reportDetail('Correctly denied connecting with ill-defined message') try: a1.signal(Concept('A well-defined source'), Concept('A well-defined message'), 'An ill-defined parameter') reportDetailFailure('Parameters are ill-defined') except SelfException: reportDetail('Correctly denied connecting with ill-defined parameters') reportSection('connect') a1.connect(Relationship('A well-defined relationship', a1, a2)) reportDetail('Correctly connected the agent') a1.connect(Relationship('A well-defined relationship', a1, a2), Concept('A well-formed parameter')) reportDetail('Correctly connected the agent') try: a1.connect('An ill-formed relationship', Concept('A well-formed parameter')) reportDetailFailure('Channel is ill-formed') except SelfException: reportDetail('Correctly denied connecting with ill-formed channel') try: a1.connect(Relationship('A well-formed relationship', a1, a2), 'An ill-formed parameter') reportDetailFailure('Parameters are ill-defined') except SelfException: reportDetail('Correctly denied connecting wiht ill-formed parameters') # Test all of Self's foundational classes arguments = parseArguments() testConcept() testProperty() testRelationship() testOntology() testBlackboard() testAgent() # Clean up the output stream if reporting concisely if arguments.concise == True: print()
58,889
16,207
''' Say you have an array for which the ith element is the price of a given stock on day i. Design an algorithm to find the maximum profit. You may complete at most two transactions. Note: You may not engage in multiple transactions at the same time (i.e., you must sell the stock before you buy again). Example 1: Input: [3,3,5,0,0,3,1,4] Output: 6 Explanation: Buy on day 4 (price = 0) and sell on day 6 (price = 3), profit = 3-0 = 3. Then buy on day 7 (price = 1) and sell on day 8 (price = 4), profit = 4-1 = 3. ''' class Solution(object): def maxProfit(self, prices): """ :type prices: List[int] :rtype: int """ if len(prices) < 2: return 0 dp = [[0 for _ in range(len(prices))] for _ in range(3)] for i in range(1,3): maxDiff = -prices[0] for j in range(1,len(prices)): dp[i][j] = max(dp[i][j-1], prices[j] + maxDiff) maxDiff = max(maxDiff, dp[i-1][j] -prices[j]) return dp[2][len(prices)-1]
1,042
390
import json from package.query_db import query from package.dictionary_to_list import dictionary_to_list from package.lambda_exception import LambdaException from boto3 import client as boto3_client def verify_admin(event, context): user_id = int(event['user_id']) user_id_dic = {} if user_id == None: #Making sure user_id was passed raise LambdaException("400: user_id was not given") user_id_dic['user_id'] = user_id sql_parameters = dictionary_to_list(user_id_dic) sql_select = """SELECT users.id FROM users WHERE users.id = :user_id""" #This query is ensuring that the user exists response = query(sql_select, sql_parameters) if response['records'] == []: #Returning error if user does not exist raise LambdaException("404: user does not exist") sql_select = """SELECT users.id FROM users WHERE users.id = :user_id and is_admin = true""" #This query is ensuring user is not already an admin response = query(sql_select, sql_parameters) if response['records'] != []: #Returning error if user is already an admin raise LambdaException("405: user is already an admin") else: sql_update = """UPDATE users SET is_admin = true WHERE users.id = :user_id""" response = query(sql_update, sql_parameters) sql_insert = """INSERT INTO admins(admin_id, user_id, is_pending) VALUES(:user_id, :user_id, false) """ response = query(sql_insert, sql_parameters) # send approval email lambda_client = boto3_client('lambda') email_event = { "user_id": user_id, "approved_role": "admin" } try: response = lambda_client.invoke(FunctionName="approval_email", InvocationType='Event', Payload=json.dumps(email_event)) except Exception as e: raise LambdaException("404: Unable to send approval email " + str(e)) return{ "statusCode": 200 }
2,077
591
#!/usr/bin/env python3 """ Created on September 21 2019 @author: Melchior du Lac @description: Wrap rp2paths into a docker """ import argparse import tempfile import os import logging import shutil import docker import glob def main(rp_pathways, rp2paths_pathways, rp2paths_compounds, timeout=30, max_steps=0, max_paths=150, unfold_compounds=False): """Call the docker to run rp2paths :param rp_pathways: The path to the results RetroPath2.0 scope file :param rp2paths_pathways: The path to the results rp2paths out_paths file :param rp2paths_compounds: The path to the results rp2paths compounds file :param timeout: The timeout of the function in minutes (Default: 90) :param max_steps: The maximal number of steps WARNING: not used (Default: 0, ie. infinite) :param max_paths: The maximal number of pathways to return WARNING: not used (Default: 150) :param unfold_compounds: not sure WARNING: not used (Default: False) :param rp_pathways: str :param rp2paths_pathways: str :param rp2paths_compounds: str :param timeout: int :param max_steps: int :param max_paths: int :param unfold_compounds: bool :rtype: None :return: None """ docker_client = docker.from_env() image_str = 'brsynth/rp2paths-standalone' try: image = docker_client.images.get(image_str) except docker.errors.ImageNotFound: logging.warning('Could not find the image, trying to pull it') try: docker_client.images.pull(image_str) image = docker_client.images.get(image_str) except docker.errors.ImageNotFound: logging.error('Cannot pull image: '+str(image_str)) exit(1) with tempfile.TemporaryDirectory() as tmpOutputFolder: if os.path.exists(rp_pathways): shutil.copy(rp_pathways, tmpOutputFolder+'/rp_pathways.csv') command = ['python', '/home/tool_rp2paths.py', '-rp_pathways', '/home/tmp_output/rp_pathways.csv', '-rp2paths_compounds', '/home/tmp_output/rp2paths_compounds.csv', '-rp2paths_pathways', '/home/tmp_output/rp2paths_pathways.csv', '-timeout', str(timeout), '-max_steps', str(max_steps), '-max_paths', str(max_paths), '-unfold_compounds', str(unfold_compounds)] container = docker_client.containers.run(image_str, command, detach=True, stderr=True, volumes={tmpOutputFolder+'/': {'bind': '/home/tmp_output', 'mode': 'rw'}}) container.wait() err = container.logs(stdout=False, stderr=True) err_str = err.decode('utf-8') if 'ERROR' in err_str: print(err_str) elif 'WARNING' in err_str: print(err_str) if not os.path.exists(tmpOutputFolder+'/rp2paths_compounds.csv') or not os.path.exists(tmpOutputFolder+'/rp2paths_pathways.csv'): print('ERROR: Cannot find the output file: '+str(tmpOutputFolder+'/rp2paths_compounds.csv')) print('ERROR: Cannot find the output file: '+str(tmpOutputFolder+'/rp2paths_pathways.csv')) else: shutil.copy(tmpOutputFolder+'/rp2paths_pathways.csv', rp2paths_pathways) shutil.copy(tmpOutputFolder+'/rp2paths_compounds.csv', rp2paths_compounds) container.remove() else: logging.error('Cannot find one or more of the input files: '+str(rp_pathways)) exit(1) if __name__ == "__main__": parser = argparse.ArgumentParser('Enumerate the individual pathways from the results of Retropath2') parser.add_argument('-rp_pathways', type=str) parser.add_argument('-rp2paths_pathways', type=str) parser.add_argument('-rp2paths_compounds', type=str) parser.add_argument('-max_steps', type=int, default=0) parser.add_argument('-timeout', type=int, default=30) parser.add_argument('-max_paths', type=int, default=150) parser.add_argument('-unfold_compounds', type=str, default='False') params = parser.parse_args() if params.timeout<0: logging.error('Timeout cannot be <0 :'+str(params.timeout)) exit(1) main(params.rp_pathways, params.rp2paths_pathways, params.rp2paths_compounds, params.timeout, params.max_steps, params.max_paths, params.unfold_compounds)
4,821
1,422
import pytest from framework.services.booker_client import BookerClient booker_client = BookerClient() @pytest.fixture() def clear_env(): pass
150
49
spiFile = open('spiflash.bin','wb') # 128KB is reserved for bitstream bitFile = open('../bitstream/mf8a18_rv32i.bin','rb') bitData = bitFile.read(0x20000) riscvFile = open('riscv.bin','rb') riscvData = riscvFile.read(32768) spiFile.write(bitData) spiFile.seek(0x20000) spiFile.write(riscvData) nullData = bytearray([0]) spiFile.seek(0x27fff) spiFile.write(nullData) spiFile.close bitFile.close
402
194
"""Options for saving user prefences, etc. """ import json import os import pyvista class RcParams(dict): """Internally used class to manage the rcParams""" filename = os.path.join(pyvista.USER_DATA_PATH, 'rcParams.json') def save(self): with open(self.filename, 'w') as f: json.dump(self, f) return def load(self): with open(self.filename, 'r') as f: data = json.load(f) self.update(data) def __setitem__(self, key, value): dict.__setitem__(self, key, value) self.save() # The options rcParams = RcParams( dark_mode=False, ) # Load user prefences from last session if none exist, save defaults try: rcParams.load() except: rcParams.save()
753
252
"""This script validates an ArcGIS Enterprise deployment to ensure it is configured properly with all the required components such as Portal for ArcGIS, ArcGIS Server, ArcGIS Data Store and the associated configuration. Designed for ArcGIS Enterprise 10.5 and higher.""" # Author: Philip Heede <pheede@esri.com> # Last modified: 2017-02-18 import os import sys import ssl import socket import urllib.request import getopt import getpass import json import traceback def main(argv): parameters = parseInputParameters(argv) portalUrl = parameters['portalUrl'] token = parameters['token'] if token == '': adminUsername = parameters['adminUsername'] adminPassword = parameters['adminPassword'] token = generateToken(adminUsername, adminPassword, portalUrl) if token == 'Failed': print('Invalid administrator username or password.') sys.exit(1) portalSelf = getPortalSelf(portalUrl, token) supportsHostedServices = portalSelf['supportsHostedServices'] supportsSceneServices = portalSelf['supportsSceneServices'] # check analysis tools helper service registration and valid url helperServices = portalSelf['helperServices'] analysisHelperServiceRegistered = False if 'analysis' in helperServices: if 'url' in helperServices['analysis']: analysisHelperService = portalSelf['helperServices']['analysis']['url'] analysisHelperServiceRegistered = analysisHelperService != '' geoanalyticsHelperServiceRegistered = False if 'geoanalytics' in portalSelf['helperServices']: if 'url' in helperServices['geoanalytics']: geoanalyticsHelperService = helperServices['geoanalytics']['url'] geoanalyticsHelperServiceRegistered = geoanalyticsHelperService != '' rasterAnalyticsHelperServiceRegistered = False if 'rasterAnalytics' in portalSelf['helperServices']: if 'url' in helperServices['rasterAnalytics']: rasterAnalyticsHelperService = helperServices['rasterAnalytics']['url'] rasterAnalyticsHelperServiceRegistered = rasterAnalyticsHelperService != '' # enumerate federated servers and find hosting server federatedServers = getFederatedServers(portalUrl, token) hostingServer = None for server in federatedServers: if 'serverRole' in server: serverRole = server['serverRole'] if serverRole == 'HOSTING_SERVER': hostingServer = server print() print("ArcGIS Enterprise deployment characteristics") print("- Hosting server configured: %s" % (hostingServer is not None)) if hostingServer is None: print("-- WARNING: lack of a hosting server will prevent many functions from working") else: hostingServerValid, validationMsgs = validateHostingServer(portalUrl, hostingServer['id'], token) if not hostingServerValid: print('-- ERROR: unable to validate hosting server') for msg in validationMsgs: print('-- ' + msg) if hostingServerValid: hasRelationalDataStore = checkArcGISDataStoreRelational(hostingServer['adminUrl'], hostingServer['url'], token) print("- ArcGIS Data Store (relational) configured with hosting server: %s" % hasRelationalDataStore) if not hasRelationalDataStore: print("-- WARNING: you must use ArcGIS Data Store to configure a relational database") print('- Analysis Tools helper service is configured: %s' % analysisHelperServiceRegistered) if not analysisHelperServiceRegistered: print('-- WARNING: analysis tools helper service not configured') analysisServiceStarted = checkAnalysisServices(hostingServer['url'], token) print("- Hosting server's spatial analysis service is started and available: %s" % analysisServiceStarted) if not analysisServiceStarted: print("-- WARNING: analysis service not started or unreachable") print("- Hosted feature services are supported: %s" % supportsHostedServices) if not supportsHostedServices: print("-- WARNING: this indicates a lack of ArcGIS Data Store configured with the relational data store type") print("- Scene services are supported: %s" % supportsSceneServices) if not supportsSceneServices: print("-- WARNING: this indicates a lack of ArcGIS Data Store (tile cache)") print('- GeoAnalytics configured: %s' % geoanalyticsHelperServiceRegistered) print('- Raster Analytics configured: %s' % rasterAnalyticsHelperServiceRegistered) def parseInputParameters(argv): currentHost = socket.getfqdn().lower() currentDir = os.getcwd() portalHost = '' context = '' adminUsername = '' adminPassword = '' outputDir = '' token = '' if len(sys.argv) > 0: try: opts, args = getopt.getopt(argv, "?hn:c:u:p:t:", ("help", "portalurl=", "context=", "user=", "password=", "token=", "ignoressl")) except: print('One or more invalid arguments') print('validatebasedeployment.py [-n <portal hostname>] [-c <portal context>] [-u <admin username>] [-p <admin password>] [-t <token>]') sys.exit(2) for opt, arg in opts: if opt in ('-n', '--portalurl'): portalHost = arg elif opt in ('-c', '--context'): context = arg elif opt in ('-u', '--user'): adminUsername = arg elif opt in ('-p', '--password'): adminPassword = arg elif opt == '--ignoressl': # disable SSL certificate checking to avoid errors with self-signed certs # this is NOT a generally recommended practice _create_unverified_https_context = ssl._create_unverified_context ssl._create_default_https_context = _create_unverified_https_context elif opt in ('-t', '--token'): token = arg elif opt in ('-h', '-?', '--help'): print('validatebasedeployment.py [-n <portal hostname>] [-c <portal context>] [-u <admin username>] [-p <admin password>] [-t <token>]') sys.exit(0) # Prompt for portal hostname if portalHost == '': portalHost = input('Enter ArcGIS Enterprise FQDN [' + currentHost + ']: ') if portalHost == '': portalHost = currentHost # Prompt for portal context if context == '': context = input('Enter context of the portal instance [\'arcgis\']: ') if context == '': context = 'arcgis' # Prompt for admin username if adminUsername == '' and token == '': while adminUsername == '': adminUsername = input('Enter administrator username: ') # Prompt for admin password if adminPassword == '' and token == '': while adminPassword == '': adminPassword = getpass.getpass(prompt='Enter administrator password: ') portalUrl = 'https://' + portalHost + '/' + context parameters = {'adminPassword':adminPassword, 'adminUsername':adminUsername, 'portalUrl':portalUrl, 'token':token} return parameters def validateHostingServer(portalUrl, hostingServerID, token): params = {'token':token, 'f':'pjson', 'types':'egdb'} request = urllib.request.Request(portalUrl + '/portaladmin/federation/servers/' + hostingServerID + '/validate?' + urllib.parse.urlencode(params)) try: response = urllib.request.urlopen(request) result = json.loads(response.read().decode('utf-8')) msgs = [] if 'messages' in result: msgs = result['messages'] if 'status' in result and result['status'] == 'success': return True, msgs except: pass return False, msgs def checkArcGISDataStoreRelational(serverAdminUrl, serverUrl, portalToken): params = {'token':portalToken, 'f':'pjson', 'types':'egdb'} request = urllib.request.Request(serverAdminUrl + '/admin/data/findItems', urllib.parse.urlencode(params).encode('ascii')) try: response = urllib.request.urlopen(request) except: request = urllib.request.Request(serverUrl + '/admin/data/findItems', urllib.parse.urlencode(params).encode('ascii')) try: response = urllib.request.urlopen(request) print('-- WARNING: hosting server administrative endpoint not') print(' accessible from this machine; this may cause') print(' publishing issues from ArcGIS Pro') except: print('-- ERROR: unable to reach hosting server administrative endpoint') print(' maybe the administrative endpoint is only accessible internally?') egdbs = json.loads(response.read().decode('utf-8')) if 'error' in egdbs: return False else: managedegdb = None for egdb in egdbs['items']: if egdb['info']['isManaged']: managedegdb = egdb if managedegdb is None: return False return managedegdb['provider'] == 'ArcGIS Data Store' def checkAnalysisServices(serverUrl, portalToken): params = {'token':portalToken, 'f':'json'} request = urllib.request.Request(serverUrl + '/rest/services/System/SpatialAnalysisTools/GPServer?%s' % urllib.parse.urlencode(params)) try: response = urllib.request.urlopen(request) serviceInfo = json.loads(response.read().decode('utf-8')) if 'error' in serviceInfo: return False else: return True except: return False def getFederatedServers(portalUrl, token): params = {'token':token, 'f':'json'} request = urllib.request.Request(portalUrl + '/portaladmin/federation/servers?%s' % urllib.parse.urlencode(params)) response = urllib.request.urlopen(request) federatedServers = json.loads(response.read().decode('utf-8')) if 'servers' not in federatedServers: print('Unable to enumerate federated servers. Not an administrator login?') sys.exit(1) return federatedServers['servers'] def getPortalSelf(portalUrl, token): params = {'token':token, 'f':'json'} request = urllib.request.Request(portalUrl + '/sharing/portals/self', urllib.parse.urlencode(params).encode('ascii')) response = urllib.request.urlopen(request) portalSelf = json.loads(response.read().decode('utf-8')) return portalSelf def generateToken(username, password, portalUrl): params = {'username':username, 'password':password, 'referer':portalUrl, 'f':'json'} try: request = urllib.request.Request(portalUrl + '/sharing/rest/generateToken', urllib.parse.urlencode(params).encode('ascii')) response = urllib.request.urlopen(request) genToken = json.loads(response.read().decode('utf-8')) if 'token' in genToken.keys(): return genToken.get('token') else: return 'Failed' except urllib.error.URLError as urlError: print('Unable to access ArcGIS Enterprise deployment at ' + portalUrl) if isinstance(urlError.reason, ssl.SSLError): print("SSL certificate validation error. Maybe you're using a self-signed certificate?") print("Pass the --ignoressl parameter to disable certificate validation") else: print(urlError.reason) sys.exit(1) except Exception as ex: print('Unable to access ArcGIS Enterprise deployment at ' + portalUrl) exc_type, exc_value, exc_traceback = sys.exc_info() print("*** print_exception:") traceback.print_exception(exc_type, exc_value, exc_traceback, limit=2, file=sys.stdout) sys.exit(0) if not sys.version_info >= (3, 4): print('This script requires Python 3.4 or higher: found Python %s.%s' % sys.version_info[:2]) if __name__ == "__main__": sys.exit(main(sys.argv[1:]))
11,978
3,203
import itertools import re import math from typing import List, Tuple def ints(text: str) -> Tuple[int, ...]: "Return a tuple of all ints in a string" return tuple(map(int, re.findall(r'-?\b\d+\b', text))) def powerset(iterable): "powerset([1,2,3]) --> () (1,) (2,) (3,) (1,2) (1,3) (2,3) (1,2,3)" s = list(iterable) return itertools.chain.from_iterable(itertools.combinations(s, r) for r in range(len(s)+1)) def manhattan(p: Tuple[int, ...], q=itertools.repeat(0)) -> Tuple[int, ...]: "Return the manhattan distance between 2 (multi-dimensional) points" return sum([abs(a-b) for a, b in zip(p, q)]) def king_distance(p: Tuple[int, ...], q=itertools.repeat(0)) -> Tuple[int, ...]: "Return thenNumber of chess King moves between two points" return max(abs(a - b) for a, b in zip(p, q)) def neighbors4(p: Tuple[int, int]) -> List[Tuple[int, int]]: "Return the 4 neighboring cells for a given position" x, y = p return [ (x, y-1), (x, y+1), (x-1, y), (x+1, y) ] def neighbors8(p: Tuple[int, int]) -> List[Tuple[int, int]]: "Return the 8 neighboring cells for a given position" x, y = p return [ (x-1, y-1), (x, y-1), (x+1, y-1), (x-1, y), (x+1, y), (x-1, y+1), (x, y+1), (x+1, y+1) ] def neighbors_cube(p: Tuple[int, int, int]) -> List[Tuple[int, int, int]]: "Return the 26 neighboring cells for a given position in a 3d cube" x, y, z = p n = [] for i in range(-1, 2): for j in range(-1, 2): for k in range(-1, 2): if (i, j, k) != (0, 0, 0): n.append((x+i, y+j, z+k)) return n def neighbors_cube4(p: Tuple[int, int, int, int]) -> List[Tuple[int, int, int, int]]: "Return the 80 neighboring cells for a given position in a 4-d cube" x, y, z, w = p n = [] for i in range(-1, 2): for j in range(-1, 2): for k in range(-1, 2): for l in range(-1, 2): if (i, j, k, l) != (0, 0, 0, 0): n.append((x+i, y+j, z+k, w+l)) return n moves = { 'n': lambda p: (p[0], p[1]-1), 's': lambda p: (p[0], p[1]+1), 'e': lambda p: (p[0]+1, p[1]), 'w': lambda p: (p[0]-1, p[1]), } left_turn = { 'n': 'w', 's': 'e', 'e': 'n', 'w': 's', } right_turn = { 'n': 'e', 's': 'w', 'e': 's', 'w': 'n', } opposite = { 'n': 's', 's': 'n', 'e': 'w', 'w': 'e', } facing_dir = { 'n': (0, -1), 's': (0, 1), 'e': (1, 0), 'w': (-1, 0), } origin = (0, 0) hex_origin = (0, 0, 0) hex_moves = { 'ne': lambda p: (p[0]+1, p[1], p[2]-1), 'nw': lambda p: (p[0], p[1]+1, p[2]-1), 'se': lambda p: (p[0], p[1]-1, p[2]+1), 'sw': lambda p: (p[0]-1, p[1], p[2]+1), 'w': lambda p: (p[0]-1, p[1]+1, p[2]), 'e': lambda p: (p[0]+1, p[1]-1, p[2]), } def hex_neighbors(p: Tuple[int, int, int]) -> List[Tuple[int, int, int]]: return [move(p) for move in hex_moves.values()] def add_pos(a: Tuple[int, int], b: Tuple[int, int], factor: int = 1) -> Tuple[int, int]: "Adds two position tuples" return (a[0]+b[0]*factor, a[1]+b[1]*factor) def sub_pos(a: Tuple[int, int], b: Tuple[int, int]) -> Tuple[int, int]: "Subtracts the position tuple b from a" return (a[0]-b[0], a[1]-b[1]) def mult_pos(a: Tuple[int, int], factor: int) -> Tuple[int, int]: "Multiplies a position tuple with a given factor" return (a[0]*factor, a[1]*factor) def rot_left(pos: Tuple[int, int], rel: Tuple[int, int] = origin) -> Tuple[int, int]: "Rotates a position 90 degrees left (counter clock-wise) relative to the given location (default origin)" rel_pos = sub_pos(pos, rel) new_pos = (rel_pos[1], -rel_pos[0]) return add_pos(new_pos, rel) def rot_right(pos: Tuple[int, int], rel: Tuple[int, int] = origin) -> Tuple[int, int]: "Rotates a position 90 degrees right (clock-wise) relative to the given location (default origin)" rel_pos = sub_pos(pos, rel) new_pos = (-rel_pos[1], rel_pos[0]) return add_pos(new_pos, rel) def min_max(lst: List[Tuple[int, ...]]) -> Tuple[int, ...]: "Returns the min and max values for every index in the given list of tuples" return tuple((min(e), max(e)) for e in zip(*lst)) def mod1(a: int, b: int) -> int: "Returns 1-based modulo" return 1 + (a-1) % b
4,445
1,926
import math import time from max30105 import MAX30105, HeartRate import smbus from bme280 import BME280 import socket #from matplotlib import pyplot as plt class DataPoint(): def __init__(self,value,time): self.time_stamp = time self.value = value class Device(): def __init__(self): self.humidity = [] self.temperature = [] self.smoke_level = [] self.mean_size = 100 self.identifier = "0,0" def setup_network(self): self.network = socket.socket(socket.AF_INET,socket.SOCK_STREAM) connected = False while not connected: try: self.network.connect(("192.168.88.167",25565)) connected = True except: a = 1 def upload_data(self): network_string = (#str(round(self.calculate_humidity_trend(),5)) + "," + str(round(self.humidity[-1].value,5)) + "," + #str(round(self.calculate_temperature_trend(),5)) + "," + str(round(self.temperature[-1].value,5)) + "," + #str(round(self.calculate_smoke_level_trend(),5)) + "," + str(round(self.smoke_level[-1].value,5)) + "," + str(round(self.pressure.value,5)) + "," + str(self.identifier)) network_string = network_string.encode() self.network.sendall(network_string) def update(self): dev.get_smoke_data() dev.get_humi_temp_data() def setup_particle_sensor(self): self.MAX30105 = MAX30105() self.MAX30105.setup(leds_enable=3) self.MAX30105.set_led_pulse_amplitude(1,0.0) self.MAX30105.set_led_pulse_amplitude(2,0.0) self.MAX30105.set_led_pulse_amplitude(3,12.5) self.MAX30105.set_slot_mode(1,"red") self.MAX30105.set_slot_mode(2,"ir") self.MAX30105.set_slot_mode(3,"green") self.MAX30105.set_slot_mode(4,"off") self.hr = HeartRate(self.MAX30105) def setup_temp_humi_sensor(self): bus = smbus.SMBus(1) self.bme280 = BME280(i2c_dev=bus) def setup_sensors(self): self.setup_particle_sensor() self.setup_temp_humi_sensor() def get_smoke_data(self): data = [] for i in range(self.mean_size*3+1): samples = self.MAX30105.get_samples() if samples is not None: for sample in samples: r = samples[2] & 0xff d = self.hr.low_pass_fir(r) data.append(d) mean = sum(data)/(self.mean_size*3) self.smoke_level.append(DataPoint(mean,time.time)) def get_humi_temp_data(self): temp_data = [] humi_data = [] pres_data = [] for i in range(self.mean_size): temp_data.append(self.bme280.get_temperature()) humi_data.append(self.bme280.get_humidity()) pres_data.append(self.bme280.get_pressure()) mean_temp = sum(temp_data)/self.mean_size mean_humi = sum(humi_data)/self.mean_size mean_pres = sum(pres_data)/self.mean_size self.humidity.append(DataPoint(mean_humi,time.time())) self.temperature.append(DataPoint(mean_temp,time.time())) self.pressure = DataPoint(mean_pres,time.time()) """def calculate_humidity_trend(self): return self.lin_reg(self.humidity) def calculate_temperature_trend(self): return self.lin_reg(self.temperature) def calculate_smoke_level_trend(self): return self.lin_reg(self.smoke_level) def lin_reg(self,data_set): x = 0 Sxy = 0 Sx = 0 Sx2 = 0 Sy = 0 Sy2 = 0 sample_size = len(data_set) for y in data_set: y=y.value x += 1 Sxy += x * y Sx += x Sx2 += x**2 Sy += y Sy2 += y**2 lin_reg = ((sample_size*Sxy)-(Sx*Sy))/((sample_size*Sx2)-(Sx)**2) return lin_reg""" dev = Device() dev.setup_sensors() dev.setup_network() for i in range(2): dev.update() while True: try: dev.update() dev.upload_data() print("sending_data") except: dev.setup_network()
4,431
1,566
from .client_http import HttpClient from .client_tcp import TcpClient from .client_udp import UdpClient from .client import Client class ClientFactory: """ Client factory for different types of protocols """ def getClient(self, ip, port=None, protocol="http") -> Client: """Creates new client Args: ip (str): robot IP port (int, optional): robot port. Defaults to None. protocol (str, optional): communication protocol. Defaults to "http". Returns: Client: requested client """ if protocol == "http": return HttpClient(ip) if port is None else HttpClient(ip, port) elif protocol == "tcp": return TcpClient(ip) if port is None else TcpClient(ip, port) elif protocol == "udp": return UdpClient(ip) if port is None else UdpClient(ip, port) else: return None def getClientClass(self, protocol="http"): """Get client class Args: protocol (str, optional): communication protocol. Defaults to "http". Returns: Client: client class """ if protocol == "http": return HttpClient elif protocol == "tcp": return TcpClient elif protocol == "udp": return UdpClient else: return None
1,417
355
import scrapy ''' 现在您已经知道如何从页面中提取数据,我们来看看如何跟踪它们的链接。 首先是提取我们想要跟踪的页面的链接。检查我们的页面,我们可以看到有一个链接到下一个页面与以下标记: <ul class="pager"> <li class="next"> <a href="/page/2/">Next <span aria-hidden="true">&rarr;</span></a> </li> </ul> 我们可以尝试在shell中提取它: >>> response.css('li.next a').extract_first() '<a href="/page/2/">Next <span aria-hidden="true">→</span></a>' 这得到了锚点元素,但是我们需要该属性href。为此,Scrapy支持CSS扩展,您可以选择属性内容,如下所示: >>> response.css('li.next a::attr(href)').extract_first() '/page/2/' 让我们看看现在我们的蜘蛛修改为递归地跟随链接到下一页,从中提取数据: ''' import scrapy class QuotesSpider(scrapy.Spider): name = "demo5" start_urls = [ 'http://quotes.toscrape.com/page/1/', ] def parse(self, response): for quote in response.css('div.quote'): yield { 'text': quote.css('span.text::text').extract_first(), 'author': quote.css('small.author::text').extract_first(), 'tags': quote.css('div.tags a.tag::text').extract(), } next_page = response.css('li.next a::attr(href)').extract_first() if next_page is not None: next_page = response.urljoin(next_page) yield scrapy.Request(next_page, callback=self.parse) ''' 现在,在提取数据之后,该parse()方法会查找到下一页的链接,使用该urljoin()方法构建完整的绝对URL (由于链接可以是相对的),并且向下一页产生一个新的请求,将其注册为回调以处理下一页的数据提取,并保持爬行遍历所有页面。 您在这里看到的是Scrapy的以下链接机制:当您以回调方式生成请求时,Scrapy将安排该请求发送,并注册一个回调方法,以在该请求完成时执行。 使用它,您可以根据您定义的规则构建复杂的跟踪链接,并根据访问页面提取不同类型的数据。 在我们的示例中,它创建一个循环,跟随到所有到下一页的链接,直到它找不到一个方便的抓取博客,论坛和其他站点分页。 ''' ''' 启动项目 scrapy crawl demo5 '''
1,569
961
import pytest import numpy as np import pandas as pd from scipy.special import binom import os import sys sys.path.insert(0, "..") from autogenes import objectives as ga_objectives def test_distance(): arr = np.ones((3,3)) assert ga_objectives.distance(arr) == 0 arr = np.identity(3) assert np.isclose(ga_objectives.distance(arr), 3 * np.sqrt(2)) def test_correlation(): arr = np.ones((3,3)) # Should'nt throw a warning assert ga_objectives.correlation(arr) == 0 arr = np.zeros((3,3)) assert ga_objectives.correlation(arr) == 0 arr = np.identity(5) # Let i!=j # cov(e_i, e_j) = 1/4 * ( 2 * (4/5) * (-1/5) + 3 * (-1/5) * (-1/5)) = -1/20 # var(e_i) = 1/4 * ( (4/5)^2 + 4 * (-1/5)^2) = 1/5 # (i,j)th entry in the corrcoef matrix is (-1/20)/sqrt(var(e_i) * var(e_j)) = -1/4 # Result is 1/4 * (5 over 2) assert np.isclose(ga_objectives.correlation(arr), 1/4*binom(5,2))
909
407
# -*- coding: utf-8 -*- # Copyright (c) 2014-2017, NVIDIA CORPORATION. All rights reserved. import os import flask from flask_wtf import FlaskForm import wtforms from wtforms import validators from digits.config import config_value from digits.device_query import get_device, get_nvml_info from digits import utils from digits.utils import sizeof_fmt from digits.utils.forms import validate_required_iff from digits import frameworks from flask_babel import lazy_gettext as _ class ModelForm(FlaskForm): # Methods def selection_exists_in_choices(form, field): found = False for choice in field.choices: if choice[0] == field.data: found = True if not found: raise validators.ValidationError(_("Selected job doesn't exist. Maybe it was deleted by another user.")) def validate_NetParameter(form, field): fw = frameworks.get_framework_by_id(form['framework'].data) try: # below function raises a BadNetworkException in case of validation error fw.validate_network(field.data) except frameworks.errors.BadNetworkError as e: raise validators.ValidationError(_('Bad network: %(message)s', message=e.message)) def validate_file_exists(form, field): from_client = bool(form.python_layer_from_client.data) filename = '' if not from_client and field.type == 'StringField': filename = field.data if filename == '': return if not os.path.isfile(filename): raise validators.ValidationError(_('Server side file, %(filename)s, does not exist.', filename=filename)) def validate_py_ext(form, field): from_client = bool(form.python_layer_from_client.data) filename = '' if from_client and field.type == 'FileField': filename = flask.request.files[field.name].filename elif not from_client and field.type == 'StringField': filename = field.data if filename == '': return (root, ext) = os.path.splitext(filename) if ext != '.py' and ext != '.pyc': raise validators.ValidationError(_('Python file, %(filename)s, needs .py or .pyc extension.', filename=filename)) # Fields # The options for this get set in the view (since they are dynamic) dataset = utils.forms.SelectField( _('Select Dataset'), choices=[], tooltip=_("Choose the dataset to use for this model.") ) python_layer_from_client = utils.forms.BooleanField( _('Use client-side file'), default=False, ) python_layer_client_file = utils.forms.FileField( _('Client-side file'), validators=[ validate_py_ext ], tooltip=_("Choose a Python file on the client containing layer definitions.") ) python_layer_server_file = utils.forms.StringField( _('Server-side file'), validators=[ validate_file_exists, validate_py_ext ], tooltip=_("Choose a Python file on the server containing layer definitions.") ) train_epochs = utils.forms.IntegerField( _('Training epochs'), validators=[ validators.NumberRange(min=1) ], default=30, tooltip=_("How many passes through the training data?") ) snapshot_interval = utils.forms.FloatField( _('Snapshot interval (in epochs)'), default=1, validators=[ validators.NumberRange(min=0), ], tooltip=_("How many epochs of training between taking a snapshot?") ) val_interval = utils.forms.FloatField( _('Validation interval (in epochs)'), default=1, validators=[ validators.NumberRange(min=0) ], tooltip=_("How many epochs of training between running through one pass of the validation data?") ) traces_interval = utils.forms.IntegerField( _('Tracing Interval (in steps)'), validators=[ validators.NumberRange(min=0) ], default=0, tooltip=_("Generation of a timeline trace every few steps") ) random_seed = utils.forms.IntegerField( _('Random seed'), validators=[ validators.NumberRange(min=0), validators.Optional(), ], tooltip=_('If you provide a random seed, then back-to-back runs with ' 'the same model and dataset should give identical results.') ) batch_size = utils.forms.MultiIntegerField( _('Batch size'), default=100, validators=[ utils.forms.MultiNumberRange(min=1), utils.forms.MultiOptional(), ], tooltip=_("How many images to process at once. If blank, values are used from the network definition.") ) batch_accumulation = utils.forms.IntegerField( _('Batch Accumulation'), validators=[ validators.NumberRange(min=1), validators.Optional(), ], tooltip=_("Accumulate gradients over multiple batches (useful when you " "need a bigger batch size for training but it doesn't fit in memory).") ) # Solver types solver_type = utils.forms.SelectField( _('Solver type'), choices=[ ('SGD', _('SGD (Stochastic Gradient Descent)')), ('MOMENTUM', _('Momentum')), ('NESTEROV', _("NAG (Nesterov's accelerated gradient)")), ('ADAGRAD', _('AdaGrad (Adaptive Gradient)')), ('ADAGRADDA', _('AdaGradDA (AdaGrad Dual Averaging)')), ('ADADELTA', _('AdaDelta')), ('ADAM', _('Adam (Adaptive Moment Estimation)')), ('RMSPROP', _('RMSprop')), ('FTRL', _('FTRL (Follow-The-Regularized-Leader)')), ], default='SGD', tooltip=_("What type of solver will be used?"), ) def validate_solver_type(form, field): fw = frameworks.get_framework_by_id(form.framework) if fw is not None: if not fw.supports_solver_type(field.data): raise validators.ValidationError( _('Solver type not supported by this framework')) # Additional settings specific to selected solver rms_decay = utils.forms.FloatField( _('RMS decay value'), default=0.99, validators=[ validators.NumberRange(min=0), ], tooltip=_("If the gradient updates results in oscillations the gradient is reduced " "by times 1-rms_decay. Otherwise it will be increased by rms_decay.") ) # Learning rate learning_rate = utils.forms.MultiFloatField( _('Base Learning Rate'), default=0.01, validators=[ utils.forms.MultiNumberRange(min=0), ], tooltip=_("Affects how quickly the network learns. If you are getting " "NaN for your loss, you probably need to lower this value.") ) lr_policy = wtforms.SelectField( _('Policy'), choices=[ ('fixed', _('Fixed')), ('step', _('Step Down')), ('multistep', _('Step Down (arbitrary steps)')), ('exp', _('Exponential Decay')), ('inv', _('Inverse Decay')), ('poly', _('Polynomial Decay')), ('sigmoid', _('Sigmoid Decay')), ], default='step' ) lr_step_size = wtforms.FloatField(_('Step Size'), default=33) lr_step_gamma = wtforms.FloatField(_('Gamma'), default=0.1) lr_multistep_values = wtforms.StringField(_('Step Values'), default="50,85") def validate_lr_multistep_values(form, field): if form.lr_policy.data == 'multistep': for value in field.data.split(','): try: float(value) except ValueError: raise validators.ValidationError(_('invalid value')) lr_multistep_gamma = wtforms.FloatField(_('Gamma'), default=0.5) lr_exp_gamma = wtforms.FloatField(_('Gamma'), default=0.95) lr_inv_gamma = wtforms.FloatField(_('Gamma'), default=0.1) lr_inv_power = wtforms.FloatField(_('Power'), default=0.5) lr_poly_power = wtforms.FloatField(_('Power'), default=3) lr_sigmoid_step = wtforms.FloatField(_('Step'), default=50) lr_sigmoid_gamma = wtforms.FloatField(_('Gamma'), default=0.1) # Network # Use a SelectField instead of a HiddenField so that the default value # is used when nothing is provided (through the REST API) method = wtforms.SelectField( _('Network type'), choices=[ ('standard', _('Standard network')), ('previous', _('Previous network')), ('pretrained', _('Pretrained network')), ('custom', _('Custom network')), ], default='standard', ) # framework - hidden field, set by Javascript to the selected framework ID framework = wtforms.HiddenField( _('framework'), validators=[ validators.AnyOf( [fw.get_id() for fw in frameworks.get_frameworks()], message=_('The framework you choose is not currently supported.') ) ], default=frameworks.get_frameworks()[0].get_id() ) # The options for this get set in the view (since they are dependent on the data type) standard_networks = wtforms.RadioField( _('Standard Networks'), validators=[ validate_required_iff(method='standard'), ], ) previous_networks = wtforms.RadioField( _('Previous Networks'), choices=[], validators=[ validate_required_iff(method='previous'), selection_exists_in_choices, ], ) pretrained_networks = wtforms.RadioField( _('Pretrained Networks'), choices=[], validators=[ validate_required_iff(method='pretrained'), selection_exists_in_choices, ], ) custom_network = utils.forms.TextAreaField( _('Custom Network'), validators=[ validate_required_iff(method='custom'), validate_NetParameter, ], ) custom_network_snapshot = utils.forms.TextField( _('Pretrained model(s)'), tooltip=_("Paths to pretrained model files, separated by '%(pathsep)s'. " "Only edit this field if you understand how fine-tuning " "works in caffe or torch.", pathsep=os.path.pathsep) ) def validate_custom_network_snapshot(form, field): pass # if form.method.data == 'custom': # for filename in field.data.strip().split(os.path.pathsep): # if filename and not os.path.lexists(filename): # raise validators.ValidationError('File "%s" does not exist' % filename) # Select one of several GPUs select_gpu = wtforms.RadioField( _('Select which GPU you would like to use'), choices=[('next', 'Next available')] + [( index, '#%s - %s (%s memory)' % ( index, get_device(index).name, sizeof_fmt( get_nvml_info(index)['memory']['total'] if get_nvml_info(index) and 'memory' in get_nvml_info(index) else get_device(index).totalGlobalMem) ), ) for index in config_value('gpu_list').split(',') if index], default='next', ) # Select N of several GPUs select_gpus = utils.forms.SelectMultipleField( _('Select which GPU[s] you would like to use'), choices=[( index, '#%s - %s (%s memory)' % ( index, get_device(index).name, sizeof_fmt( get_nvml_info(index)['memory']['total'] if get_nvml_info(index) and 'memory' in get_nvml_info(index) else get_device(index).totalGlobalMem) ), ) for index in config_value('gpu_list').split(',') if index], tooltip=_("The job won't start until all of the chosen GPUs are available.") ) # XXX For testing # The Flask test framework can't handle SelectMultipleFields correctly select_gpus_list = wtforms.StringField(_('Select which GPU[s] you would like to use (comma separated)')) def validate_select_gpus(form, field): if form.select_gpus_list.data: field.data = form.select_gpus_list.data.split(',') # Use next available N GPUs select_gpu_count = wtforms.IntegerField(_('Use this many GPUs (next available)'), validators=[ validators.NumberRange(min=1, max=len( config_value('gpu_list').split(','))) ], default=1, ) def validate_select_gpu_count(form, field): if field.data is None: if form.select_gpus.data: # Make this field optional field.errors[:] = [] raise validators.StopValidation() model_name = utils.forms.StringField(_('Model Name'), validators=[ validators.DataRequired() ], tooltip=_("An identifier, later used to refer to this model in the Application.") ) group_name = utils.forms.StringField(_('Group Name'), tooltip=_("An optional group name for organization on the main page.") ) # allows shuffling data during training (for frameworks that support this, as indicated by # their Framework.can_shuffle_data() method) shuffle = utils.forms.BooleanField(_('Shuffle Train Data'), default=True, tooltip=_('For every epoch, shuffle the data before training.') ) steps = utils.forms.IntegerField("训练总步长", default=4000, validators=[ validators.NumberRange(min=1) ], tooltip="本次训练总步长数(迭代次数)") iter_store_step = utils.forms.IntegerField("步长间隔", default=1000, validators=[ validators.NumberRange(min=1) ], tooltip="要间隔多少个步长来进行快照保存") train_batch_size = utils.forms.IntegerField("批处理大小", default=100, validators=[ validators.NumberRange(min=1) ], tooltip="一次处理多少图片,默认为100") # bottleneck_dir = utils.forms.StringField("瓶颈值目录", # tooltip="计算出每个图片的瓶颈值并存储于此目录下")
15,563
4,376
# encoding: utf-8 # usando python 3 # Afinação - Alberto toca violão e é programador. # Precisando afinar o violão e sem diapasão por perto, # resolveu fazer um programa para ajudá-lo. # O que ele queria era a nota Lá soando sem parar até que ele conseguisse afinar a # respectiva corda do violão; as demais cordas ele poderia afinar com base na primeira. # Escreva um programa que faz soar no alto-falante do computador a nota Lá (440 Hz) # e só para quando for pressionada alguma tecla. import numpy as np import simpleaudio as sa frequency = 440 # Our played note will be 440 Hz fs = 44100 # 44100 samples per second seconds = 3 # Note duration of 3 seconds # Generate array with seconds*sample_rate steps, ranging between 0 and seconds t = np.linspace(0, seconds, seconds * fs, False) # Generate a 440 Hz sine wave note = np.sin(frequency * t * 2 * np.pi) # Ensure that highest value is in 16-bit range audio = note * (2**15 - 1) / np.max(np.abs(note)) # Convert to 16-bit data audio = audio.astype(np.int16) # Start playback play_obj = sa.play_buffer(audio, 1, 2, fs) # Wait for playback to finish before exiting play_obj.wait_done()
1,155
422
fav_beer = {'Julian': 'White Rabbit Dark Ale', 'Bob': 'Some sort of light beer I assume', 'Mike': 'Oregano Beer'}
138
46
# Copyright 2014 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Actual WSGI app instantiations used from app.yaml. Extracted to a separate module to avoid calling 'initialize' in unit tests during module loading time. """ import gae_ts_mon import main endpoints_app, frontend_app, backend_app = main.initialize() gae_ts_mon.initialize() gae_ts_mon.instrument_wsgi_application(frontend_app) gae_ts_mon.instrument_wsgi_application(backend_app)
548
182
"""Layer classes.""" from __future__ import absolute_import import abc import six import tensorflow as tf @six.add_metaclass(abc.ABCMeta) class BaseLayer(object): """Base layer interface.""" @abc.abstractmethod def forward(self): """Layer forward propagation.""" pass @abc.abstractmethod def backward(self): """Layer backward propagation.""" pass @abc.abstractmethod def get_variables(self): """Get layer's tf variables.""" pass @abc.abstractmethod def get_parameters(self): """Get the layer parameters.""" pass class Linear(BaseLayer): """Fully-Connected layer.""" def __init__(self, shape, name="linear", vnames=["W", "b"]): """Create a new linear layer instance.""" self.name = name self.vnames = vnames with tf.name_scope(self.name): self.W = tf.Variable( tf.truncated_normal(shape=shape, stddev=0.1), name=vnames[0]) self.b = tf.Variable( tf.constant(0.1, shape=[shape[1]]), name=vnames[1]) def forward(self, X): """Forward propagate X through the fc layer.""" with tf.name_scope(self.name): return tf.add(tf.matmul(X, self.W), self.b) def backward(self, H): """Backward propagate H through the fc layer.""" pass def get_variables(self): """Get layer's variables.""" return [self.W, self.b] def get_parameters(self): """Return all the parameters of this layer.""" with tf.Session() as sess: return { self.names[0]: sess.run(self.W), self.names[1]: sess.run(self.b) } class Activation(BaseLayer): """Activation function layer.""" def __init__(self, func, name="act_func"): """Create a new Activation layer instance.""" self.name = name if func is not None: self.func = func else: self.func = tf.identity def forward(self, X): """Forward propagate X.""" return self.func(X) def backward(self, H): """Backward propagate H through the fc layer.""" pass def get_variables(self): """Return the layer's variables.""" pass def get_parameters(self): """Return all the parameters of this layer.""" pass class SoftMax(BaseLayer): """SoftMax layer.""" def __init__(self, prev_layer, n_classes, name="softmax"): """Create a new SoftMax layer instance.""" self.prev_layer = prev_layer self.shape = (prev_layer.get_shape()[1].value, n_classes) self.n_classes = n_classes self.name = name self.vs = ['softmax_W', 'softmax_b'] with tf.name_scope(self.name): self.W = tf.Variable( tf.truncated_normal(self.shape, stddev=0.1), name=self.vs[0]) self.b = tf.Variable( tf.constant(0.1, shape=[n_classes]), name=self.vs[0]) def forward(self, X): """Forward propagate X.""" with tf.name_scope(self.name): return tf.add(tf.matmul(self.prev_layer, self.W), self.b) def backward(self, H): """Backward propagate H through the fc layer.""" pass def get_variables(self): """Return the layer's variables.""" return (self.W, self.b) def get_parameters(self): """Return all the parameters of this layer.""" with tf.Session() as sess: return { self.vs[0]: sess.run(self.W), self.vs[1]: sess.run(self.b) } class Regularization(BaseLayer): """Regularization function layer.""" def __init__(self, variables, C, regtype="l2", name="act_func"): """Create a new Regularization layer instance.""" assert regtype in ["l1", "l2"] self.variables = variables self.C = C self.regtype = regtype self.name = name def forward(self, X): """Forward propagate X.""" regs = tf.constant(0.0) for v in self.variables: if self.regtype == "l1": regs = tf.add(regs, tf.reduce_sum(tf.abs(v))) elif self.regtype == "l2": regs = tf.add(regs, tf.nn.l2_loss(v)) return tf.mul(self.C, regs) def backward(self, H): """Backward propagate H through the fc layer.""" pass def get_variables(self): """Return the layer's variables.""" pass def get_parameters(self): """Return all the parameters of this layer.""" pass class Loss(BaseLayer): """Loss function layer.""" def __init__(self, mod_y, ref_y, loss_type, regterm=None, summary=True, name="loss_func"): """Create a new Loss layer instance.""" assert loss_type in ["cross_entropy", "softmax_cross_entropy", "mean_squared"] self.mod_y = mod_y self.ref_y = ref_y self.loss_type = loss_type self.regterm = regterm self.name = name if loss_type == "cross_entropy": clip_inf = tf.clip_by_value(self.mod_y, 1e-10, float('inf')) clip_sup = tf.clip_by_value(1 - self.mod_y, 1e-10, float('inf')) loss = - tf.reduce_mean(tf.add( tf.mul(self.ref_y, tf.log(clip_inf)), tf.mul(tf.sub(1.0, self.ref_y), tf.log(clip_sup)))) elif loss_type == "softmax_cross_entropy": loss = tf.contrib.losses.softmax_cross_entropy( self.mod_y, self.ref_y) elif loss_type == "mean_squared": loss = tf.sqrt(tf.reduce_mean( tf.square(tf.sub(self.ref_y, self.mod_y)))) self.loss = loss + regterm if regterm is not None else loss if summary: tf.summary.scalar(self.name, self.loss) def forward(self, X): """Forward propagate X.""" pass def backward(self, H): """Backward propagate H through the fc layer.""" pass def get_variables(self): """Return the layer's variables.""" pass def get_parameters(self): """Return all the parameters of this layer.""" pass
6,298
1,971
import unittest from malcolm.core import call_with_params from malcolm.modules.builtin.parts import GroupPart class TestGroupPart(unittest.TestCase): def setUp(self): self.o = call_with_params( GroupPart, name="things", description="A group of things") self.setter = list(self.o.create_attribute_models())[0][2] def test_init(self): assert self.o.name == "things" assert self.o.attr.value == "expanded" assert self.o.attr.meta.description == "A group of things" assert self.o.attr.meta.tags == ("widget:group", "config") def test_setter(self): assert self.o.attr.value == "expanded" self.setter("collapsed") assert self.o.attr.value == "collapsed" with self.assertRaises(ValueError): self.setter("anything else")
835
263
import os from jose import jwt from datetime import datetime, timedelta JWT_SECRET = 'secret' JWT_ALGORITHM = 'HS256' JWT_EXP_DELTA_SECONDS = 31556952 # year def get_token(request): return jwt.decode(request.headers.get('Authorization'), os.environ['JWT_SECRET']) def create_token(user_id): payload = { 'user_id': user_id, 'exp': datetime.utcnow() + timedelta(seconds=JWT_EXP_DELTA_SECONDS) } jwt_token = jwt.encode(payload, JWT_SECRET, JWT_ALGORITHM) return jwt_token
512
213
import os from django.db import models from ai4all_api.detection_items import DETECTION_ITEMS from ai4all_api.notification_types import NOTIFICATION_TYPES class SubmitCameraItem(models.Model): date_created = models.DateTimeField(auto_now_add=True) tg_chat_id = models.BigIntegerField() name = models.CharField(max_length=255, blank=True, default='') url = models.CharField(max_length=2048, blank=False) what_to_detect = models.CharField(choices=DETECTION_ITEMS, max_length=100) detection_threshold = models.IntegerField() detection_enabled = models.BooleanField(default=True) notification_type = models.CharField(choices = NOTIFICATION_TYPES, max_length=100) # used for 'appearance/disappearance' notification type last_time_object_presented = models.BooleanField(default=False) # Region of image where the detection to be run. # Specified in percents from the image size. edge_left = models.IntegerField(default=0) edge_top = models.IntegerField(default=0) edge_right = models.IntegerField(default=100) edge_bottom = models.IntegerField(default=100) def __str__(self): return "{}".format(self.name) class Meta: ordering = ('date_created',)
1,238
397
# # ------------------------------------------------------------------------- # Copyright (c) 2019 AT&T Intellectual Property # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # ------------------------------------------------------------------------- # class CPUFilter(object): def __init__(self): self.name = "cpu" self.status = None def init_condition(self): self.status = None def check_pre_condition(self, _level, _v, _avail_hosts, _avail_groups): return True def filter_candidates(self, _level, _v, _candidate_list): candidate_list = [] for c in _candidate_list: if self._check_candidate(_level, _v, c): candidate_list.append(c) return candidate_list def _check_candidate(self, _level, _v, _candidate): """Return True if host has sufficient CPU cores.""" avail_vcpus = _candidate.get_vcpus(_level) instance_vcpus = _v.vCPUs # TODO: need to check against original CPUs? # Do not allow an instance to overcommit against itself, # only against other instances. # if instance_vcpus > vCPUs: # return False if avail_vcpus < instance_vcpus: return False return True
1,797
525
import tensorflow as tf x1_data=[73.,93.,89.,96.,73.] x2_data=[80.,88.,91.,98.,66.] x3_data=[75.,93.,90.,100.,70.] y_data=[152.,185.,180.,196.,142.] x1=tf.placeholder(tf.float32) x2=tf.placeholder(tf.float32) x3=tf.placeholder(tf.float32) Y=tf.placeholder(tf.float32) w1=tf.Variable(tf.random_normal([1]),name='weight1') w2=tf.Variable(tf.random_normal([1]),name='weight2') w3=tf.Variable(tf.random_normal([1]),name='weight1') b=tf.Variable(tf.random_normal([1]),name='bias') hypothesis=x1*w1+x2*w2+x3*w3+b cost=tf.reduce_mean(tf.square(hypothesis-Y)) #minimize optimizer=tf.train.GradientDescentOptimizer(learning_rate=1e-5) train=optimizer.minimize(cost) #launch graph sess=tf.Session() #initialize sess.run(tf.global_variables_initializer()) for step in range(2001): cost_val,hy_val,_=sess.run([cost,hypothesis,train], feed_dict={x1:x1_data,x2:x2_data,x3:x3_data,Y:y_data}) if step%10==0: print(step,"Cost:",cost_val,"\nPrediction:\n",hy_val)
996
461
import pandas as pd import nltk import re from nltk.stem import wordnet from nltk import pos_tag from nltk import word_tokenize from datetime import datetime data = pd.read_csv('traindata.csv', encoding='utf-8') train_counter = 0 def text_normalize(text): global train_counter if train_counter % 10000 == 0: print(str(train_counter) + " sets lemmatized..., "+"Time now: " + str(datetime.now())) train_counter += 1 text = str(text).lower() spl_char_text = re.sub(r'[^ a-z]', '', text) tokens = nltk.word_tokenize(spl_char_text) lema = wordnet.WordNetLemmatizer() tags_list = pos_tag(tokens, tagset = None) lema_words = [] for token, pos_token in tags_list: if pos_token.startswith('V'): pos_value = 'v' elif pos_token.startswith('J'): pos_value = 'a' elif pos_token.startswith('R'): pos_value = 'r' else: pos_value = 'n' lema_token = lema.lemmatize(token, pos_value) lema_words.append(lema_token) return " ".join(lema_words) if __name__ == '__main__': print("Time now: " + str(datetime.now())) print(data.info()) print("\nData Imported...") print("----------------------------------------------------------------------------------------------------------") data['lemmatized text'] = data['Content'].apply(text_normalize) print("Training Data Lemmatized..., Time now: " + str(datetime.now())) data.to_csv('traindata.csv', encoding='utf-8', index = False) print(data['lemmatized text']) print(type(data['lemmatized text'])) print("\nTraining data...") print("----------------------------------------------------------------------------------------------------------")
1,759
546
import io import logging import queue from collections import namedtuple import torch import zmq from tensor_twister.status_codes import StatusCode UnpackedMessage = namedtuple("UnpackedMessage", ["tensor", "name", "ip"]) def serve(host: str, port: int): """ Listen for incoming tensor data from clients. Print comparisons between pairs of tensor data. Args: host (str): The hostname to listen on; for example "localhost" port (int): The port to listen on; for example 5555 """ logger = logging.getLogger(__name__) logger.debug("libzmq version: %s", zmq.zmq_version()) logger.debug(" pyzmq version: %s", zmq.__version__) tensor_queue = queue.Queue() context = zmq.Context() socket = context.socket(zmq.REP) server_uri = f"tcp://{host}:{port}" logger.info("Attempting to listen on %s.", server_uri) socket.bind(server_uri) logger.info("Listening on %s.", server_uri) while True: # Get the next message, blocking. message = socket.recv_pyobj() try: tensor = torch.load(message.tensor) except Exception: socket.send_pyobj(StatusCode.TensorLoadFailure) continue tensor_queue.put(UnpackedMessage(tensor, message.name, message.ip)) socket.send_pyobj(StatusCode.OK) # If the queue has at least 2 messages, compare the first 2. if tensor_queue.qsize() >= 2: m1 = tensor_queue.get() m2 = tensor_queue.get() print(f"{m1.name}@{m1.ip}: tensor min: {m1.tensor.min()}; max: {m1.tensor.max()}; mean: {m1.tensor.mean()}") print(f"{m2.name}@{m2.ip}: tensor min: {m2.tensor.min()}; max: {m2.tensor.max()}; mean: {m2.tensor.mean()}") print(f"t1 and t2 are {'' if (m1.tensor == m2.tensor).all() else 'not'} equal")
1,848
601
#! /usr/bin/env python # Copyright (c) 2014 KU Leuven, ESAT-STADIUS # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # # 1. Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # 3. Neither name of copyright holders nor the names of its contributors # may be used to endorse or promote products derived from this software # without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR # CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, # EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, # PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR # PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING # NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. import math import functools from .solver_registry import register_solver from .util import Solver, _copydoc from . import util _numpy_available = True try: import numpy as np except ImportError: _numpy_available = False _deap_available = True try: import deap import deap.creator import deap.base import deap.tools import deap.cma import deap.algorithms except ImportError: _deap_available = False except TypeError: # this can happen because DEAP is in Python 2 # install needs to take proper care of converting # 2 to 3 when necessary _deap_available = False class CMA_ES(Solver): """ .. include:: /global.rst Please refer to |cmaes| for details about this algorithm. This solver uses an implementation available in the DEAP library [DEAP2012]_. .. warning:: This solver has dependencies on DEAP_ and NumPy_ and will be unavailable if these are not met. .. _DEAP: https://code.google.com/p/deap/ .. _NumPy: http://www.numpy.org """ def __init__(self, num_generations, sigma=1.0, Lambda=None, **kwargs): """blah .. warning:: |warning-unconstrained| """ if not _deap_available: raise ImportError('This solver requires DEAP but it is missing.') if not _numpy_available: raise ImportError('This solver requires NumPy but it is missing.') self._num_generations = num_generations self._start = kwargs self._sigma = sigma self._lambda = Lambda @staticmethod def suggest_from_seed(num_evals, **kwargs): """Verify that we can effectively make a solver. The doctest has to be skipped from automated builds, because DEAP may not be available and yet we want documentation to be generated. >>> s = CMA_ES.suggest_from_seed(30, x=1.0, y=-1.0, z=2.0) >>> solver = CMA_ES(**s) #doctest:+SKIP """ fertility = 4 + 3 * math.log(len(kwargs)) d = dict(kwargs) d['num_generations'] = int(math.ceil(float(num_evals) / fertility)) # num_gen is overestimated # this will require slightly more function evaluations than permitted by num_evals return d @property def num_generations(self): return self._num_generations @property def start(self): """Returns the starting point for CMA-ES.""" return self._start @property def lambda_(self): return self._lambda @property def sigma(self): return self._sigma @_copydoc(Solver.optimize) def optimize(self, f, maximize=True, pmap=map): toolbox = deap.base.Toolbox() if maximize: fit = 1.0 else: fit = -1.0 deap.creator.create("FitnessMax", deap.base.Fitness, weights=(fit,)) Fit = deap.creator.FitnessMax deap.creator.create("Individual", list, fitness=Fit) Individual = deap.creator.Individual if self.lambda_: strategy = deap.cma.Strategy(centroid=list(self.start.values()), sigma=self.sigma, lambda_=self.lambda_) else: strategy = deap.cma.Strategy(centroid=list(self.start.values()), sigma=self.sigma) toolbox.register("generate", strategy.generate, Individual) toolbox.register("update", strategy.update) @functools.wraps(f) def evaluate(individual): return (util.score(f(**dict([(k, v) for k, v in zip(self.start.keys(), individual)]))),) toolbox.register("evaluate", evaluate) toolbox.register("map", pmap) hof = deap.tools.HallOfFame(1) deap.algorithms.eaGenerateUpdate(toolbox=toolbox, ngen=self._num_generations, halloffame=hof, verbose=False) return dict([(k, v) for k, v in zip(self.start.keys(), hof[0])]), None # CMA_ES solver requires deap > 1.0.1 # http://deap.readthedocs.org/en/latest/examples/cmaes.html if _deap_available and _numpy_available: CMA_ES = register_solver('cma-es', 'covariance matrix adaptation evolutionary strategy', ['CMA-ES: covariance matrix adaptation evolutionary strategy', ' ', 'This method requires the following parameters:', '- num_generations :: number of generations to use', '- sigma :: (optional) initial covariance, default 1', '- Lambda :: (optional) measure of reproducibility', '- starting point: through kwargs' ' ', 'This method is described in detail in:', 'Hansen and Ostermeier, 2001. Completely Derandomized Self-Adaptation in Evolution Strategies. Evolutionary Computation' ])(CMA_ES)
6,744
2,049
""" This problem was asked by Twitter. Implement an autocomplete system. That is, given a query string s and a set of all possible query strings, return all strings in the set that have s as a prefix. For example, given the query string de and the set of strings [dog, deer, deal], return [deer, deal]. Hint: Try preprocessing the dictionary into a more efficient data structure to speed up queries. """ def autocomplete_bruteforce(words, s): result = [] for word in words: if s in word: result.append(word) return result print(autocomplete_bruteforce(['dog','deer','deal'], 'de'))
622
173
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Created on 09.08.2018 This is basically a python port of of a script by berwinter https://github.com/berwinter/uvr1611/blob/master/lib/backend/blnet-connection.inc.php author: Niels """ import struct from datetime import datetime # Parser constant # 1 bit DIGITAL_ON = 1 DIGITAL_OFF = 0 # 8 bit SPEED_ACTIVE = 0x80 SPEED_MASK = 0x1F # 16 bit INT16_POSITIVE_MASK = 0xFFFF SIGN_BIT = 0x8000 POSITIVE_VALUE_MASK = 0x0FFF TYPE_MASK = 0x7000 TYPE_NONE = 0x0000 TYPE_DIGITAL = 0x1000 TYPE_TEMP = 0x2000 TYPE_VOLUME = 0x3000 TYPE_RADIATION = 0x4000 TYPE_RAS = 0x7000 RAS_POSITIVE_MASK = 0x01FF # 32 bit INT32_MASK = 0xFFFFFFFF INT32_SIGN = 0x80000000 class BLNETParser: def __init__(self, data): """ parse a binary string containing a dataset Provides access to the values of a dataset as object properties @param data: byte string """ # check if dataset contains time information # (fetched from bootloader storage) if len(data) == 61: (_, seconds, minutes, hours, days, months, years) = struct.unpack( '<55sBBBBBB', data) self.date = datetime(2000 + years, months, days, hours, minutes, seconds) # Only parse preceding data data = data[:55] power = [0, 0] kWh = [0, 0] MWh = [0, 0] (_, digital, speed, active, power[0], kWh[0], MWh[0], power[1], kWh[1], MWh[1]) = struct.unpack('<32sH4sBLHHLHH', data) analog = struct.unpack( '<{}{}'.format('H' * 16, 'x' * (len(data) - 32)), data) self.analog = {} for channel in range(0, 16): self.analog[channel + 1] = round( self._convert_analog(analog[channel]), 3) self.digital = {} for channel in range(0, 16): self.digital[channel + 1] = self._convert_digital(digital, channel) self.speed = {} for channel in range(0, 4): self.speed[channel + 1] = round( self._convert_speed(speed[channel]), 3) self.energy = {} for channel in range(0, 2): self.energy[channel + 1] = round( self._convert_energy(MWh[channel], kWh[channel], active, channel), 3) self.power = {} for channel in range(0, 2): self.power[channel + 1] = round( self._convert_power(power[channel], active, channel), 3) def to_dict(self): """ Turn parsed data into parser object @return dict """ return self.__dict__ def _convert_analog(self, value): """ Convert int to correct float @param value: short unsigned int that was returned by blnet @return float with correct sensor value """ mask = value & TYPE_MASK if mask == TYPE_TEMP: return self._calculate_value(value, 0.1) elif mask == TYPE_VOLUME: return self._calculate_value(value, 4) elif mask == TYPE_DIGITAL: if value & SIGN_BIT: return 1 else: return 0 elif mask == TYPE_RAS: return self._calculate_value(value, 0.1, RAS_POSITIVE_MASK) elif mask in [TYPE_RADIATION, TYPE_NONE] or True: return self._calculate_value(value) def _convert_digital(self, value, position): """ Check if bit at given position is set (=1) """ if value & (0x1 << (position)): return DIGITAL_ON else: return DIGITAL_OFF def _convert_speed(self, value): """ Check if speed is activated and return its value """ if value & SPEED_ACTIVE: return None else: return value & SPEED_MASK def _convert_energy(self, mwh, kwh, active, position): """ Check if heat meter is activated on a given position @return its energy """ if active & position: kwh = self._calculate_value(kwh, 0.1, INT16_POSITIVE_MASK) return mwh * 1000 + kwh else: return None def _convert_power(self, value, active, position): """ checks if heat meter is activated at given position @return its power """ if active & position: return self._calculate_value(value, 1 / 2560, INT32_MASK, INT32_SIGN) else: return None def _calculate_value(self, value, multiplier=1, positive_mask=POSITIVE_VALUE_MASK, signbit=SIGN_BIT): result = value & positive_mask if value & signbit: result = -((result ^ positive_mask) + 1) return result * multiplier
4,974
1,627
from typing import Optional from typing import List from fastapi import APIRouter, Depends, Body from models import User, Content, Node, Group, ExternalContent from routers import get_current_user, admin_only from schemas import NodeAdd, NodeEdit, NodeFind # router = APIRouter() @router.post("/push_content") async def push_content(node_id: str = Body(..., embed=True), content_id: str = Body(..., embed=True), admin: User = Depends(admin_only)): assert admin is not None node = await Node.find_one_and_add_to_set( find={"id": node_id}, data={"contents": Content.ref(content_id)} ) return node.export() @router.post("/pull_content") async def pull_content(node_id: str = Body(..., embed=True), content_id: str = Body(..., embed=True), admin: User = Depends(admin_only)): assert admin is not None node = await Node.find_one_and_pull( find={"id": node_id}, data={"contents": Content.ref(content_id)} ) return node.export() @router.post("/push_external_content") async def push_external_content(node_id: str = Body(..., embed=True), external_content_id: str = Body(..., embed=True), admin: User = Depends(admin_only)): assert admin is not None node = await Node.find_one_and_add_to_set( find={"id": node_id}, data={"external_contents": ExternalContent.ref(external_content_id)} ) return node.export() @router.post("/pull_external_content") async def pull_external_content(node_id: str = Body(..., embed=True), external_content_id: str = Body(..., embed=True), admin: User = Depends(admin_only)): assert admin is not None node = await Node.find_one_and_pull( find={"id": node_id}, data={"external_contents": ExternalContent.ref(external_content_id)} ) return node.export() @router.get("/current") async def current_nodes(current_user: User = Depends(get_current_user)): groups = await Group.find({"members": current_user}) nodes_ids = [node.id for group in groups for node in group.nodes] return [node.export() for node in await Node.find({"id": {"$in": nodes_ids}})] @router.post("/browse", dependencies=[Depends(admin_only)]) async def browse_nodes(find: NodeFind) -> List[Node]: return await Node.find(find=find.dict(exclude_unset=True)) @router.post("/read", dependencies=[Depends(admin_only)]) async def read_node(find: NodeFind, with_contents: bool = Body(False), with_other_contents: bool = Body(False)): node = await Node.find_one(find=find.dict(exclude_unset=True)) node_export = node.dict() if with_contents: node_export["contents"] = await Content.find({"id": {"$in": [ content.id for content in node.contents ]}}) if with_other_contents: node_export["other_contents"] = await Content.find({"id": {"$nin": [ content.id for content in node.contents ]}}) return node_export @router.post("/edit", dependencies=[Depends(admin_only)]) async def edit_node(find: NodeFind, data: NodeEdit): print("find", find) print("data", data) return await Node.find_one_and_set(find=find.dict(exclude_unset=True), data=data.dict(exclude_unset=True)) @router.post("/add", dependencies=[Depends(admin_only)]) async def add_node(data: NodeAdd): return await Node.insert_one(data=data.dict(exclude_unset=True)) @router.post("/delete", dependencies=[Depends(admin_only)]) async def delete_node(find: NodeFind): return await Node.delete_one(find=find.dict(exclude_unset=True))
3,568
1,118
class vector: def __init__(self, *vals): self.x = list(vals) for val in vals: float(val) self.dimension = len(self.x) def __getitem__(self, key): return self.x[key] def __setitem__(self, key, value): self.x[key] = value return self def __add__(self, other): if type(other) == type(int): # This also needs to support floating point types for i in range(self.dimension): self[i] += other return self self._checkDimension(other) newx = [] for i in range(self.dimension): newx.append(self[i] + other[i]) return vector(*newx) def __eq__(self, other): if self.dimension != other.dimension: return False for i in range(self.dimension): if self[i] != other[i]: return False return True def __mul__(self, other): if type(other) == type(int): x = [] for i in range(self.dimension): x.append(self[i] * other) return vector(*x) self._checkDimension(other) value = 0 for i in range(self.dimension): value += self[i] * other[i] return value def __rmul__(self, other): return self * other def __matmul__(self, other): if self.dimension != other.dimension != 3: raise TypeError("Vector dimensions must be 3") v = vector(0, 0, 0) v[0] = (self[1] * other[2]) - (self[2] * other[1]) v[1] = (self[2] * other[0]) - (self[0] * other[2]) v[2] = (self[0] * other[1]) - (self[1] * other[0]) return v def __sub__(self, other): return self + ( - other) def __neg__(self): v = [] for i in range(self): v.append( - self[i]) return vector(*v) def __abs__(self): value = self.magnitude() return value**0.5 def _checkDimension(self, other): if self.dimension != other.dimension: raise TypeError("Vector dimensions must agree") def magnitude(self): # Returns the value of the sum of all values of the vector squared powerMagnitude = 0 for a in self.x: powerMagnitude += a*a return powerMagnitude
2,427
794
# https://programmers.co.kr/learn/courses/30/lessons/1845 def solution(nums): if len(set(nums)) <= len(nums)/2: return int(len(set(nums))) else: return int(len(nums)/2) print(solution([3,1,2,3])) print(solution([3,3,3,2,2,4])) print(solution([3,3,3,2,2,2]))
283
131
# coding: utf-8 # # Copyright 2022 :Barry-Thomas-Paul: Moss # # Licensed under the Apache License, Version 2.0 (the "License") # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http: // www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Interface Class # this is a auto generated file generated by Cheetah # Libre Office Version: 7.3 # Namespace: com.sun.star.xml.crypto.sax import typing from abc import abstractmethod from ....uno.x_interface import XInterface as XInterface_8f010a43 if typing.TYPE_CHECKING: from ...sax.x_document_handler import XDocumentHandler as XDocumentHandler_9b90e28 from ...wrapper.xxml_element_wrapper import XXMLElementWrapper as XXMLElementWrapper_66c0107c class XSAXEventKeeper(XInterface_8f010a43): """ Interface of SAX Event Keeper. This interface is used to manipulate element marks in a SAX event stream. There are two kinds of element mark, one is element collector, which is used to collect a particular element from the SAX event stream; the other is blocker, which is used to block the SAX event stream. See Also: `API XSAXEventKeeper <https://api.libreoffice.org/docs/idl/ref/interfacecom_1_1sun_1_1star_1_1xml_1_1crypto_1_1sax_1_1XSAXEventKeeper.html>`_ """ __ooo_ns__: str = 'com.sun.star.xml.crypto.sax' __ooo_full_ns__: str = 'com.sun.star.xml.crypto.sax.XSAXEventKeeper' __ooo_type_name__: str = 'interface' __pyunointerface__: str = 'com.sun.star.xml.crypto.sax.XSAXEventKeeper' @abstractmethod def addBlocker(self) -> int: """ Adds a new blocker on the next element in the SAX event stream. No SAX event starting from the next element will be forwarded until this blocker is removed. """ @abstractmethod def addElementCollector(self) -> int: """ Adds a new element collector on the next element in the SAX event stream. """ @abstractmethod def getCurrentBlockingNode(self) -> 'XXMLElementWrapper_66c0107c': """ Gets the element which current blocking happens. This element is the working element of the first blocker in tree order. """ @abstractmethod def getElement(self, id: int) -> 'XXMLElementWrapper_66c0107c': """ Gets the element of an element mark. """ @abstractmethod def isBlocking(self) -> bool: """ Checks whether the SAX event stream is blocking. """ @abstractmethod def printBufferNodeTree(self) -> str: """ Prints information about all buffered elements. """ @abstractmethod def removeBlocker(self, id: int) -> None: """ Removes a blocker. """ @abstractmethod def removeElementCollector(self, id: int) -> None: """ Removes an element collector. """ @abstractmethod def setElement(self, id: int, aElement: 'XXMLElementWrapper_66c0107c') -> None: """ Sets the element of an element mark. When an element is replaced outside of this interface, then uses this method can restore the link between an element mark and its working element. """ @abstractmethod def setNextHandler(self, nextHandler: 'XDocumentHandler_9b90e28') -> 'XDocumentHandler_9b90e28': """ Sets the next document handler in the SAX chain. This handler will receive SAX events forwarded by the SAXEventKeeper. """ __all__ = ['XSAXEventKeeper']
3,903
1,201
from itertools import count from tqdm import tqdm from neodroid.environments.droid_environment import VectorUnityEnvironment def run(self, environment: VectorUnityEnvironment, render: bool = True) -> None: state = environment.reset().observables F = count(1) F = tqdm(F, leave=False, disable=not render) for frame_i in F: F.set_description(f"Frame {frame_i}") action, *_ = self.sample(state, deterministic=True) state, signal, terminated, info = environment.react(action, render=render) if terminated.all(): state = environment.reset().observables def infer(self, env, render=True): for episode_i in count(1): print(f"Episode {episode_i}") state = env.reset() for frame_i in count(1): action, *_ = self.sample(state) state, signal, terminated, info = env.act(action) if render: env.render() if terminated: break
992
290
# -*- coding: utf-8 -*- import re import copy import os import string import xlrd import pickle from .get_tokens import * keywords_0 = ('auto', 'typedf', 'const', 'extern', 'register', 'static', 'volatile', 'continue', 'break', 'default', 'return', 'goto', 'else', 'case') keywords_1 = ('catch', 'sizeof', 'if', 'switch', 'while', 'for') keywords_2 = ('memcpy', 'wmemcpy', '_memccpy', 'memmove', 'wmemmove', 'memset', 'wmemset', 'memcmp', 'wmemcmp', 'memchr', 'wmemchr', 'strncpy', 'lstrcpyn', 'wcsncpy', 'strncat', 'bcopy', 'cin', 'strcpy', 'lstrcpy', 'wcscpy', '_tcscpy', '_mbscpy', 'CopyMemory', 'strcat', 'lstrcat', 'fgets', 'main', '_main', '_tmain', 'Winmain', 'AfxWinMain', 'getchar', 'getc', 'getch', 'getche', 'kbhit', 'stdin', 'm_lpCmdLine', 'getdlgtext', 'getpass', 'istream.get', 'istream.getline', 'istream.peek', 'istream.putback', 'streambuf.sbumpc', 'streambuf.sgetc', 'streambuf.sgetn', 'streambuf.snextc', 'streambuf.sputbackc', 'SendMessage', 'SendMessageCallback', 'SendNotifyMessage', 'PostMessage', 'PostThreadMessage', 'recv', 'recvfrom', 'Receive', 'ReceiveFrom', 'ReceiveFromEx', 'CEdit.GetLine', 'CHtmlEditCtrl.GetDHtmlDocument', 'CListBox.GetText', 'CListCtrl.GetItemText', 'CRichEditCtrl.GetLine', 'GetDlgItemText', 'CCheckListBox.GetCheck', 'DISP_FUNCTION', 'DISP_PROPERTY_EX', 'getenv', 'getenv_s', '_wgetenv', '_wgetenv_s', 'snprintf', 'vsnprintf', 'scanf', 'sscanf', 'catgets', 'gets', 'fscanf', 'vscanf', 'vfscanf', 'printf', 'vprintf', 'CString.Format', 'CString.FormatV', 'CString.FormatMessage', 'CStringT.Format', 'CStringT.FormatV', 'CStringT.FormatMessage', 'CStringT.FormatMessageV', 'vsprintf', 'asprintf', 'vasprintf', 'fprintf', 'sprintf', 'syslog', 'swscanf', 'sscanf_s', 'swscanf_s', 'swprintf', 'malloc', 'readlink', 'lstrlen', 'strchr', 'strcmp', 'strcoll', 'strcspn', 'strerror', 'strlen', 'strpbrk', 'strrchr', 'strspn', 'strstr', 'strtok', 'strxfrm', 'kfree', '_alloca') keywords_3 = ('_strncpy*', '_tcsncpy*', '_mbsnbcpy*', '_wcsncpy*', '_strncat*', '_mbsncat*', 'wcsncat*', 'CEdit.Get*', 'CRichEditCtrl.Get*', 'CComboBox.Get*', 'GetWindowText*', 'istream.read*', 'Socket.Receive*', 'DDX_*', '_snprintf*', '_snwprintf*') keywords_5 = ('*malloc',) xread = xlrd.open_workbook('./sysevr/ml_models/function.xls') keywords_4 = [] for sheet in xread.sheets(): col = sheet.col_values(0)[1:] keywords_4 += col #print keywords_4 typewords_0 = ('short', 'int', 'long', 'float', 'doubule', 'char', 'unsigned', 'signed', 'void' ,'wchar_t', 'size_t', 'bool') typewords_1 = ('struct', 'union', 'enum') typewords_2 = ('new', 'delete') operators = ('+', '-', '*', '/', '=', '%', '?', ':', '!=', '==', '<<', '&&', '||', '+=', '-=', '++', '--', '>>', '|=') function = '^[_a-zA-Z][_a-zA-Z0-9]*$' variable = '^[_a-zA-Z][_a-zA-Z0-9(->)?(\.)?]*$' number = '[0-9]+' stringConst = '(^\'[\s|\S]*\'$)|(^"[\s|\S]*"$)' constValue = ['NULL', 'false', 'true'] phla = '[^a-zA-Z0-9_]' space = '\s' spa = '' def isinKeyword_3(token): for key in keywords_3: if len(token) < len(key)-1: return False if key[:-1] == token[:len(key)-1]: return True else: return False def isinKeyword_5(token): for key in keywords_5: if len(token) < len(key)-1: return False if token.find(key[1:]) != -1: if "_" in token: return False else: return True else: return False def isphor(s, liter): m = re.search(liter, s) if m is not None: return True else: return False def var(s): m = re.match(function, s) if m is not None: return True else: return False def CreateVariable(string, token): length = len(string) stack1 = [] s = '' i = 0 while (i < length): if var(string[i]): #if i + 1 < length and (string[i + 1] == '->' or string[i + 1] == '.'): # stack1.append(string[i]) # stack1.append(string[i + 1]) # i = i + 2 #else: while stack1 != []: s = stack1.pop() + s s = s + string[i] token.append(s) s = '' i = i + 1 else: token.append(string[i]) i = i + 1 def mapping(list_sentence): list_code = [] list_func = [] for code in list_sentence: #print code _string = '' for c in code: _string = _string + ' ' + c _string = _string[1:] list_code.append(_string) #print list_code _func_dict = {} _variable_dict = {} index = 0 while index < len(list_code): string = [] token = [] j = 0 str1 = copy.copy(list_code[index]) i = 0 tag = 0 strtemp = '' while i < len(str1): if tag == 0: if isphor(str1[i], space): if i > 0: string.append(str1[j:i]) j = i + 1 else: j = i + 1 i = i + 1 elif i + 1 == len(str1): string.append(str1[j:i + 1]) break elif isphor(str1[i], phla): if i + 1 < len(str1) and str1[i] == '-' and str1[i + 1] == '>': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '<' and str1[i + 1] == '<': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '>' and str1[i + 1] == '>': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '&' and str1[i + 1] == '&': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '|' and str1[i + 1] == '|': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '|' and str1[i + 1] == '=': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '=' and str1[i + 1] == '=': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '!' and str1[i + 1] == '=': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '+' and str1[i + 1] == '+': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '-' and str1[i + 1] == '-': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '+' and str1[i + 1] == '=': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif i + 1 < len(str1) and str1[i] == '-' and str1[i + 1] == '=': string.append(str1[i] + str1[i + 1]) j = i + 2 i = i + 2 elif str1[i] == '"': strtemp = strtemp + str1[i] i = i + 1 tag = 1 elif str1[i] == '\'': strtemp = strtemp + str1[i] i = i + 1 tag = 2 else: string.append(str1[i]) j = i + 1 i += 1 else: i += 1 elif tag == 1: if str1[i] != '"': strtemp = strtemp + str1[i] i = i + 1 else: strtemp = strtemp + str1[i] string.append(strtemp) strtemp = '' tag = 0 j = i + 1 i += 1 elif tag == 2: if str1[i] != '\'': strtemp = strtemp + str1[i] i = i + 1 else: strtemp = strtemp + str1[i] string.append(strtemp) strtemp = '' tag = 0 j = i + 1 i += 1 count = 0 for sub in string: if sub == spa: count += 1 for i in range(count): string.remove('') CreateVariable(string, token) j = 0 while j < len(token): if token[j] in constValue: token[j] = token[j] j += 1 elif j < len(token) and isphor(token[j], variable): if (token[j] in keywords_0) or (token[j] in typewords_0) or (token[j] in typewords_1 or token[j] in typewords_2): j += 1 elif j - 1 >= 0 and j + 1 < len(token) and token[j-1] == 'new' and token[j + 1] == '[': j = j + 2 elif j + 1 < len(token) and token[j + 1] == '(': #print(token[j]) if token[j] in keywords_1: j = j + 2 elif token[j] in keywords_2: #print('3', token[j]) j = j + 2 elif isinKeyword_3(token[j]): #print('4', token[j]) j = j + 2 elif token[j] in keywords_4: #print('5', token[j]) j = j + 2 elif isinKeyword_5(token[j]): #print('6', token[j]) j = j + 2 else: #print('7',token[j]) if "good" in token[j] or "bad" in token[j]: list_func.append(str(token[j])) if token[j] in _func_dict.keys(): token[j] = _func_dict[token[j]] else: list_values = _func_dict.values() if len(list_values) == 0: _func_dict[token[j]] = 'func_0' token[j] = _func_dict[token[j]] else: if token[j] in _func_dict.keys(): token[j] = _func_dict[token[j]] else: list_num = [] for value in list_values: list_num.append(int(value.split('_')[-1])) _max = max(list_num) _func_dict[token[j]] = 'func_' + str(_max+1) token[j] = _func_dict[token[j]] j = j + 2 elif j + 1 < len(token) and (not isphor(token[j + 1], variable)): if token[j + 1] == '*': if j + 2 < len(token) and token[j + 2] == 'const': j = j + 3 elif j - 1 >= 0 and token[j - 1] == 'const': j = j + 2 elif j - 1 > 0 and (token[j - 1] in operators): list_values = _variable_dict.values() if len(list_values) == 0: _variable_dict[token[j]] = 'variable_0' token[j] = _variable_dict[token[j]] else: if token[j] in _variable_dict.keys(): token[j] = _variable_dict[token[j]] else: list_num = [] for value in list_values: list_num.append(int(value.split('_')[-1])) _max = max(list_num) _variable_dict[token[j]] = 'variable_' + str(_max+1) token[j] = _variable_dict[token[j]] j = j + 2 elif j + 2 < len(token) and token[j + 2] == ')': j = j + 2 elif j - 2 > 0 and (token[j - 1] == '(' and token[j - 2] in operators): list_values = _variable_dict.values() if len(list_values) == 0: _variable_dict[token[j]] = 'variable_0' token[j] = _variable_dict[token[j]] else: if token[j] in _variable_dict.keys(): token[j] = _variable_dict[token[j]] else: list_num = [] for value in list_values: list_num.append(int(value.split('_')[-1])) _max = max(list_num) _variable_dict[token[j]] = 'variable_' + str(_max+1) token[j] = _variable_dict[token[j]] j = j + 2 else: list_values = _variable_dict.values() if len(list_values) == 0: _variable_dict[token[j]] = 'variable_0' token[j] = _variable_dict[token[j]] else: if token[j] in _variable_dict.keys(): token[j] = _variable_dict[token[j]] else: list_num = [] for value in list_values: list_num.append(int(value.split('_')[-1])) _max = max(list_num) _variable_dict[token[j]] = 'variable_' + str(_max+1) token[j] = _variable_dict[token[j]] j = j + 2 else: list_values = _variable_dict.values() if len(list_values) == 0: _variable_dict[token[j]] = 'variable_0' token[j] = _variable_dict[token[j]] else: if token[j] in _variable_dict.keys(): token[j] = _variable_dict[token[j]] else: list_num = [] for value in list_values: list_num.append(int(value.split('_')[-1])) _max = max(list_num) _variable_dict[token[j]] = 'variable_' + str(_max+1) token[j] = _variable_dict[token[j]] j = j + 2 elif j + 1 == len(token): list_values = _variable_dict.values() if len(list_values) == 0: _variable_dict[token[j]] = 'variable_0' token[j] = _variable_dict[token[j]] else: if token[j] in _variable_dict.keys(): token[j] = _variable_dict[token[j]] else: list_num = [] for value in list_values: list_num.append(int(value.split('_')[-1])) _max = max(list_num) _variable_dict[token[j]] = 'variable_' + str(_max+1) token[j] = _variable_dict[token[j]] break else: j += 1 elif j < len(token) and isphor(token[j], number): j += 1 elif j < len(token) and isphor(token[j], stringConst): j += 1 else: j += 1 stemp = '' i = 0 while i < len(token): if i == len(token) - 1: stemp = stemp + token[i] else: stemp = stemp + token[i] + ' ' i += 1 list_code[index] = stemp index += 1 #print list_code #print _variable_dict return list_code, list_func
17,995
5,499
""" Copyright (c) 2020 Flatipie Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ from PyQt5.QtWidgets import QGroupBox, QGraphicsDropShadowEffect from PyQt5.QtCore import Qt, pyqtSignal, QObject, QPoint class MaterialCard(QGroupBox): clicked = pyqtSignal() def __init__(self, parent=None, shadow=True): super(MaterialCard, self).__init__(parent) self._is_shadow = shadow @property def is_shadow(self): return self._is_shadow def enterEvent(self, event): if self.is_shadow: shadow_effect = QGraphicsDropShadowEffect( blurRadius=10, offset=QPoint(0, 0) ) self.setGraphicsEffect(shadow_effect) def leaveEvent(self, event): if self.is_shadow: self.setGraphicsEffect(None) def mousePressEvent(self, event): self.clicked.emit()
1,929
631
import os import shutil import xml.dom.minidom def output_pascal_format(mid_data, output_path): # mid_data: # filename # width # height # depth # objects # -- xmin # -- ymin # -- xmax # -- ymax # -- name # -- pose # -- truncated # -- difficult additional_data = dict() additional_data['folder'] = 'VOC2007' additional_data['s_database'] = 'The VOC2007 Database' additional_data['s_annotation'] = 'PASCAL VOC2007' additional_data['s_image'] = 'flickr' additional_data['s_flickrid'] = '123456789' additional_data['o_flickrid'] = 'Tom' additional_data['o_name'] = 'Tom' additional_data['segmented'] = '0' des_xml_dom = xml.dom.minidom.Document() # annotation des_root_node = des_xml_dom.createElement('annotation') # folder des_folder_node = des_xml_dom.createElement('folder') des_folder = des_xml_dom.createTextNode(additional_data['folder']) des_folder_node.appendChild(des_folder) des_root_node.appendChild(des_folder_node) # filename des_filename_node = des_xml_dom.createElement('filename') des_filename = des_xml_dom.createTextNode(mid_data['filename']) des_filename_node.appendChild(des_filename) des_root_node.appendChild(des_filename_node) # source des_dataset_name = des_xml_dom.createTextNode(additional_data['s_database']) des_dataset_node = des_xml_dom.createElement('database') des_dataset_node.appendChild(des_dataset_name) des_annotation = des_xml_dom.createTextNode(additional_data['s_annotation']) des_annotation_node = des_xml_dom.createElement('annotation') des_annotation_node.appendChild(des_annotation) des_image = des_xml_dom.createTextNode(additional_data['s_image']) des_image_node = des_xml_dom.createElement('image') des_image_node.appendChild(des_image) des_flickrid = des_xml_dom.createTextNode(additional_data['s_flickrid']) des_flickrid_node = des_xml_dom.createElement('flickrid') des_flickrid_node.appendChild(des_flickrid) des_source_node = des_xml_dom.createElement('source') des_source_node.appendChild(des_dataset_node) des_source_node.appendChild(des_annotation_node) des_source_node.appendChild(des_image_node) des_source_node.appendChild(des_flickrid_node) des_root_node.appendChild(des_source_node) # owner des_owner_flickrid = des_xml_dom.createTextNode(additional_data['o_flickrid']) des_owner_flickrid_node = des_xml_dom.createElement('flickrid') des_owner_flickrid_node.appendChild(des_owner_flickrid) des_owner_name = des_xml_dom.createTextNode(additional_data['o_name']) des_owner_name_node = des_xml_dom.createElement('name') des_owner_name_node.appendChild(des_owner_name) des_owner_node = des_xml_dom.createElement('owner') des_owner_node.appendChild(des_owner_flickrid_node) des_owner_node.appendChild(des_owner_name_node) des_root_node.appendChild(des_owner_node) # size des_size_node = des_xml_dom.createElement('size') des_width_node = des_xml_dom.createElement('width') des_height_node = des_xml_dom.createElement('height') des_depth_node = des_xml_dom.createElement('depth') des_width = des_xml_dom.createTextNode(str(mid_data['width'])) des_height = des_xml_dom.createTextNode(str(mid_data['height'])) des_depth = des_xml_dom.createTextNode(str(mid_data['depth'])) des_width_node.appendChild(des_width) des_height_node.appendChild(des_height) des_depth_node.appendChild(des_depth) des_size_node.appendChild(des_width_node) des_size_node.appendChild(des_height_node) des_size_node.appendChild(des_depth_node) des_root_node.appendChild(des_size_node) # segmented des_segmented = des_xml_dom.createTextNode(additional_data['segmented']) des_segmented_node = des_xml_dom.createElement('segmented') des_segmented_node.appendChild(des_segmented) des_root_node.appendChild(des_segmented_node) # object org_objects = mid_data['objects'] for j in range(0, len(org_objects)): org_object = org_objects[j] des_object_node = des_xml_dom.createElement('object') x_min = int(org_object['xmin']) y_min = int(org_object['ymin']) x_max = int(org_object['xmax']) y_max = int(org_object['ymax']) # prevent box scale out # pixel coordinate is 1 based if x_min <= 0: org_object['xmin'] = '1' if y_min <= 0: org_object['ymin'] = '1' if y_max > mid_data['height']: org_object['ymax'] = mid_data['height'] if x_max > mid_data['width']: org_object['xmax'] = mid_data['width'] # name des_object_name = des_xml_dom.createTextNode(org_object['name']) des_object_name_node = des_xml_dom.createElement('name') des_object_name_node.appendChild(des_object_name) des_object_node.appendChild(des_object_name_node) # pose des_pose = des_xml_dom.createTextNode(org_object['pose']) des_pose_node = des_xml_dom.createElement('pose') des_pose_node.appendChild(des_pose) des_object_node.appendChild(des_pose_node) # truncated des_truncated = des_xml_dom.createTextNode(str(org_object['truncated'])) des_truncated_node = des_xml_dom.createElement('truncated') des_truncated_node.appendChild(des_truncated) des_object_node.appendChild(des_truncated_node) # difficult des_object_difficult = des_xml_dom.createTextNode(str(org_object['difficult'])) des_object_difficult_node = des_xml_dom.createElement('difficult') des_object_difficult_node.appendChild(des_object_difficult) des_object_node.appendChild(des_object_difficult_node) # bndbox des_xmin_node = des_xml_dom.createElement('xmin') des_xmin = des_xml_dom.createTextNode(str(org_object['xmin'])) des_xmin_node.appendChild(des_xmin) des_ymin_node = des_xml_dom.createElement('ymin') des_ymin = des_xml_dom.createTextNode(str(org_object['ymin'])) des_ymin_node.appendChild(des_ymin) des_xmax_node = des_xml_dom.createElement('xmax') des_xmax = des_xml_dom.createTextNode(str(org_object['xmax'])) des_xmax_node.appendChild(des_xmax) des_ymax_node = des_xml_dom.createElement('ymax') des_ymax = des_xml_dom.createTextNode(str(org_object['ymax'])) des_ymax_node.appendChild(des_ymax) des_object_box_node = des_xml_dom.createElement('bndbox') des_object_box_node.appendChild(des_xmin_node) des_object_box_node.appendChild(des_ymin_node) des_object_box_node.appendChild(des_xmax_node) des_object_box_node.appendChild(des_ymax_node) des_object_node.appendChild(des_object_box_node) des_root_node.appendChild(des_object_node) with open(output_path, 'w') as des_file: des_root_node.writexml(des_file, addindent='\t', newl='\n')
7,035
2,373
""" Contains path planning logic. """ import math import numpy as np from heapq import heappush, heappop def heuristics(a, b): """Heuristics function using the Euclidian Distance.""" weight = 1.0 x1, y1 = a x2, y2 = b distance = np.sqrt(np.square(x2-x1) + np.square(y2-y1)) # distance = math.hypot(x1 - x2, y1 - y2) return distance def motion_model_4(): return [ [1, 0, 1], [0, 1, 1], [-1, 0, 1], [0, -1, 1], [-1, -1, 1], [-1, 1, 1], [1, -1, 1], [1, 1, 1] ] class AStarPlanner: def __init__(self): pass def plan(self, occupancy_grid_map, start_node, goal_node): """Plans a path through the occupancy grid map. Args: occupancy_grid_map: The occupancy grid map. start_node: Coordinates of the start node. goal_ndoe: Coordinates of the goal node. Returns: A list of coordinates of the planned path or None, if no path could be constructed. """ # Node; Cost to Goal; Node cost, previous node start_node_costs = 0 node_to_goal = heuristics(start_node, goal_node) + start_node_costs frontier = [(node_to_goal, start_node_costs, start_node, None)] visited = [] history = {} possible_movements = motion_model_4() # Safety guard (TODO: Remove after DEV) i = 0 break_if_count_reached = 10000 while frontier or i >= break_if_count_reached: i += 1 element = heappop(frontier) total_cost, cost, position, previous = element # If we have already traversed this node (x,y), then skip it if position in visited: continue # Mark this position as visited visited.append(position) history[position] = previous # Have already reached our goal, we can abort. if position == goal_node: break for dx, dy, dcost in possible_movements: xn = position[0] + dx yn = position[1] + dy if xn < 0 or yn < 0: continue if (xn, yn) in visited: continue if yn >= occupancy_grid_map.shape[0] or xn >= occupancy_grid_map.shape[1]: continue # Check if that cell is free! cell = occupancy_grid_map[yn][xn] if cell <= 0: potential_cost = 0 # abs(cell) # * 3 new_cost = cost + dcost + potential_cost new_total_cost_to_goal = new_cost + \ heuristics((xn, yn), goal_node) + potential_cost heappush( frontier, (new_total_cost_to_goal, new_cost, (xn, yn), position)) path = [] while position: path.append(position) position = history[position] return list(reversed(path)) def smoothen(self, occupancy_grid_map, path): """Smoothens the planned path. Utilizes gradient descent to smoothen the path. """ from copy import deepcopy # Create a deep copy of the path smoothed_path = deepcopy(path) weight_data = 0.01 weight_smooth = 0.8 tolerance = 0.0000001 smoothed_path = [list(elem) for elem in smoothed_path] while True: # Keep track of the total of changes made to check if we # reached convergence total_of_changes = 0 for i in range(len(path)): # Do not smoothen start and endpoint if i == 0 or i == (len(path) - 1): continue for dimension in range(len(path[i])): previous = smoothed_path[i][dimension] smoothed_path[i][dimension] = smoothed_path[i][dimension] + \ weight_data * (path[i][dimension] - smoothed_path[i][dimension]) + \ weight_smooth * \ (smoothed_path[i+1][dimension] + smoothed_path[i-1] [dimension] - 2 * smoothed_path[i][dimension]) total_of_changes += abs(previous - smoothed_path[i][dimension]) if total_of_changes < tolerance: break return smoothed_path def get_nearest_point(robot, aa): r = (robot.x, robot.y) a = [edist(k, r) for k in list(reference_trajectory)] i = np.argmin(a) p1 = reference_trajectory[i] p2 = reference_trajectory[i+5] aaa = np.arctan2(p2[1] - p1[1], p2[0] - p1[0]) print(f"l => {aaa:.4f}, {p1}, {p2}") return reference_trajectory[i]
4,880
1,528
from libcity.model.traffic_od_prediction.GEML import GEML __all__ = [ "GEML" ]
84
37
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('core', '0007_auto_20151202_1434'), ] operations = [ migrations.AddField( model_name='certificationprocess', name='active', field=models.BooleanField(default=True, verbose_name='Active'), ), migrations.AlterField( model_name='certificationprocess', name='course_certification', field=models.ForeignKey(related_name='processes', verbose_name='Certificate', to='core.CourseCertification', null=True), ), ]
699
207
import json import logging import sys log = logging.getLogger(__name__) def die(s): log.error(s) sys.exit(-1) def load_json(path): try: fp = open(path) except OSError as err: die("Could not open {}: {}".format(path, str(err))) try: value = json.load(fp) except ValueError as err: die("Invalid JSON in {}: {}".format(path, str(err))) return value def load_jsons(*paths): return [load_json(path) for path in paths] def merge_dicts(a, b): c = a.copy() c.update(b) return c def usage(): log.error("usage: hipshare <strategy>") sys.exit(-1)
628
228
identifier = 'org.vistrails.test.upgrades_layout' name ='test_upgrades_layout' version = '0.3'
95
39
#for all read file code at the end of the code file should closed. f = open('test.txt', 'r') file_contents = f.read() f.close() words = file_contents.split(' ') word_count = len(words) print(word_count) """ OUTPUT: 3 """ #for all the time use a file closing it is became a big job. with open('test.txt', 'r') as f: #python manage the file life cycle file_contents = f.read() words = file_contents.split(' ') word_count = len(words) print(word_count) """ OUTPUT: 3 """
480
175
import torch from einops import rearrange """ @openai - VD-VAE https://github.com/openai/vdvae/blob/main/vae_helpers.py Nice helper module as calling super.__init__() gets annoying """ class HelperModule(torch.nn.Module): def __init__(self, *args, **kwargs): super().__init__() self.build(*args, **kwargs) def build(self, *args, **kwargs): raise NotImplementedError def get_parameter_count(net: torch.nn.Module) -> int: return sum(p.numel() for p in net.parameters() if p.requires_grad) """ @lucidrains - Phil Wang (nystrom-attention) https://github.com/lucidrains/nystrom-attention/blob/main/nystrom_attention/nystrom_attention.py """ def exists(val): return val is not None def moore_penrose_iter_pinv(x, iters = 6): device = x.device abs_x = torch.abs(x) col = abs_x.sum(dim = -1) row = abs_x.sum(dim = -2) z = rearrange(x, '... i j -> ... j i') / (torch.max(col) * torch.max(row)) I = torch.eye(x.shape[-1], device = device) I = rearrange(I, 'i j -> () i j') for _ in range(iters): xz = x @ z z = 0.25 * z @ (13 * I - (xz @ (15 * I - (xz @ (7 * I - xz))))) return z
1,192
477
# coding=utf-8 import sys import numpy as np from p2_mahjong.card import MahjongCard as Card log_head = "utils.py" CARD_USED_TYPE = ['characters', 'green', 'red', 'white', 'east', 'west', 'north', 'south', 'spring', 'summer', 'autumn', 'winter', 'mei', 'lan', 'zhu', 'ju'] card_encoding_dict = {} card_id = 0 DIC_CHOW = {} character_list = [] wind_list = [] dragon_list = [] card_used = {} for _type in ['bamboo', 'characters', 'dots']: for _trait in ['1', '2', '3', '4', '5', '6', '7', '8', '9']: card = _type+"-"+_trait card_encoding_dict[card] = card_id DIC_CHOW[card_id] = 1 if _type in ['characters']: card_used[card_id] = 1 character_list.append(card_id) card_id += 1 for _trait in ['green', 'red', 'white']: card = 'dragons-'+_trait card_encoding_dict[card] = card_id if _trait in CARD_USED_TYPE: card_used[card_id] = 1 dragon_list.append(card_id) card_id += 1 for _trait in ['east', 'west', 'north', 'south']: card = 'winds-'+_trait card_encoding_dict[card] = card_id if _trait in CARD_USED_TYPE: card_used[card_id] = 1 wind_list.append(card_id) card_id += 1 for _trait in ['spring', 'summer', 'autumn', 'winter', 'mei', 'lan', 'zhu', 'ju']: card = 'flowers-'+_trait card_encoding_dict[card] = card_id if _trait in CARD_USED_TYPE: card_used[card_id] = 1 card_id += 1 card_decoding_dict = {card_encoding_dict[key]: key for key in card_encoding_dict.keys()} def init_deck(game_id=""): func_head = "init_deck()" + game_id deck = [] idx = 0 for card_id in card_decoding_dict: for _ in range(4): if card_id not in card_used: continue card = Card(runtime_id=idx, card_id=card_id) card.type = card_decoding_dict[card_id].split("-")[0] card.trait = card_decoding_dict[card_id].split("-")[1] deck.append(card) idx += 1 if card.type == "flowers": break return deck
2,078
797
class Solution: def mySqrt(self, x: int) -> int: # Base cases if (x == 0 or x == 1): return x # Staring from 1, try all numbers until # i*i remains less than to x. i = 1 while (i*i < x):i += 1 return i if i*i == x else i-1 ''' class Solution: def mySqrt(self,x) : # Base cases if (x == 0 or x == 1) : return x # Do Binary Search for integer square root start = 1 end = x while (start <= end) : mid = (start + end) // 2 # If x is a perfect square if (mid*mid == x) : return mid # when mid^2 is smaller than x, check if (mid+1)^2 >x if (mid * mid < x) : if (mid+1)*(mid+1) > x:return mid start = mid + 1 else : # If mid*mid is greater than x end = mid-1 ''' sol=Solution() for i in range(1,10): print(i,sol.mySqrt(i))
1,062
385
"""MEG UI Manager """ import pkg_resources from PyQt5 import QtCore, QtWidgets, QtGui, uic from meg_runtime.config import Config from meg_runtime.logger import Logger from meg_runtime.app import App class UIManager(QtWidgets.QMainWindow): """Main UI manager for the MEG system.""" UI_FILE = 'mainwindow.ui' # The window class widgets __widgets = None def __init__(self, **kwargs): """UI manager constructor.""" # Load window resource if needed if UIManager.__widgets is None: # Load the resource setup from the package UIManager.__widgets = uic.loadUiType(pkg_resources.resource_filename(__name__, UIManager.UI_FILE)) # Initialize the super class super().__init__(**kwargs) # Setup window resource UIManager.__widgets[0]().setupUi(self) # Set the window panel stack self._panels = [] self._current_panel = None self._current_popup = None # Set handler for closing a panel self._panel = self.findChild(QtWidgets.QTabWidget, 'panelwidget') self._panel.tabCloseRequested.connect(self.remove_view_by_index) self._panel.currentChanged.connect(self._show_view_by_index) # Get status widget self._statusbar = self.findChild(QtWidgets.QStatusBar, 'statusbar') # Set handlers for main buttons # TODO: Add more handlers for these self._action_clone = self.findChild(QtWidgets.QAction, 'action_Clone') self._action_clone.triggered.connect(App.open_clone_panel) self._action_open = self.findChild(QtWidgets.QAction, 'action_Open') self._action_open.triggered.connect(App.open_repo_panel) self._action_quit = self.findChild(QtWidgets.QAction, 'action_Quit') self._action_quit.triggered.connect(App.quit) self._action_about = self.findChild(QtWidgets.QAction, 'action_About') self._action_about.triggered.connect(App.open_about) self._action_preferences = self.findChild(QtWidgets.QAction, 'action_Preferences') self._action_preferences.triggered.connect(App.open_prefs_panel) self._action_manage_plugins = self.findChild(QtWidgets.QAction, 'action_Manage_Plugins') self._action_manage_plugins.triggered.connect(App.open_plugins_panel) # Set the default title self.set_title() # Set the icon icon_path = App.get_icon() if icon_path is not None: self.setWindowIcon(QtGui.QIcon(icon_path)) # Restore the state from the configuration if needed window_state = Config.get('window/state', 'none') state = self.windowState() if window_state == 'maximized': state &= ~(QtCore.Qt.WindowMinimized | QtCore.Qt.WindowFullScreen) state |= QtCore.Qt.WindowMaximized elif window_state == 'minimized': state &= ~(QtCore.Qt.WindowMaximized | QtCore.Qt.WindowFullScreen) state |= QtCore.Qt.WindowMinimized elif window_state == 'fullscreen': state &= ~(QtCore.Qt.WindowMinimized | QtCore.Qt.WindowMaximized) state |= QtCore.Qt.WindowFullScreen self.setWindowState(state) # Restore the window geometry from the configuration if needed geometry = Config.get('window/geometry', None) if isinstance(geometry, list) and len(geometry) == 4: self.setGeometry(geometry[0], geometry[1], geometry[2], geometry[3]) def closeEvent(self, event): """The window was closed""" # Determine the window state state = self.windowState() window_state = 'none' if state & QtCore.Qt.WindowFullScreen: window_state = 'fullscreen' elif state & QtCore.Qt.WindowMaximized: window_state = 'maximized' elif state & QtCore.Qt.WindowMinimized: window_state = 'minimized' else: # Save the window geometry for normal state geometry = self.geometry() Config.set('window/geometry', [ geometry.x(), geometry.y(), geometry.width(), geometry.height() ]) # Save the window state Config.set('window/state', window_state) # Save the configuration Config.save() # Continue to close the window QtWidgets.QMainWindow.closeEvent(self, event) def set_title(self, panel=None): """Update the window title from the current panel""" # Set the new window title, if provided by the panel if panel is not None and panel.get_title(): title = panel.get_title() self.setWindowTitle(f'{App.get_name()} - {title}') container = self.get_panel_container() if container is not None: index = container.indexOf(panel.get_widgets()) if index >= 0: container.setTabText(index, title) container.setTabIcon(index, panel.get_icon()) else: self.setWindowTitle(f'{App.get_name()}') def set_status(self, panel=None, timeout=0): """Update the window status from the current panel""" self.set_status_text('' if panel is None else panel.get_status(), timeout) def set_status_text(self, message, timeout=0): """Update the window status from the current panel""" if self._statusbar is not None: self._statusbar.showMessage('' if message is None else message, timeout) def get_panel_container(self): """Get the panel container widget""" return self._panel def get_panels(self): """Get all the panels in the window panel stack""" if not isinstance(self._panels, list): self._panels = [] return self._panels def get_panel(self, name): """Get a panel in the window panel stack by name""" # Check panels by name for panel in self.get_panels(): if panel.get_name() == name: # Return the panel return panel # Panel not found return None def get_panel_by_index(self, index): """Get a panel in the window panel stack by index""" # Get panel container container = self.get_panel_container() if container is not None: # Get the widgets of the panel widgets = container.widget(index) if widgets is not None: # Check the panels for matching widgets for panel in self.get_panels(): if panel.get_widgets() == widgets: # Found the panel return panel # Panel not found return None def get_current_panel(self): """Get the current panel in the window stack""" return self._current_panel def get_current_popup(self): """Get the current popup dialog""" return self._current_popup def push_view(self, panel): """Push a panel onto the stack being viewed.""" if panel is not None: Logger.debug(f'MEG UI: Adding panel "{panel.get_name()}"') # Hide the current panel current_panel = self.get_current_panel() if current_panel is not None: current_panel.on_hide() # Show the current panel panel.on_show() # Update the title for the panel self.set_title(panel) # Update the status for the panel self.set_status(panel) # Get the window central widget container = self.get_panel_container() if container is not None: # Add the panel to the view stack widgets = panel.get_widgets() widgets.setParent(container) title = panel.get_title() index = container.addTab(widgets, 'Home' if not title else title) # Remove the close button if not closable tabbar = container.tabBar() if not panel.get_is_closable(): tabbar.tabButton(index, QtWidgets.QTabBar.RightSide).deleteLater() tabbar.setTabButton(index, QtWidgets.QTabBar.RightSide, None) # Add the panel icon tabbar.setTabIcon(index, panel.get_icon()) # Add the panel to the panel stack self.get_panels().append(panel) # Set the panel to the view container.setCurrentIndex(index) def set_view(self, panel): """Set the panel to be viewed in the stack or push the panel onto the stack being viewed.""" if panel is not None: # Get the window central widget container = self.get_panel_container() if container is not None: # Get the index of the panel index = container.indexOf(panel.get_widgets()) if index >= 0: # Set the new panel container.setCurrentIndex(index) # Do not continue since the panel was found do not push Logger.debug(f'MEG UI: Setting panel "{panel.get_name()}"') return # Push the panel instead because it was not found self.push_view(panel) def popup_view(self, panel, resizable=False): """Popup a dialog containing a panel.""" if panel is None or self._current_popup is not None: return QtWidgets.QDialog.Rejected # Create a dialog window to popup dialog = QtWidgets.QDialog(None, QtCore.Qt.WindowSystemMenuHint | QtCore.Qt.WindowTitleHint | QtCore.Qt.WindowCloseButtonHint) dialog.setModal(True) dialog.setSizeGripEnabled(resizable) # Set the current popup self._current_popup = dialog # Set dialog layout layout = QtWidgets.QGridLayout() layout.setContentsMargins(0, 0, 0, 0) dialog.setLayout(layout) # Add the panel widgets to the popup widgets = panel.get_widgets() layout.addWidget(widgets) widgets.setParent(dialog) # Set the dialog icon icon = panel.get_icon() dialog.setWindowIcon(icon if icon else QtWidgets.QIcon(App.get_icon())) title = panel.get_title() # Set the dialog title dialog.setWindowTitle(title if title else App.get_name()) previous_panel = self._current_panel # Hide the current panel if previous_panel is not None: previous_panel.on_hide() # Make the panel the current self._current_panel = panel # Show the panel panel.on_show() # Show the dialog if not resizable: dialog.setFixedSize(dialog.size()) result = dialog.exec_() # Hide the panel panel.on_hide() # Remove the popup self._current_popup = None # Restore the previous panel to current self._current_panel = previous_panel # Show the previous panel if previous_panel is not None: previous_panel.on_show() return result def remove_view(self, panel): """Remove a panel from the stack being viewed.""" # Check if the panel is closable if panel is not None and panel.get_is_closable(): Logger.debug(f'MEG UI: Removing panel "{panel.get_name()}"') # Close the panel panel.on_hide() panel.on_close() # Remove the panel from the list panels = self.get_panels() if panel in panels: panels.remove(panel) if self._current_panel == panel: self._current_panel = None # Get the window central widget container = self.get_panel_container() if container: # Get the index of this panel index = container.indexOf(panel.get_widgets()) if index >= 0: # Remove the panel from the view stack container.removeTab(index) panel.get_widgets().setParent(None) def remove_view_by_index(self, index): """Remove a panel from the stack being viewed.""" # Get the panel by index Logger.debug(f'MEG UI: Removing panel by index ({index})') panel = self.get_panel_by_index(index) if panel is not None and panel.get_is_closable(): # Remove the panel self.remove_view(panel) def _show_view_by_index(self, index): """Show the panel on click""" # Get the panel by index panel = self.get_panel_by_index(index) if panel is not None: # Get the current panel current_panel = self.get_current_panel() # Check if the panel is not the current panel if current_panel != panel: # Hide the current panel if current_panel is not None: current_panel.on_hide() # Set the current panel self._current_panel = panel # Update the title self.set_title(panel) # Update the status self.set_status(panel) # Show the new panel if panel is not None: panel.on_show()
13,517
3,576
"""testing """ from energy_demand.scripts import s_rs_raw_shapes import numpy as np def test_assign_hes_data_to_year(): hes_data = { 'working_day': { 0: np.zeros((24, 2)) + 10, 1: np.zeros((24, 2)) + 20, 2: np.zeros((24, 2)) + 30, 3: np.zeros((24, 2)) + 40, 4: np.zeros((24, 2)) + 50, 5: np.zeros((24, 2)) + 60, 6: np.zeros((24, 2)) + 70, 7: np.zeros((24, 2)) + 80, 8: np.zeros((24, 2)) + 90, 9: np.zeros((24, 2)) + 100, 10: np.zeros((24, 2)) + 110, 11: np.zeros((24, 2)) + 120}, 'holiday' : { 0: np.zeros((24, 2)) + 1, 1: np.zeros((24, 2)) + 2, 2: np.zeros((24, 2)) + 3, 3: np.zeros((24, 2)) + 4, 4: np.zeros((24, 2)) + 5, 5: np.zeros((24, 2)) + 6, 6: np.zeros((24, 2)) + 7, 7: np.zeros((24, 2)) + 8, 8: np.zeros((24, 2)) + 9, 9: np.zeros((24, 2)) + 10, 10: np.zeros((24, 2)) + 11, 11: np.zeros((24, 2)) + 12} } result = s_rs_raw_shapes.assign_hes_data_to_year( nr_of_appliances=2, hes_data=hes_data, base_yr=2015) #daytype, month_python, appliances assert result[10][0][1] == 1 # yearday, hour, appliance_nr--> sun assert result[11][0][1] == 10 # yearday, hour, appliance_nr--> mon
1,459
692
import argparse import logging import yaml from pathlib import Path from script_utils.common import common_setup from release.davis16.compute_flow import link_splits from release.helpers.misc import msg, subprocess_call def check_tracks(track_output, splits): for split in splits: np_dir = track_output / split if not np_dir.exists(): raise ValueError(f'Did not find tracks in {np_dir}; ' f'did you run release/davis17/track.py?') def evaluate_proposed(config, output_stage): if output_stage == 'detection': input_dir = (Path(config['davis16']['output_dir']) / 'detections') elif output_stage == 'tracking': input_dir = (Path(config['davis16']['output_dir']) / 'tracks') else: raise ValueError(f'Unknown output stage: {output_stage}') for split in config['davis16']['splits']: masks_dir = input_dir / split / 'masks' / 'masks' cmd = [ 'python', 'davis/eval_fgbg.py', '--masks-dir', masks_dir ] msg(f'Evaluating {split}') subprocess_call(cmd) def main(): # Use first line of file docstring as description if it exists. parser = argparse.ArgumentParser( description=__doc__.split('\n')[0] if __doc__ else '', formatter_class=argparse.ArgumentDefaultsHelpFormatter) parser.add_argument('output_stage', choices=['detection', 'tracking'], default='detection') parser.add_argument('--config', default=Path('./release/config.yaml')) args = parser.parse_args() logging.getLogger().setLevel(logging.INFO) logging.basicConfig(format='%(asctime)s.%(msecs).03d: %(message)s', datefmt='%H:%M:%S') with open(args.config, 'r') as f: config = yaml.load(f) evaluate_proposed(config, args.output_stage) if __name__ == "__main__": main()
1,935
618
# The MIT License (MIT) # Copyright (c) 2021-present foxwhite25 # # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the "Software"), # to deal in the Software without restriction, including without limitation # the rights to use, copy, modify, merge, publish, distribute, sublicense, # and/or sell copies of the Software, and to permit persons to whom the # Software is furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS # OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. from __future__ import annotations from typing import Type, TypeVar, List, TYPE_CHECKING, Any, Union __all__ = ( 'AllowedMentions', ) if TYPE_CHECKING: from .types.message import AllowedMentions as AllowedMentionsPayload from .member import Member from .role import Role class _FakeBool: def __repr__(self): return 'True' def __eq__(self, other): return other is True def __bool__(self): return True default: Any = _FakeBool() A = TypeVar('A', bound='AllowedMentions') class AllowedMentions: """一个类,表示消息中允许提及的内容。 这个类可以在 :class:`Client` 初始化期间设置,以应用于每条发送的消息。 它也可以通过 :meth:`abc.Messageable.send` 在每条消息的基础上应用,以获得更细粒度的控制。 Attributes ------------ everyone: :class:`bool` 是否允许所有人和这里提到。 默认为 ``True``。 users: Union[:class:`bool`, List[:class:`Member`]] 控制被提及的用户。 如果为 ``True`` (默认值),则根据消息内容提及用户。 如果 ``False`` 则根本不会提及用户。 如果给出了 :class:`Member` 的列表,则只提及所提供的用户,前提是这些用户在消息内容中。 roles: Union[:class:`bool`, List[:class:`Role`]] 控制提到的用户组。 如果为 ``True`` (默认值),则根据消息内容提及用户组。 如果 ``False`` 则根本不提及用户组。 如果给出了 :class:`Role` 的列表,则只提及所提供的用户组,前提是这些用户组在消息内容中。 replied_user: :class:`bool` 是否提及正在回复的消息的作者。 默认为 ``True`` 。 """ __slots__ = ('everyone', 'users', 'roles', 'replied_user') def __init__( self, *, everyone: bool = default, users: Union[bool, List[Member]] = default, roles: Union[bool, List[Role]] = default, replied_user: bool = default, ): self.everyone = everyone self.users = users self.roles = roles self.replied_user = replied_user @classmethod def all(cls: Type[A]) -> A: """返回一个 :class:`AllowedMentions` 的工厂方法,其中所有字段都显式设置为 ``True``""" return cls(everyone=True, users=True, roles=True, replied_user=True) @classmethod def none(cls: Type[A]) -> A: """一个工厂方法,返回一个 :class:`AllowedMentions`,所有字段都设置为 ``False``""" return cls(everyone=False, users=False, roles=False, replied_user=False) def to_dict(self) -> AllowedMentionsPayload: parse = [] data = {} if self.everyone: parse.append('everyone') if self.users == True: parse.append('users') elif self.users != False: data['users'] = [x.id for x in self.users] if self.roles == True: parse.append('roles') elif self.roles != False: data['roles'] = [x.id for x in self.roles] if self.replied_user: data['replied_user'] = True data['parse'] = parse return data # type: ignore def merge(self, other: AllowedMentions) -> AllowedMentions: # Creates a new AllowedMentions by merging from another one. # Merge is done by using the 'self' values unless explicitly # overridden by the 'other' values. everyone = self.everyone if other.everyone is default else other.everyone users = self.users if other.users is default else other.users roles = self.roles if other.roles is default else other.roles replied_user = self.replied_user if other.replied_user is default else other.replied_user return AllowedMentions(everyone=everyone, roles=roles, users=users, replied_user=replied_user) def __repr__(self) -> str: return ( f'{self.__class__.__name__}(everyone={self.everyone}, ' f'users={self.users}, roles={self.roles}, replied_user={self.replied_user})' )
4,712
1,740
import tkinter as tk class Kinds: def __init__(self, master, *args, **kwargs): self.master = master self.imethods = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAPCAYAAAA71pVKAAAACXBIWXMAAA7DAAAOwwHHb6hk AAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAddJREFUKJGNkjFrk3EQxn/3vtEW8QMICoK CNZYqDn4BndQ2NIMZRHES0TqILomW9toG7aJiW4cOQkGKoAhNagUHEZwspHQQm9SiDoI4iBqJJNb3/Z9D+0IaE+ iz3d3z4+7gEZpoLru0OwjkDpAE5p3H5eRAfKHRJ/XF5GRhy46v2y9hNmxITiCNcCqqfeRqj3Z8+w/Oa+kouAlDq oL1JfTAm2j2VFd2+YQ3PeykiQwvWse4qjhZA5eyhlxAuL5o+x+oimv2Tn5o+biZGxMoJjSe8AAMuQJUzPg9qFgz cM1nNYGyQc8zfb/HW+/7AvcFRvJaKswOvTtWD80MF7vyWprD7IkTmwYqq4QxAchpseog/hf3pU38i5j1AwXDbnv IaYOUIRNGdTSph3/mtFgO4UisfkNKO1eBe49HP0y111bTIC8MHlksiPf2d31ufMNrbACk0nvLCY1ngKrzY4PNwJ bwZhXBoeC2bQZ4pZ/aAX8rfrAO213Be50bKp5XtZbXzOjSiV/U3gIvu3Xfxw0JM2wcqEUJy2mx4vzYIRfyp2XCI j0fW2kLvgfXDDIC0wbnwKZAzoA99NuCG92Zgz8i/wY40mx2eWcY2C3BzgosOM/19Q50zjf6/gEMUNa2RFgfkAAA AABJRU5ErkJggg==""") self.ivariables = None self.ifields = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAANCAYAAAB2HjRBAAAACX BIWXMAAA7DAAAOwwHHb6hkAAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAcBJREFUKJF1k jFoUwEQhr97SQzaqODi5KCTRRA3K6iQxA5BnIpZHCwIioJK074XqEZSi2hIK0gpilBBcNStVMX4nkWtFDqVLi5F wUlECBrRNu/9Tq0xbW68+7/j/uM3OtSQr8MOlAQ5jPkoYnQ8a89bNbYBqumIGcMYPSYmYwkmm01yghLQMHGrmuE pZlqH3ZqO4XADOGgwts24X07bz7X5+QUldv7gLGLY4HsUcdncQIeAu4huQTXV4EH5lP3qaGdWey1kBkjFEc8Mmq HoGc/a505Q0deBEFxC+gxmBHkH2CfjoxlLnq+H3qy6W6HB1zru+ZqO4D3wNRT7V1e4CBAHaP6hP5lkRyiuEPLB8 /UOMS2jH9gjuOeEnKn0Wh1g4IV2rcMAd9L2CSgUX2kkinEVYwK40PWNJ+W8rWxmxWlvVHqt3hRTwGo1Y4/awYE5 bU1s4Ryw7ABfYklOlwPF2xf99+WX6hryVYj/ZllGDqMvroi84zDRAM8NdLu+ncfU/0HlQKmGuAQMAotE5Ksn7C2 sJUwy9w0nESVgt2DK4JrBiKAALDhitJK1udZrNsSzGOhoBDcRaaAmuD6WsfnNrPwFMR6ziXNIHQ8AAAAASUVORK 5CYII=""") self.iclasses = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAPCAYAAAA71pVKAAAAC XBIWXMAAA7DAAAOwwHHb6hkAAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAdFJREFUKJGN 0k1IVFEYxvH/c+descDgzsaCFhHUeDPBatcmIheCIi2slasZpDZB22hTBrWqhSEhzUdRIAxB1KpshJYRtTAoxgp qYYILvVMimDP3vK2KYZSxZ3U4nN/Lw8sRTYlLvf3mkgpmlXBx35iuvW7QJt7fQ60YHTeXVIA7SFG8f7lkZVI74j h/5JgzmxVcT+eqN801hpCN1taikbY4LvX2m9ysZBNhrnrXyr0d8vx7oHf1De9VO+ybS+aA22F2YfJb6UBnvJY8B XZvBsHg3tyH9VawUsyM4tSX3lOd0Go+eoSsz/luyGt4eWBQsrNhduFZK1wtZsYwTQGLwLysTCr+FT1ANgq8BdLA T0lLzdCZ+YIzeN6wkuSUSZcFYGVStbVo5HcQzHbUN+dM/BD25Z80pYFxTMMShwy7KtyAtlQr9LyR7FZzbZs+EcT ++hOgD+gSbiDMfZ73WvF20YX39bArdQ5ZwWDEl1uCpk+y44DzHzelXVPAZN3873Ehc9rf7qGh7lrx8MHmu4Sg09 zGY0EVrGzo+RZs8EnGtGspJRKAGZmumKgAD7csbLss3z/aHXiNr0g3MBsHXobZ6qX/wgArxZ6TMl4AM2G2elHC/ gCskr1TRm4JBAAAAABJRU5ErkJggg==""") self.iinterfaces = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAICAYAAAAm06XyAA AACXBIWXMAAA7DAAAOwwHHb6hkAAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAQRJREFUG JWFkD9LgnEAhJ/7qbODQw211JLRVlMEofYFgqIhglpsqSnfCJeaChRXgz5Aa30A/1ARCbnXJ7Aai5rC95oEkcAb D4577sQYFbtOpb+IMLuIKeBJCQ4qq3pJAkRNzwoylTzPSB4Opz+pI5YtIsf0QmDDfTrRnZdUantfpg4EzG21oPV BsNT2gsx9KiZ7vqaPgR+1fYHJKmr5DZgcKrsGvgEEc8B7Ja+tYZqjhudDoBPGbf5PyYAAkoJTwyUQEDfVnLZHsB /KDU8MY/fNjqApgJOGZ34hUyvQHT3suOUrmxUFysT0HNjEFBNiUeMQz9pO/sSUJPYM08BjLA5rOb3+AQBaXk0WW bcyAAAAAElFTkSuQmCC""") self.imodules = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAPCAYAAAA71pVKAAAAC XBIWXMAAA7DAAAOwwHHb6hkAAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAXtJREFUKJGN 0TuLE2EYxfH/eZLIikkweEFb2xQWostMOhUCUwha+BG0007wEwh22ilabmNpF1CbZCbeECzS2iqsMrmJu2bzvBa OQkJMfMr3nB8ceEVxnU7nSLVavQ9cb7VaJ1i6NE13gWfT6fROu93+DmB/wlqt9gC4FEK4uQwBJN0ALha9328Ag8 Hg0HA4HJnZ5SiK0lUYoN/vt9z9xWg0Opokyb4B5Hl+HNiazWaf/gUBinyrXq8fW5gNYGZah5dzAyiVSqeAAzPLN +AcOAghnAZQmqbngYeSduM4vrIOA2RZ9hw4CdwqA2+Bubuf3QQB3P2upI/AG5O0Lem9md37H1z03gEXBNDtds+Z 2WtJ9TiOf6yZfDiEMDaz7SiKPhiApM9A2d0bGyY3gPJ8Pv8CS1/l7mEDXsgNYDwefwP2KpXKmXW4yPcajcbXvzh Jkn1JO+7+JMuyq6tgr9e75u5PJe00m82fC7Mnk8lt4FUI4fEqLOkR8LLoAfALta2R2TiY6xMAAAAASUVORK5CYI I=""") self.iproperties = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAPCAYAAAA71pVKAAAACXB IWXMAAA7DAAAOwwHHb6hkAAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAdFJREFUKJGN0r 9rFEEUB/Dvd+8uxzUhoBCCEkhtEbTbHVdZIiQQfySKWmkjWtkIWtj4D1imsDQIRlCDoBbauOLNzR5aGrBVEY0g2 gQCtzfztTALinenrxt4n++8Nwwxorrd7ni/379O8gyAKQDvAdwyxqwAQDQCTnrvLcndJBcBnAewC0Be9QzF3vub JJ8aYy6GEGYkrZBcMMZsWGv3WWsvcxDM83xibGzsQ6/X29NoNPaTfExyPkmS1+12ezaKoucAfgy8udVqTZP8lGX ZVlmW7yR9BLBYFMUBks9IXgEwM2zsr5Km8jyvZ1n2rSzLuRDCUgjhhaRLkt4C2ByIvfezANBsNicBoAqQtJym6R NJ1wCs/7VzURRHQgh3oyg6Gcdxx1o7XZbl5yzL+t1ud9x7fwPA0RBC/Ad2zs1JWvsNZiQfAQgANgHsBfCQ5NUkS b7XK2itzSStSToVx3GnCgJwnOSG936iVqt9SZJkuzLcgQdJrocQzqZp+tI5d1jS/eo85FHBTqdzCMADAKeNMa+c c/OS7lSjD4MAUAdwG8C5HbggaRXAUhzHxSgI/PqeLUnbzrllSashhBPGmH9CAKhLukDynqQtAMfSNH3zPxAAfgL 2/u9cQzl88QAAAABJRU5ErkJggg==""") self.ikeywords = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAPCAYAAAA71pVKAAAACXBIW XMAAA7DAAAOwwHHb6hkAAAAGXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAAPRJREFUKJGlkrFK xEAURc8bHltbWqytYG0TYiC2VnZ+hK3iByhouX7HfkcmhAgigpXFgt02gpXETObZiEVMWI23PecOl+HBPyLe+5c xaGaLLMsWY1yBHRE5NLO3Hrtwzm0BFEVxJCI3Q2XM7BIIPbYLPAOo6n2M8exHWUROxmaFEJ4AkiRZA+sxb1KkLM vrTVKM8SHLsmVVVfMY4ynwnqbplZrZbFPZOacATdOIqs7MrP8/E2Z774+HQAjhMc/zFUBVVfOu6/b7jgLnQ2VVv QVWXw/ticigNzlS1/X2b+UQwkeapq/f69q2Hb3tgdwBB39ZN5pP4uJac+7GJRAAAAAASUVORK5CYII=""") self.iwords = tk.PhotoImage(data="""iVBORw0KGgoAAAANSUhEUgAAAA8AAAAICAYAAAAm06XyAAAACXBIWXMAAA7DAAAOwwHHb6hkAAAA GXRFWHRTb2Z0d2FyZQB3d3cuaW5rc2NhcGUub3Jnm+48GgAAATBJREFUGJWFjyFLQ2EYhc95t31JEQYiGExaLBO DYbvJaFBEGBqHGNW4n6BFBUGw2ZUJCoaxMGT3Xm+1LwkLgrILwzLZdwy7wTD0iec9z4EXmEAcx40wDPcn3X5j/x X+gmEYLgA4JLkE4KFSqdzEcdwA0Jc0A0CSzoMgSNrt9pRzrg5gRVLXSNZIfki6AnDa6XRKGBsbku5IPpNsJkky5 5x7AjAv6ZLkV344HJ4VCoUdM9uUNMrlcsuZfBEEwS0ARFG0670/BrDY6/XWq9XqCEAr75xrABgAeASwKilHElmG bIgkpwF8ZuL45yiKUgCBpG+SLwCOSG5LmvXebwEomdk9yTVJife+ViwWm/1+f9tI1iW1ssIrgIGkd0lmZm9mdk1 yr1wudwFUzewkTdOU5MEPhImOHsSTJnYAAAAASUVORK5CYII=""")
6,395
4,509
# Generated by Django 3.0 on 2020-03-29 14:09 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.CreateModel( name='UserProfile', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('location', models.CharField(blank=True, default=0, max_length=30, null=True)), ('age', models.IntegerField()), ('occupation', models.IntegerField(choices=[(0, 'Undergraduate Student'), (1, 'Graduate Student'), (2, 'Faculty Member'), (3, 'Staff')], default=4)), ('first_name', models.CharField(max_length=30)), ('last_name', models.CharField(max_length=150)), ('user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], ), ]
1,125
344
import struct import sys import math import numpy as np # # TODO 1: (NOW FIXED) Find the first occurrence of magic and start from there # TODO 2: Warn if we cannot parse a specific section and try to recover # TODO 3: Remove error at end of file if we have only fragment of TLV # def tlvHeaderDecode(data): tlvType, tlvLength = struct.unpack('2I', data) return tlvType, tlvLength def parseDetectedObjects(data, numObj, tlvLength): detected_points = struct.unpack(str(numObj * 4) + 'f', data[:tlvLength]) detected_points = np.asarray(detected_points).reshape(numObj, 4) return detected_points def parseRangeProfile(data, tlvLength): # an integer is 2 byte long range_bins = tlvLength / 2 range_profile = struct.unpack(str(int(range_bins)) + 'H', data[:tlvLength]) return range_profile def parseRDheatmap(data, tlvLength, range_bins, rm_clutter=True): """ range bins times doppler bins times 2, doppler bins = chirps/ frame divided by num of antennas TX (3) #default chirps per frame is (128/3) = 42 * 2 * 256 the call to replace_left_right mirror-flips left and right after reshaping. replace_left_right is equivalent to this line from mmWave.js in the visualizer code # rangeDoppler = rangeDoppler.slice((rangeDoppler.length + 1) / 2).concat( # rangeDoppler.slice(0, (rangeDoppler.length + 1) / 2)); :param range_bins: :param data: the incoming byte stream to be interpreted as range-doppler heatmap/profile :param tlvLength: :return: """ doppler_bins = (tlvLength / 2) / range_bins rd_heatmap = struct.unpack(str(int(range_bins * doppler_bins)) + 'H', data[:tlvLength]) rd_heatmap = np.reshape(rd_heatmap, (int(range_bins), int(doppler_bins))) overall_mean = np.mean(rd_heatmap) if rm_clutter: rd_heatmap = np.array([row - np.mean(row) for row in rd_heatmap]) return replace_left_right(rd_heatmap) def chg_val(val): return val - 65536 if val > 32767 else val def parseAziheatmap(data, tlvLength, range_bins): """ :param range_bins: :param data: the incoming byte stream to be interpreted as range-doppler heatmap/profile :param tlvLength: :return: """ # range_bins = 256 azi_bins = (tlvLength / 2) / range_bins azi_heatmap = struct.unpack(str(int(range_bins * azi_bins)) + 'H', data[:tlvLength]) # azi_heatmap = [chg_val(x) for x in azi_heatmap] azi_heatmap = np.reshape(azi_heatmap, (int(range_bins), int(azi_bins))) # use the default order of 3 Tx's and ordering is TX0, TX1, TX2 row_indices = [7, 5, 11, 9] qrows = 4 qcols = range_bins rowSizeBytes = 48 q = data[:tlvLength] qq = [] for col in range(qcols): real = [] img = [] for row in range(qrows): index = col * rowSizeBytes + 4 * row_indices[row] real.append(q[index + 1] * 256 + q[index]) img.append(q[index + 3] * 256 + q[index + 2]) real = [chg_val(x) for x in real] img = [chg_val(x) for x in img] # convert to complex numbers data = np.array([real, img]).transpose() data = np.pad(data, ((0, 60), (0, 0)), 'constant', constant_values=0) data = data[..., 0] + 1j * data[..., 1] transformed = np.fft.fft(data) # take the magnitude transformed = np.absolute(transformed) qq.append(np.concatenate((transformed[int(len(transformed) / 2):], transformed[:int(len(transformed) / 2)]))) qq = np.array(qq) return qq def replace_left_right(a): rtn = np.empty(shape=a.shape) rtn[:, :int(rtn.shape[1] / 2)] = a[:, int(rtn.shape[1] / 2):] rtn[:, int(rtn.shape[1] / 2):] = a[:, :int(rtn.shape[1] / 2)] return rtn def parseStats(data): interProcess, transmitOut, frameMargin, chirpMargin, activeCPULoad, interCPULoad = struct.unpack('6I', data[:24]) return interProcess, transmitOut, frameMargin, chirpMargin, activeCPULoad, interCPULoad # print("\tOutputMsgStats:\t%d " % (6)) # print("\t\tChirpMargin:\t%d " % (chirpMargin)) # print("\t\tFrameMargin:\t%d " % (frameMargin)) # print("\t\tInterCPULoad:\t%d " % (interCPULoad)) # print("\t\tActiveCPULoad:\t%d " % (activeCPULoad)) # print("\t\tTransmitOut:\t%d " % (transmitOut)) # print("\t\tInterprocess:\t%d " % (interProcess)) negative_rtn = False, None, None, None, None, None class tlv_header_decoder(): def __init__(self): pass def decode_iwr_tlv(in_data): """ Must disable range profile for the quick RD heatmap to work, this way the number of range bins will be be calculated from the absent range profile. You can still get the range profile by inferring it from the RD heatmap :param in_data: :return: if no detected point at this frame, the detected point will be an empty a """ magic = b'\x02\x01\x04\x03\x06\x05\x08\x07' header_length = 36 offset = in_data.find(magic) data = in_data[offset:] if len(data) < header_length: return negative_rtn try: data_magic, version, length, platform, frameNum, cpuCycles, numObj, numTLVs = struct.unpack('Q7I', data[ :header_length]) except struct.error: print("Improper TLV structure found: ", (data,)) return negative_rtn # print("Packet ID:\t%d "%(frameNum)) # print("Version:\t%x "%(version)) # print("Data Len:\t\t%d", length) # print("TLV:\t\t%d "%(numTLVs)) # print("Detect Obj:\t%d "%(numObj)) # print("Platform:\t%X "%(platform)) if version >= 50462726 and len(data) >= length: # if version > 0x01000005 and len(data) >= length: try: sub_frame_num = struct.unpack('I', data[36:40])[0] header_length = 40 # print("Subframe:\t%d "%(subFrameNum)) pending_bytes = length - header_length data = data[header_length:] detected_points = None range_profile = None rd_heatmap = None azi_heatmap = None range_bins = 8 statistics = None for i in range(numTLVs): tlvType, tlvLength = tlvHeaderDecode(data[:8]) data = data[8:] if tlvType == 1: # print('Outputting Points') detected_points = parseDetectedObjects(data, numObj, tlvLength) # if no detected points, tlvType won't have 1 elif tlvType == 2: # the range bins is modified in the range profile is enabled range_profile = parseRangeProfile(data, tlvLength) elif tlvType == 4: # resolving static azimuth heatmap pass elif tlvType == 5: # try: # assert range_bins # except AssertionError: # raise Exception('Must enable range-profile while enabling range-doppler-profile, in order to' # 'interpret the number of range bins') rd_heatmap = parseRDheatmap(data, tlvLength, range_bins) elif tlvType == 6: # TODO why is the states' TLV not present? interProcess, transmitOut, frameMargin, chirpMargin, activeCPULoad, interCPULoad = parseStats(data) pass elif tlvType == 7: pass elif tlvType == 8: # resolving static azimuth-elevation heatmap try: azi_heatmap = parseAziheatmap(data, tlvLength, range_bins) except: print('bad azimuth') azi_heatmap = None pass elif tlvType == 9: # only for AoP EV2 pass else: # print("Unidentified tlv type %d" % tlvType, '. Its len is ' + str(tlvLength)) n_offset = data.find(magic) if n_offset != offset and n_offset != -1: print('New magic found, discarding previous frame with unknown tlv') data = data[n_offset:] return True, data, detected_points, range_profile, rd_heatmap, azi_heatmap data = data[tlvLength:] pending_bytes -= (8 + tlvLength) data = data[pending_bytes:] # data that are left # infer range profile from heatmap is the former is not enabled if range_profile is None and rd_heatmap is not None and len(rd_heatmap) > 0: range_profile = rd_heatmap[:, 0] return True, data, detected_points, range_profile, rd_heatmap, azi_heatmap except struct.error as se: print('Failed to parse tlv message, type = ' + str(tlvType) + ', error: ') print(se) pass return negative_rtn if __name__ == "__main__": magic = b'\x02\x01\x04\x03\x06\x05\x08\x07' fileName = 'D:/PycharmProjects/mmWave_gesture_iwr6843/test_data2.dat' rawDataFile = open(fileName, "rb") rawData = rawDataFile.read() rawDataFile.close() offset = rawData.find(magic) rawData = rawData[offset:] # for i in range(len(rawData/36)) # # for length, frameNum in tlvHeader(rawData): # print
9,686
3,238
from io import StringIO from os import path, listdir, remove from math import radians, tan, cos, pi, atan, sin from pandas import read_csv import sympy as sy import numpy as np # these variables are used to solve symbolic mathematical equations # x is the control variable over the height ... max(x) = H_cross_section x = sy.Symbol('x', real=True, positive=True) accuracy = 10 def to_num(x): if x == '': return None elif x.replace('-', '').isdecimal(): return int(x) elif ('.' in x) and (x.lower().replace('.', '').replace('-', '').replace('e', '').isdecimal()): return float(x) else: return x def csv(txt, comment=None): """ Read the string in txt as csv file and return the content as DataFrame. Args: txt (str): content of csv comment (str): comment sign Returns: dict: profile label and values """ df = read_csv(StringIO(txt), index_col=0, skipinitialspace=True, skip_blank_lines=True, comment=comment) df = df[df.index.notnull()].copy() df.index = df.index.astype(str) return df def to_xs_dict(txt, comment=None): """ Read the string in txt as csv file and return the content as DataFrame. Args: txt (str): content of csv comment (str): comment sign Returns: dict: profile label and values """ di = dict() names = [] for line in txt.split('\n'): if line == '': continue elif isinstance(comment, str) and line.startswith(comment): continue elif not names: names = [n.strip() for n in line.split(',')[1:]] di['_names'] = names else: name, *values = [n.strip() for n in line.split(',')] # di[name] = {k: to_num(v) for k, v in zip(names, values)} di[name] = [to_num(v) for v in values] return di def deg2slope(degree): """ convert degrees to a slope (:math:`\\Delta x / \\Delta y`) Args: degree (float): angle in degree Returns: float: slope .. figure:: images/slope.gif :align: center :alt: slope :figclass: align-center Slope """ return tan(radians(degree)) def channel_end(r, end_degree): """ get vertical end of the channel based on the radius of the channel and an end angle Args: r (float): radius of the channel end_degree (float): end angle in degree (°) Returns: float: height of the channel when the circle reaches a certain angle .. figure:: images/channel_end.gif :align: center :alt: channel end :figclass: align-center Channel end """ return r * (1 - cos(radians(end_degree))) def sqrt(i): """ Return the square root of x. """ return i ** (1 / 2) def combine_input_files(shape_path, delete_original=False): """combine all generated shape text files to a single inp-like text file When running the :func:`shape_generator.shape_generator.Profile.input_file` function, a .txt file will be created. Those txt files will be combines to a single file with this function. This makes it easier to import all shapes to the .inp file. Args: shape_path (str): path where the shapes are stored delete_original (bool): whether to delete the original single files """ with open(os.path.join(shape_path, 'all_shapes.txt'), 'w') as outfile: for fname in listdir(shape_path): if not fname.endswith('_shape.txt'): continue in_fn = os.path.join(shape_path, fname) with open(in_fn) as infile: outfile.write(infile.read()) outfile.write('\n\n') if delete_original: remove(in_fn) print('Files are combined and originals {}deleted.'.format('' if delete_original else 'NOT ')) #################################################################################################################### class CustomExpr: def __init__(self): pass def __repr__(self): return 'Custom Function' def expr(self): pass def solve(self, i): pass def length(self, i0, i1): pass def area(self, i0, i1): pass #################################################################################################################### class Slope(CustomExpr): """ get function/expression of a straight line with a given point which it intercepts Args: slope (float): slope p0 (set[float, float]): point as a set of a x and a y coordinate Returns: sympy.core.expr.Expr: linear function .. figure:: images/gerade.gif :align: center :alt: straight line :figclass: align-center Straight line """ def __init__(self, slope, unit=None): if unit is None or unit == '': self.slope = slope elif unit == '°': self.slope = deg2slope(slope) elif unit == '%': self.slope = slope / 100 else: raise NotImplementedError('Unknown Unit for slope function') self.x0 = None self.y0 = None self.x1 = None self.y1 = None CustomExpr.__init__(self) def __repr__(self): return f'Slope Function (k={self.slope:0.2f}, zero=[{self.x0:0.2f}, {self.y0:0.2f}])' def set_start_point(self, point): """set start point""" x0, y0 = point self.x0 = x0 self.y0 = y0 def set_end_point(self, point): """set end point""" x1, y1 = point self.x1 = x1 self.y1 = y1 def expr(self): """get sympy expression""" return self.y0 + (x - self.x0) / self.slope def solve(self, i): """get y value""" return self.y0 + (i - self.x0) / self.slope @classmethod def from_points(cls, start, end): """ set the slope by giving the start and end point Args: start (): end (): Returns: """ x0, f0 = start x1, f1 = end if abs(f0 - f1) < 1.0e-6: return Vertical(f0) elif abs(x0 - x1) < 1.0e-6: return Horizontal.from_points(start, end) slope = (x1 - x0) / (f1 - f0) new_slope = cls(slope) new_slope.set_start_point(start) new_slope.set_end_point(end) return new_slope def end_point(self): """get the end point""" return self.x1, self.y1 def length(self, i0, i1): """get shape length between two values""" return sqrt((self.solve(i0) - self.solve(i1)) ** 2 + (i0 - i1) ** 2) def area(self, i0, i1): """get shape area between two values""" return (self.solve(i0) + self.solve(i1)) / 2 * np.abs(i0 - i1) #################################################################################################################### class Vertical(CustomExpr): """ function of a vertical line """ def __init__(self, y): """ Args: y (float): y value of the vertical line """ self.y = y CustomExpr.__init__(self) def __repr__(self): return f'Vertical Function (y={self.y:0.2f})' def expr(self): return self.y + x * 0 def solve(self, i): return self.y + i * 0 def length(self, i0, i1): return i1 - i0 def area(self, i0, i1): return self.length(i0, i1) * self.y #################################################################################################################### class Horizontal(CustomExpr): """ function of a horizontal line """ def __init__(self): CustomExpr.__init__(self) self.x = None self.y0 = None self.y1 = None def set_x(self, i): self.x = i def set_points(self, start, end): x0, y0 = start x1, y1 = end if x0 == x1: self.x = x0 else: if x0 is not None: self.x = x0 elif x1 is not None: self.x = x1 self.y0 = y0 self.y1 = y1 def __repr__(self): return 'Horizontal Function' def expr(self): return self.y1 def solve(self, i): return self.y1 def length(self, i0, i1): return np.abs(self.y1 - self.y0) def area(self, i0, i1): return 0 @classmethod def from_points(cls, start, end): h = cls() h.set_points(start, end) return h def start_point(self): return self.x, self.y0 def end_point(self): return self.x, self.y1 #################################################################################################################### class Circle(CustomExpr): """ function of a circle .. figure:: images/kreis.gif :align: center :alt: circle :figclass: align-center Circle """ def __init__(self, r, x_m=0, y_m=0, clockwise=False): """ Args: r (float): radius x_m (float): x axis value of the mid point y_m (float): y axis value of the mid point clockwise (bool): whether the circle is clockwise or anticlockwise """ self.r = float(r) self.x_m = float(x_m) self.y_m = float(y_m) self.clockwise = clockwise CustomExpr.__init__(self) def __repr__(self): return f'Circle Function (radius={self.r:0.2f}, mid=[{self.x_m:0.2f}, {self.y_m:0.2f}])' def expr(self): """ get function/expression of a circle with a given mid point Returns: sympy.core.expr.Expr: function of the circle """ return sy.sqrt(sy.Float(self.r) ** 2 - (x - sy.Float(self.x_m)) ** 2) * (-1 if self.clockwise else 1) + \ sy.Float(self.y_m) def _alpha(self, i): """ angle in the circle of a point to the horizontal Args: i: variable Returns: float: angle in rad """ if isinstance(i, np.ndarray): return np.arctan((i - self.x_m) / (self.solve(i) - self.y_m)) else: if (self.solve(i) - self.y_m) == 0: a = pi / 2 if (i - self.x_m) < 0: a *= -1 else: a = np.arctan((i - self.x_m) / (self.solve(i) - self.y_m)) return a def _d_alpha(self, i0, i1): """ difference of the angle in the circle of two points Args: i0: start variable i1: end variable Returns: float: difference of the angle in rad """ return np.abs(self._alpha(i0) - self._alpha(i1)) def solve(self, i): return sqrt(self.r ** 2 - (i - self.x_m) ** 2) * (-1 if self.clockwise else 1) + self.y_m def length(self, i0, i1): return self._d_alpha(i0, i1) * self.r def area(self, i0, i1): alpha = self._d_alpha(i0, i1) return self.r ** 2 / 2 * (alpha - np.sin(alpha)) + (self.solve(i0) + self.solve(i1)) / 2 * (i1 - i0)
11,285
3,609
from __future__ import annotations import json import socket import requests from bunq.sdk.context.api_context import ApiContext, ApiEnvironmentType from bunq.sdk.exception.bunq_exception import BunqException from bunq.sdk.http.api_client import ApiClient from bunq.sdk.model.generated import endpoint from bunq.sdk.model.generated.endpoint import SandboxUser __UNIQUE_REQUEST_ID = "uniqueness-is-required" __FIELD_API_KEY = "ApiKey" __INDEX_FIRST = 0 __FIELD_RESPONSE = "Response" __ENDPOINT_SANDBOX_USER = "sandbox-user" _ERROR_COULD_NOT_CREATE_NEW_SANDBOX_USER = "Could not create new sandbox user." def automatic_sandbox_install() -> ApiContext: sandbox_user = __generate_new_sandbox_user() return ApiContext.create(ApiEnvironmentType.SANDBOX, sandbox_user.api_key, socket.gethostname() ) def __generate_new_sandbox_user() -> SandboxUser: url = ApiEnvironmentType.SANDBOX.uri_base + __ENDPOINT_SANDBOX_USER headers = { ApiClient.HEADER_REQUEST_ID: __UNIQUE_REQUEST_ID, ApiClient.HEADER_CACHE_CONTROL: ApiClient.CACHE_CONTROL_NONE, ApiClient.HEADER_GEOLOCATION: ApiClient.GEOLOCATION_ZERO, ApiClient.HEADER_LANGUAGE: ApiClient.LANGUAGE_EN_US, ApiClient.HEADER_REGION: ApiClient.REGION_NL_NL, } response = requests.request(ApiClient.METHOD_POST, url, headers=headers) if response.status_code is ApiClient.STATUS_CODE_OK: response_json = json.loads(response.text) return endpoint.SandboxUser.from_json( json.dumps(response_json[__FIELD_RESPONSE][__INDEX_FIRST][ __FIELD_API_KEY])) raise BunqException(_ERROR_COULD_NOT_CREATE_NEW_SANDBOX_USER)
1,772
609