text
string
size
int64
token_count
int64
# Copyright 2017 The Nuclio Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # uses vader lib (will be installed automatically via build commands) to identify sentiments in the body string # return score result in the form of: {'neg': 0.0, 'neu': 0.323, 'pos': 0.677, 'compound': 0.6369} # # @nuclio.configure # # function.yaml: # apiVersion: "nuclio.io/v1beta1" # kind: "Function" # spec: # runtime: "python" # # build: # commands: # - "pip install requests vaderSentiment" # from vaderSentiment.vaderSentiment import SentimentIntensityAnalyzer def handler(context, event): body = event.body.decode('utf-8') context.logger.debug_with('Analyzing ', 'sentence', body) analyzer = SentimentIntensityAnalyzer() score = analyzer.polarity_scores(body) return str(score)
1,323
440
import requests class KakaoMapClient: def __init__(self, kakao_auth_key): self.kakao_map_url = "https://dapi.kakao.com/v2/local/search/address.json" self.kakao_auth_key = kakao_auth_key def get_latitude_and_longitudes(self, pharmacies): searched_pharamacies = [] for pharmacy in pharmacies: (latitude, longitude) = self.get_latitude_and_longitude(pharmacy[1]) pharmacy.append(latitude) pharmacy.append(longitude) searched_pharamacies.append(pharmacy) return searched_pharamacies def get_latitude_and_longitude(self, addr): # parse addr by road name parsed_addr="" for idx, entity in reversed(list(enumerate(addr.split()))): if ("로" in entity) or ("길" in entity): parsed_addr = " ".join(addr.split()[:(idx+2)]) print("searching address : " + parsed_addr) query_pharams = { 'query': parsed_addr, 'analyze_type': 'similar', 'page': 1, 'size': 10 } auth_header = { 'Authorization': self.kakao_auth_key } response = requests.get(self.kakao_map_url, params=query_pharams, headers=auth_header) latitude = "0" longitude = "0" try: response_addr = response.json()["documents"][0]["address"] latitude=response_addr["y"] longitude=response_addr["x"] except: print("no info : " + parsed_addr) # print("longitude: " + longitude + "latitude: " + latitude) return (latitude, longitude)
1,636
515
#!/usr/bin/env python3 # -*- coding:utf-8 -*- # Author : Alenx.Hai <alenx.hai@gmail.com> # created time: 2020/12/21-10:49 上午 import asyncio from src.mysql_elastic import MySQLElasticSearch @asyncio.coroutine def main(): elastic = MySQLElasticSearch() yield elastic.put_data() if __name__ == '__main__': loop = asyncio.get_event_loop() loop.run_until_complete(asyncio.wait([main()]))
404
176
# Crie um módulo chamado moeda.py que tenha as funções incorporadas aumentar(), diminuir(), dobro() e metade(). Faça # também um programa que importe esse módulo e use algumas dessas funções. import moeda p = float(input('Digite o preço: ')) print(f'A metade do {p} é R${moeda.metade(p)}') print(f'O dobro de {p} é R${moeda.dobro(p)}') print(f'Aumentando 10%, temos R${moeda.aumentar(p, 10)}')
397
162
"""Class to hold all cover accessories.""" import logging from homeassistant.components.cover import ATTR_CURRENT_POSITION from homeassistant.helpers.event import async_track_state_change from . import TYPES from .accessories import HomeAccessory, add_preload_service from .const import ( SERV_WINDOW_COVERING, CHAR_CURRENT_POSITION, CHAR_TARGET_POSITION, CHAR_POSITION_STATE) _LOGGER = logging.getLogger(__name__) @TYPES.register('Window') class Window(HomeAccessory): """Generate a Window accessory for a cover entity. The cover entity must support: set_cover_position. """ def __init__(self, hass, entity_id, display_name): """Initialize a Window accessory object.""" super().__init__(display_name, entity_id, 'WINDOW') self._hass = hass self._entity_id = entity_id self.current_position = None self.homekit_target = None self.serv_cover = add_preload_service(self, SERV_WINDOW_COVERING) self.char_current_position = self.serv_cover. \ get_characteristic(CHAR_CURRENT_POSITION) self.char_target_position = self.serv_cover. \ get_characteristic(CHAR_TARGET_POSITION) self.char_position_state = self.serv_cover. \ get_characteristic(CHAR_POSITION_STATE) self.char_target_position.setter_callback = self.move_cover def run(self): """Method called be object after driver is started.""" state = self._hass.states.get(self._entity_id) self.update_cover_position(new_state=state) async_track_state_change( self._hass, self._entity_id, self.update_cover_position) def move_cover(self, value): """Move cover to value if call came from HomeKit.""" if value != self.current_position: _LOGGER.debug("%s: Set position to %d", self._entity_id, value) self.homekit_target = value if value > self.current_position: self.char_position_state.set_value(1) elif value < self.current_position: self.char_position_state.set_value(0) self._hass.services.call( 'cover', 'set_cover_position', {'entity_id': self._entity_id, 'position': value}) def update_cover_position(self, entity_id=None, old_state=None, new_state=None): """Update cover position after state changed.""" if new_state is None: return current_position = new_state.attributes[ATTR_CURRENT_POSITION] if current_position is None: return self.current_position = int(current_position) self.char_current_position.set_value(self.current_position) if self.homekit_target is None or \ abs(self.current_position - self.homekit_target) < 6: self.char_target_position.set_value(self.current_position) self.char_position_state.set_value(2) self.homekit_target = None
3,024
888
import pandas as pd import numpy as np import datetime import pytz from pathlib import Path import warnings from io import IOBase from . import calpost_reader calpost_cat = calpost_reader.calpost_cat def hysplit_reader_long(f, tslice=slice(None, None), x=None, y=None, z=None, rdx_map=None): """reads hysplit output file, returns dict of numpy arrays :param FileIO f: either (1)opened hysplit output file, (2) hysplit output filename or (3) list of (1) or (2) :param slice tslice: slice of time index :param list x: list of x coords :param list y: list of y coords :return: dict, with ['v'] has data as 3d array (t, y, x) :rtype: dict """ print(type(f)) if isinstance(f, IOBase): raise ValueError('plese pass filename, not FileIO...') # assume file name passed if 'f' is string if isinstance(f, (str, Path)): df = pd.read_csv(f, sep=r'\s+') return hysplit_reader_long(df, tslice, x, y, z, rdx_map) # list of files may have different time period and locations. So # first they are grouped by time perod, then each chunk got read. # then they got joined with the time stiching routine aware of # spin-up time if isinstance(f, list): lines = [next(pd.read_csv(fn, sep=r'\s+', nrows=1).itertuples()) for fn in f] # Pandas(Index=0, JDAY=268.208, YR1=19, MO1=9, DA1=25, HR1=5, MN1=0, # YR2=19, MO2=9, DA2=25, HR2=5, MN2=1, Pol=1, Lev=1, Station=1, # Value=0.0) print(lines) dtes = [datetime.datetime(_.YR1, _.MO1, _.DA1, _.HR1, _.MN1).replace(tzinfo=pytz.utc).astimezone(pytz.timezone('Etc/GMT+6')) for _ in lines] df_fnames = pd.DataFrame({'fname': f, 'datetime': dtes}) df_fnames.to_csv('fnames.csv') # group the file names by the datetime dct_fnames = {} for fn,dte in zip(f, dtes): dct_fnames.setdefault(dte, []).append(fn) file_dates = list(dct_fnames.keys()) dat = [] for dte,fnames in dct_fnames.items(): dfs = [pd.read_csv(fn, sep=r'\s+') for fn in fnames] df = pd.concat(dfs) dat.append( hysplit_reader_long(df, tslice, x, y, z, rdx_map) ) dat = calpost_cat(dat, use_later_files=True) dat['ts'] = dat['ts'][tslice] dat['v'] = dat['v'][tslice] return dat # now i should be getting dataframe df = f units = '???' print('dt') # extremely slow! #df['Datetime'] = [datetime.datetime(_.YR1, _.MO1, _.DA1, _.HR1, # _.MN1).replace(tzinfo=pytz.utc).astimezone(pytz.timezone('Etc/GMT+6')) # for _ in df.itertuples()] df['Datetime'] = pd.to_datetime(df[['YR1', 'MO1', 'DA1', 'HR1', 'MN1']].assign( YR1= lambda df: df['YR1'] + 2000).rename( columns={'YR1':'year', 'MO1':'month', 'DA1': 'day', 'HR1': 'hour', 'MN1': 'minute'}), utc=True).dt.tz_convert('Etc/GMT+6') # bad idea! #df['Datetime_tup'] = [_ for _ in df[['YR1', 'MO1', 'DA1', 'HR1', # 'MN1']].itertuples(index=False)] df = df[['Datetime', 'Lev', 'Station', 'Value']] #grouped = df.groupby(['Datetime', 'Lev', 'Station']) nrec = len(df.index) print('set_index') df = df[['Datetime', 'Lev', 'Station', 'Value']].set_index( ['Datetime', 'Station', 'Lev'] ) print('dt') ts = df.index.levels[0] #xxx = pd.DataFrame(ts, columns=('year', 'month', 'day', 'hour', # 'minute')) #print(xxx) #xxx = xxx.assign(year=lambda x: x['year']+2000) #print(xxx) # #ts = pd.to_datetime( # pd.DataFrame( # ts, # columns=('year', 'month', 'day', 'hour', 'minute') # ).assign( # year=lambda x: x['year']+2000 # )) #print(ts) print('cont') stations = df.index.levels[1] nz = len(df.index.levels[2]) nsta = len(df.index.levels[1]) nt = len(df.index.levels[0]) print('nt,nz,nsta,nrec=', nt, nz, nsta, nrec) # ........ bad idea #assert nt * nz * nsta == nrec if not nt * nz * nsta == nrec: print(f'expected {nt*nz*nsta} rec, got {nrec}, short by {nt*nz*nsta-nrec}') print(' f:', f) print(' rng:', df.index.levels[0][0], df.index.levels[0][-1]) print('unstack') df = df.unstack().unstack() df.columns = df.columns.droplevel() if rdx_map: x = rdx_map.x y = rdx_map.y nx = len(x) ny = len(y) grid = rdx_map.grid v = df.to_numpy() if rdx_map.coverage == 'full, c-order' and nsta==nx*ny: v = v.reshape(nt, nz, ny, nx) elif rdx_map.coverage == 'full, f-order' and nsta==nx*ny: raise NotImplementedError( 'qa first! receptor def = "{}", '.format(rdx_map.coverage)) v = v.reshape(nt, nz, nx, ny) v = np.swapaxes(v, -1, -2) elif rdx_map.coverage in ('full, c-order', 'full, f-order', 'full, random', 'patial, random'): rdx = np.arange(nt*nz) + 1 mymap = rdx_map.get_index(stations).to_numpy() mymap = mymap[:, ::-1] vv = np.empty((nt, nz, ny, nx)) vv[...] = np.nan v = v.reshape(nt , nz, -1) for tt,t in zip(vv, v): for zz, z in zip(tt, t): for ji,p in zip(mymap,z): zz[tuple(ji)] = p v = vv else: raise ValueError('rdx_map is mandatory for now') #dct = {'v': v, 'ts': ts, 'units': units, 'df': f, 'name': None} dct = {'v': v, 'ts': ts, 'units': units, 'name': None} dct.update( {'x': x, 'y': y, 'grid': grid, }) del df return dct
5,888
2,278
import unittest from yapper import create_app, db from yapper.blueprints.user.models import User, Role class TestUserAddToDb(unittest.TestCase): def setUp(self): self.app = create_app('test') self.app_context = self.app.app_context() self.app_context.push() db.create_all() def tearDown(self): db.session.remove() db.drop_all() self.app_context.pop() def test_role_gets_id(self): role = Role(name='admin') self.assertTrue(role.id is None) db.session.add(role) db.session.commit() self.assertFalse(role.id is None) def test_user_gets_role_and_id(self): role = Role(name='administrator') self.assertTrue(role.id is None) user = User(email='b2@gmail.com', password='1234', role=role) self.assertTrue(user.id is None) db.session.add(user) db.session.commit() self.assertFalse(role.id is None) self.assertFalse(user.id is None) self.assertTrue(user.role_id == role.id) self.assertTrue(user.is_admin())
1,094
364
# Copyright (C) 2010-2011 Richard Lincoln # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from CIM16.IEC61970.Core.Curve import Curve class StartIgnFuelCurve(Curve): """The quantity of ignition fuel (Y-axis) used to restart and repay the auxiliary power consumed versus the number of hours (X-axis) the unit was off lineThe quantity of ignition fuel (Y-axis) used to restart and repay the auxiliary power consumed versus the number of hours (X-axis) the unit was off line """ def __init__(self, ignitionFuelType="oil", StartupModel=None, *args, **kw_args): """Initialises a new 'StartIgnFuelCurve' instance. @param ignitionFuelType: Type of ignition fuel Values are: "oil", "coal", "lignite", "gas" @param StartupModel: The unit's startup model may have a startup ignition fuel curve """ #: Type of ignition fuel Values are: "oil", "coal", "lignite", "gas" self.ignitionFuelType = ignitionFuelType self._StartupModel = None self.StartupModel = StartupModel super(StartIgnFuelCurve, self).__init__(*args, **kw_args) _attrs = ["ignitionFuelType"] _attr_types = {"ignitionFuelType": str} _defaults = {"ignitionFuelType": "oil"} _enums = {"ignitionFuelType": "FuelType"} _refs = ["StartupModel"] _many_refs = [] def getStartupModel(self): """The unit's startup model may have a startup ignition fuel curve """ return self._StartupModel def setStartupModel(self, value): if self._StartupModel is not None: self._StartupModel._StartIgnFuelCurve = None self._StartupModel = value if self._StartupModel is not None: self._StartupModel.StartIgnFuelCurve = None self._StartupModel._StartIgnFuelCurve = self StartupModel = property(getStartupModel, setStartupModel)
2,880
915
import pyecharts.options as opts from pyecharts.charts import Line """ Gallery 使用 pyecharts 1.1.0 参考地址: https://echarts.baidu.com/examples/editor.html?c=line-aqi 目前无法实现的功能: 1、dataZoom 放大的时候无法固定 Y 轴的上下限 """ all_data = [ ["2000-06-05", 116], ["2000-06-06", 129], ["2000-06-07", 135], ["2000-06-08", 86], ["2000-06-09", 73], ["2000-06-10", 85], ["2000-06-11", 73], ["2000-06-12", 68], ["2000-06-13", 92], ["2000-06-14", 130], ["2000-06-15", 245], ["2000-06-16", 139], ["2000-06-17", 115], ["2000-06-18", 111], ["2000-06-19", 309], ["2000-06-20", 206], ["2000-06-21", 137], ["2000-06-22", 128], ["2000-06-23", 85], ["2000-06-24", 94], ["2000-06-25", 71], ["2000-06-26", 106], ["2000-06-27", 84], ["2000-06-28", 93], ["2000-06-29", 85], ["2000-06-30", 73], ["2000-07-01", 83], ["2000-07-02", 125], ["2000-07-03", 107], ["2000-07-04", 82], ["2000-07-05", 44], ["2000-07-06", 72], ["2000-07-07", 106], ["2000-07-08", 107], ["2000-07-09", 66], ["2000-07-10", 91], ["2000-07-11", 92], ["2000-07-12", 113], ["2000-07-13", 107], ["2000-07-14", 131], ["2000-07-15", 111], ["2000-07-16", 64], ["2000-07-17", 69], ["2000-07-18", 88], ["2000-07-19", 77], ["2000-07-20", 83], ["2000-07-21", 111], ["2000-07-22", 57], ["2000-07-23", 55], ["2000-07-24", 60], ["2000-07-25", 44], ["2000-07-26", 127], ["2000-07-27", 114], ["2000-07-28", 86], ["2000-07-29", 73], ["2000-07-30", 52], ["2000-07-31", 69], ["2000-08-01", 86], ["2000-08-02", 118], ["2000-08-03", 56], ["2000-08-04", 91], ["2000-08-05", 121], ["2000-08-06", 127], ["2000-08-07", 78], ["2000-08-08", 79], ["2000-08-09", 46], ["2000-08-10", 108], ["2000-08-11", 80], ["2000-08-12", 79], ["2000-08-13", 69], ["2000-08-14", 80], ["2000-08-15", 105], ["2000-08-16", 119], ["2000-08-17", 105], ["2000-08-18", 55], ["2000-08-19", 74], ["2000-08-20", 41], ["2000-08-21", 62], ["2000-08-22", 104], ["2000-08-23", 118], ["2000-08-24", 121], ["2000-08-25", 126], ["2000-08-26", 99], ["2000-08-27", 92], ["2000-08-28", 75], ["2000-08-29", 91], ["2000-08-30", 94], ["2000-08-31", 69], ["2000-09-01", 93], ["2000-09-02", 124], ["2000-09-03", 120], ["2000-09-04", 93], ["2000-09-05", 26], ["2000-09-06", 32], ["2000-09-07", 70], ["2000-09-08", 89], ["2000-09-10", 117], ["2000-09-11", 144], ["2000-09-12", 111], ["2000-09-13", 120], ["2000-09-14", 97], ["2000-09-15", 108], ["2000-09-17", 74], ["2000-09-18", 105], ["2000-09-19", 127], ["2000-09-20", 143], ["2000-09-21", 62], ["2000-09-22", 80], ["2000-09-23", 136], ["2000-09-24", 29], ["2000-09-25", 91], ["2000-09-26", 93], ["2000-09-27", 114], ["2000-09-28", 45], ["2000-09-29", 102], ["2000-09-30", 111], ["2000-10-01", 93], ["2000-10-02", 117], ["2000-10-03", 78], ["2000-10-04", 76], ["2000-10-05", 100], ["2000-10-06", 75], ["2000-10-07", 169], ["2000-10-08", 59], ["2000-10-09", 89], ["2000-10-10", 91], ["2000-10-11", 75], ["2000-10-12", 28], ["2000-10-13", 47], ["2000-10-14", 92], ["2000-10-16", 72], ["2000-10-17", 149], ["2000-10-18", 86], ["2000-10-19", 88], ["2000-10-20", 104], ["2000-10-21", 91], ["2000-10-22", 88], ["2000-10-23", 55], ["2000-10-24", 63], ["2000-10-25", 41], ["2000-10-26", 85], ["2000-10-27", 99], ["2000-10-28", 121], ["2000-10-29", 96], ["2000-10-30", 90], ["2000-11-01", 80], ["2000-11-02", 116], ["2000-11-03", 207], ["2000-11-04", 306], ["2000-11-05", 283], ["2000-11-06", 200], ["2000-11-07", 93], ["2000-11-08", 49], ["2000-11-09", 78], ["2000-11-10", 40], ["2000-11-11", 74], ["2000-11-12", 67], ["2000-11-13", 118], ["2000-11-14", 196], ["2000-11-15", 101], ["2000-11-16", 59], ["2000-11-17", 83], ["2000-11-18", 83], ["2000-11-19", 124], ["2000-11-20", 57], ["2000-11-21", 78], ["2000-11-22", 113], ["2000-11-23", 172], ["2000-11-24", 129], ["2000-11-25", 103], ["2000-11-26", 75], ["2000-11-27", 125], ["2000-11-28", 121], ["2000-11-29", 204], ["2000-11-30", 141], ["2000-12-01", 106], ["2000-12-02", 146], ["2000-12-03", 95], ["2000-12-04", 149], ["2000-12-05", 71], ["2000-12-07", 157], ["2000-12-08", 141], ["2000-12-09", 197], ["2000-12-10", 43], ["2000-12-11", 81], ["2000-12-12", 109], ["2000-12-13", 118], ["2000-12-15", 115], ["2000-12-16", 92], ["2000-12-17", 123], ["2000-12-18", 147], ["2000-12-19", 59], ["2000-12-20", 103], ["2000-12-21", 146], ["2000-12-22", 137], ["2000-12-23", 74], ["2000-12-24", 64], ["2000-12-25", 67], ["2000-12-26", 107], ["2000-12-27", 101], ["2000-12-28", 79], ["2000-12-29", 137], ["2000-12-30", 165], ["2000-12-31", 81], ["2001-01-01", 100], ["2001-01-02", 126], ["2001-01-03", 56], ["2001-01-05", 108], ["2001-01-06", 88], ["2001-01-07", 78], ["2001-01-08", 105], ["2001-01-09", 77], ["2001-01-10", 105], ["2001-01-11", 93], ["2001-01-12", 107], ["2001-01-13", 128], ["2001-01-14", 53], ["2001-01-15", 81], ["2001-01-16", 128], ["2001-01-17", 179], ["2001-01-18", 225], ["2001-01-19", 116], ["2001-01-20", 153], ["2001-01-21", 161], ["2001-01-22", 149], ["2001-01-23", 115], ["2001-01-24", 136], ["2001-01-25", 101], ["2001-01-26", 109], ["2001-01-27", 108], ["2001-01-28", 86], ["2001-01-29", 101], ["2001-01-30", 109], ["2001-01-31", 139], ["2001-02-01", 110], ["2001-02-02", 113], ["2001-02-03", 130], ["2001-02-04", 62], ["2001-02-05", 88], ["2001-02-06", 105], ["2001-02-07", 87], ["2001-02-08", 140], ["2001-02-09", 116], ["2001-02-10", 100], ["2001-02-11", 83], ["2001-02-12", 102], ["2001-02-13", 106], ["2001-02-14", 157], ["2001-02-15", 131], ["2001-02-16", 77], ["2001-02-17", 101], ["2001-02-18", 148], ["2001-02-19", 227], ["2001-02-20", 105], ["2001-02-21", 155], ["2001-02-22", 293], ["2001-02-23", 99], ["2001-02-24", 57], ["2001-02-25", 97], ["2001-02-26", 104], ["2001-02-27", 117], ["2001-02-28", 125], ["2001-03-01", 216], ["2001-03-02", 149], ["2001-03-03", 256], ["2001-03-04", 172], ["2001-03-05", 113], ["2001-03-06", 338], ["2001-03-07", 57], ["2001-03-08", 48], ["2001-03-10", 111], ["2001-03-11", 87], ["2001-03-12", 175], ["2001-03-13", 186], ["2001-03-14", 201], ["2001-03-15", 76], ["2001-03-16", 131], ["2001-03-17", 127], ["2001-03-18", 128], ["2001-03-19", 152], ["2001-03-20", 144], ["2001-03-21", 162], ["2001-03-22", 500], ["2001-03-24", 358], ["2001-03-25", 128], ["2001-03-26", 54], ["2001-03-27", 57], ["2001-03-28", 54], ["2001-03-29", 80], ["2001-03-30", 71], ["2001-03-31", 73], ["2001-04-01", 139], ["2001-04-02", 224], ["2001-04-03", 107], ["2001-04-04", 150], ["2001-04-05", 180], ["2001-04-06", 77], ["2001-04-07", 95], ["2001-04-08", 194], ["2001-04-09", 143], ["2001-04-10", 205], ["2001-04-11", 129], ["2001-04-12", 64], ["2001-04-13", 61], ["2001-04-14", 79], ["2001-04-15", 121], ["2001-04-16", 130], ["2001-04-17", 150], ["2001-04-18", 205], ["2001-04-19", 154], ["2001-04-20", 81], ["2001-04-21", 140], ["2001-04-22", 119], ["2001-04-23", 156], ["2001-04-24", 72], ["2001-04-25", 108], ["2001-04-26", 124], ["2001-04-27", 94], ["2001-04-28", 157], ["2001-04-29", 100], ["2001-04-30", 158], ["2001-05-01", 277], ["2001-05-02", 332], ["2001-05-03", 303], ["2001-05-04", 238], ["2001-05-05", 500], ["2001-05-06", 99], ["2001-05-07", 93], ["2001-05-08", 104], ["2001-05-09", 74], ["2001-05-10", 68], ["2001-05-11", 90], ["2001-05-12", 114], ["2001-05-13", 142], ["2001-05-14", 126], ["2001-05-15", 185], ["2001-05-16", 402], ["2001-05-17", 189], ["2001-05-17", 189], ["2001-05-17", 189], ["2001-05-18", 112], ["2001-05-19", 137], ["2001-05-20", 158], ["2001-05-21", 158], ["2001-05-22", 116], ["2001-05-23", 132], ["2001-05-24", 110], ["2001-05-25", 82], ["2001-05-26", 56], ["2001-05-27", 54], ["2001-05-28", 71], ["2001-05-29", 101], ["2001-05-30", 57], ["2001-05-31", 88], ["2001-06-01", 99], ["2001-06-02", 84], ["2001-06-03", 139], ["2001-06-04", 132], ["2001-06-05", 141], ["2001-06-07", 159], ["2001-06-08", 131], ["2001-06-09", 180], ["2001-06-10", 164], ["2001-06-11", 134], ["2001-06-12", 163], ["2001-06-13", 105], ["2001-06-14", 74], ["2001-06-15", 50], ["2001-06-16", 60], ["2001-06-17", 82], ["2001-06-18", 111], ["2001-06-19", 89], ["2001-06-20", 81], ["2001-06-21", 76], ["2001-06-22", 70], ["2001-06-23", 74], ["2001-06-24", 99], ["2001-06-25", 91], ["2001-06-26", 113], ["2001-06-27", 93], ["2001-06-28", 69], ["2001-06-29", 74], ["2001-06-30", 75], ["2001-07-01", 108], ["2001-07-02", 115], ["2001-07-03", 86], ["2001-07-04", 67], ["2001-07-05", 68], ["2001-07-06", 74], ["2001-07-07", 69], ["2001-07-08", 95], ["2001-07-09", 99], ["2001-07-10", 92], ["2001-07-11", 84], ["2001-07-12", 77], ["2001-07-13", 69], ["2001-07-14", 62], ["2001-07-15", 83], ["2001-07-16", 101], ["2001-07-17", 98], ["2001-07-18", 89], ["2001-07-19", 82], ["2001-07-20", 105], ["2001-07-21", 79], ["2001-07-22", 48], ["2001-07-23", 119], ["2001-07-24", 126], ["2001-07-25", 44], ["2001-07-26", 42], ["2001-07-27", 86], ["2001-07-28", 68], ["2001-07-29", 93], ["2001-07-30", 89], ["2001-07-31", 76], ["2001-08-01", 54], ["2001-08-02", 53], ["2001-08-03", 35], ["2001-08-04", 65], ["2001-08-05", 108], ["2001-08-06", 114], ["2001-08-07", 90], ["2001-08-08", 63], ["2001-08-09", 79], ["2001-08-10", 102], ["2001-08-11", 100], ["2001-08-12", 107], ["2001-08-13", 81], ["2001-08-14", 79], ["2001-08-15", 116], ["2001-08-16", 98], ["2001-08-17", 96], ["2001-08-18", 94], ["2001-08-19", 63], ["2001-08-20", 39], ["2001-08-21", 81], ["2001-08-22", 73], ["2001-08-23", 66], ["2001-08-24", 52], ["2001-08-25", 64], ["2001-08-26", 61], ["2001-08-27", 83], ["2001-08-28", 85], ["2001-08-29", 99], ["2001-08-30", 97], ["2001-08-31", 93], ["2001-09-01", 86], ["2001-09-02", 105], ["2001-09-03", 98], ["2001-09-04", 109], ["2001-09-05", 92], ["2001-09-06", 68], ["2001-09-07", 92], ["2001-09-08", 72], ["2001-09-09", 64], ["2001-09-10", 88], ["2001-09-11", 97], ["2001-09-12", 102], ["2001-09-13", 103], ["2001-09-14", 120], ["2001-09-15", 94], ["2001-09-16", 95], ["2001-09-17", 93], ["2001-09-18", 56], ["2001-09-19", 98], ["2001-09-20", 81], ["2001-09-21", 100], ["2001-09-22", 75], ["2001-09-23", 84], ["2001-09-24", 91], ["2001-09-25", 70], ["2001-09-26", 96], ["2001-09-27", 128], ["2001-09-28", 92], ["2001-09-29", 107], ["2001-09-30", 95], ["2001-10-01", 63], ["2001-10-02", 115], ["2001-10-03", 69], ["2001-10-04", 47], ["2001-10-05", 86], ["2001-10-06", 122], ["2001-10-07", 104], ["2001-10-08", 122], ["2001-10-09", 49], ["2001-10-10", 36], ["2001-10-11", 83], ["2001-10-12", 107], ["2001-10-13", 126], ["2001-10-14", 126], ["2001-10-15", 78], ["2001-10-16", 72], ["2001-10-17", 76], ["2001-10-18", 87], ["2001-10-19", 143], ["2001-10-20", 259], ["2001-10-21", 183], ["2001-10-22", 276], ["2001-10-23", 232], ["2001-10-24", 167], ["2001-10-25", 105], ["2001-10-26", 129], ["2001-10-27", 140], ["2001-10-28", 61], ["2001-10-29", 85], ["2001-10-30", 155], ["2001-11-01", 38], ["2001-11-02", 106], ["2001-11-03", 134], ["2001-11-04", 57], ["2001-11-05", 51], ["2001-11-06", 68], ["2001-11-07", 129], ["2001-11-08", 158], ["2001-11-09", 85], ["2001-11-10", 121], ["2001-11-11", 161], ["2001-11-12", 94], ["2001-11-13", 58], ["2001-11-14", 57], ["2001-11-15", 71], ["2001-11-16", 105], ["2001-11-17", 66], ["2001-11-18", 117], ["2001-11-19", 87], ["2001-11-20", 88], ["2001-11-21", 131], ["2001-11-22", 151], ["2001-11-23", 310], ["2001-11-24", 161], ["2001-11-25", 23], ["2001-11-26", 52], ["2001-11-27", 82], ["2001-11-28", 128], ["2001-11-29", 115], ["2001-11-30", 63], ["2001-12-02", 102], ["2001-12-03", 96], ["2001-12-04", 107], ["2001-12-05", 89], ["2001-12-06", 59], ["2001-12-07", 100], ["2001-12-08", 136], ["2001-12-09", 137], ["2001-12-10", 119], ["2001-12-11", 112], ["2001-12-12", 186], ["2001-12-13", 192], ["2001-12-14", 83], ["2001-12-15", 97], ["2001-12-16", 113], ["2001-12-18", 89], ["2001-12-19", 106], ["2001-12-20", 119], ["2001-12-21", 62], ["2001-12-22", 79], ["2001-12-23", 58], ["2001-12-24", 61], ["2001-12-25", 64], ["2001-12-26", 108], ["2001-12-27", 101], ["2001-12-28", 82], ["2001-12-29", 85], ["2001-12-30", 98], ["2001-12-31", 132], ["2002-01-01", 88], ["2002-01-02", 97], ["2002-01-03", 116], ["2002-01-04", 111], ["2002-01-05", 81], ["2002-01-06", 78], ["2002-01-07", 138], ["2002-01-08", 100], ["2002-01-09", 157], ["2002-01-10", 349], ["2002-01-11", 196], ["2002-01-12", 190], ["2002-01-13", 100], ["2002-01-14", 103], ["2002-01-15", 160], ["2002-01-16", 97], ["2002-01-17", 103], ["2002-01-18", 123], ["2002-01-19", 137], ["2002-01-20", 268], ["2002-01-21", 52], ["2002-01-22", 44], ["2002-01-23", 66], ["2002-01-24", 106], ["2002-01-25", 94], ["2002-01-26", 96], ["2002-01-27", 58], ["2002-01-28", 62], ["2002-01-29", 56], ["2002-01-30", 62], ["2002-01-31", 109], ["2002-02-01", 96], ["2002-02-02", 95], ["2002-02-03", 126], ["2002-02-04", 161], ["2002-02-05", 138], ["2002-02-06", 106], ["2002-02-07", 99], ["2002-02-08", 113], ["2002-02-09", 80], ["2002-02-10", 90], ["2002-02-11", 86], ["2002-02-12", 142], ["2002-02-13", 93], ["2002-02-14", 125], ["2002-02-15", 135], ["2002-02-16", 138], ["2002-02-17", 111], ["2002-02-18", 70], ["2002-02-19", 101], ["2002-02-20", 153], ["2002-02-21", 146], ["2002-02-22", 97], ["2002-02-23", 82], ["2002-02-24", 99], ["2002-02-25", 131], ["2002-02-26", 88], ["2002-02-27", 74], ["2002-02-28", 96], ["2002-03-01", 133], ["2002-03-02", 105], ["2002-03-03", 86], ["2002-03-04", 105], ["2002-03-05", 89], ["2002-03-06", 70], ["2002-03-07", 87], ["2002-03-08", 109], ["2002-03-09", 161], ["2002-03-10", 83], ["2002-03-11", 129], ["2002-03-12", 107], ["2002-03-13", 89], ["2002-03-14", 186], ["2002-03-15", 108], ["2002-03-16", 500], ["2002-03-17", 188], ["2002-03-18", 102], ["2002-03-19", 139], ["2002-03-20", 155], ["2002-03-21", 500], ["2002-03-22", 370], ["2002-03-23", 164], ["2002-03-24", 105], ["2002-03-25", 156], ["2002-03-26", 180], ["2002-03-27", 105], ["2002-03-28", 126], ["2002-03-29", 120], ["2002-03-30", 122], ["2002-03-31", 118], ["2002-04-01", 188], ["2002-04-02", 260], ["2002-04-03", 296], ["2002-04-04", 118], ["2002-04-05", 132], ["2002-04-06", 80], ["2002-04-07", 500], ["2002-04-08", 500], ["2002-04-09", 253], ["2002-04-10", 67], ["2002-04-11", 110], ["2002-04-13", 133], ["2002-04-14", 246], ["2002-04-15", 324], ["2002-04-16", 225], ["2002-04-17", 120], ["2002-04-18", 121], ["2002-04-19", 131], ["2002-04-20", 148], ["2002-04-21", 174], ["2002-04-22", 106], ["2002-04-23", 32], ["2002-04-24", 86], ["2002-04-25", 92], ["2002-04-26", 117], ["2002-04-27", 110], ["2002-04-28", 90], ["2002-04-29", 86], ["2002-04-30", 106], ["2002-05-01", 84], ["2002-05-02", 76], ["2002-05-03", 92], ["2002-05-04", 85], ["2002-05-05", 79], ["2002-05-07", 92], ["2002-05-08", 99], ["2002-05-09", 105], ["2002-05-10", 105], ["2002-05-11", 78], ["2002-05-12", 125], ["2002-05-13", 113], ["2002-05-14", 90], ["2002-05-15", 89], ["2002-05-16", 99], ["2002-05-17", 94], ["2002-05-18", 109], ["2002-05-19", 105], ["2002-05-20", 115], ["2002-05-21", 110], ["2002-05-22", 54], ["2002-05-23", 76], ["2002-05-24", 83], ["2002-05-25", 75], ["2002-05-26", 89], ["2002-05-27", 97], ["2002-05-28", 113], ["2002-05-29", 106], ["2002-05-30", 86], ["2002-05-31", 108], ["2002-06-01", 115], ["2002-06-02", 106], ["2002-06-03", 99], ["2002-06-04", 151], ["2002-06-05", 118], ["2002-06-06", 139], ["2002-06-07", 161], ["2002-06-08", 77], ["2002-06-09", 72], ["2002-06-10", 36], ["2002-06-11", 81], ["2002-06-12", 67], ["2002-06-13", 56], ["2002-06-14", 73], ["2002-06-15", 75], ["2002-06-16", 80], ["2002-06-17", 122], ["2002-06-19", 142], ["2002-06-20", 77], ["2002-06-21", 68], ["2002-06-22", 77], ["2002-06-23", 50], ["2002-06-24", 51], ["2002-06-25", 40], ["2002-06-26", 46], ["2002-06-27", 65], ["2002-06-28", 110], ["2002-06-29", 104], ["2002-06-30", 85], ["2002-07-01", 126], ["2002-07-02", 88], ["2002-07-03", 112], ["2002-07-04", 108], ["2002-07-05", 98], ["2002-07-06", 88], ["2002-07-07", 68], ["2002-07-08", 87], ["2002-07-09", 83], ["2002-07-10", 87], ["2002-07-11", 127], ["2002-07-12", 111], ["2002-07-13", 108], ["2002-07-14", 91], ["2002-07-15", 89], ["2002-07-16", 75], ["2002-07-17", 88], ["2002-07-18", 76], ["2002-07-19", 62], ["2002-07-20", 55], ["2002-07-21", 66], ["2002-07-22", 67], ["2002-07-23", 62], ["2002-07-24", 113], ["2002-07-25", 81], ["2002-07-26", 66], ["2002-07-27", 86], ["2002-07-28", 47], ["2002-07-29", 44], ["2002-07-30", 79], ["2002-07-31", 137], ["2002-08-01", 160], ["2002-08-02", 89], ["2002-08-03", 96], ["2002-08-04", 63], ["2002-08-05", 53], ["2002-08-06", 50], ["2002-08-07", 44], ["2002-08-08", 74], ["2002-08-09", 64], ["2002-08-10", 72], ["2002-08-11", 94], ["2002-08-12", 71], ["2002-08-13", 124], ["2002-08-14", 129], ["2002-08-15", 155], ["2002-08-16", 156], ["2002-08-17", 125], ["2002-08-18", 130], ["2002-08-19", 66], ["2002-08-20", 91], ["2002-08-21", 114], ["2002-08-22", 112], ["2002-08-23", 102], ["2002-08-24", 72], ["2002-08-25", 76], ["2002-08-26", 77], ["2002-08-27", 86], ["2002-08-28", 92], ["2002-08-29", 108], ["2002-08-30", 100], ["2002-08-31", 122], ["2002-09-01", 164], ["2002-09-02", 111], ["2002-09-03", 52], ["2002-09-04", 70], ["2002-09-05", 59], ["2002-09-06", 82], ["2002-09-07", 96], ["2002-09-08", 92], ["2002-09-09", 124], ["2002-09-10", 98], ["2002-09-11", 45], ["2002-09-12", 37], ["2002-09-13", 81], ["2002-09-14", 90], ["2002-09-15", 98], ["2002-09-16", 97], ["2002-09-17", 111], ["2002-09-18", 125], ["2002-09-19", 83], ["2002-09-20", 41], ["2002-09-21", 87], ["2002-09-22", 56], ["2002-09-23", 72], ["2002-09-25", 182], ["2002-09-26", 183], ["2002-09-27", 70], ["2002-09-28", 44], ["2002-09-29", 62], ["2002-09-30", 100], ["2002-10-01", 121], ["2002-10-02", 62], ["2002-10-03", 70], ["2002-10-04", 99], ["2002-10-05", 89], ["2002-10-06", 52], ["2002-10-07", 37], ["2002-10-08", 64], ["2002-10-09", 135], ["2002-10-10", 232], ["2002-10-11", 365], ["2002-10-12", 198], ["2002-10-13", 53], ["2002-10-14", 121], ["2002-10-15", 83], ["2002-10-16", 100], ["2002-10-17", 169], ["2002-10-18", 75], ["2002-10-20", 72], ["2002-10-21", 51], ["2002-10-22", 50], ["2002-10-23", 95], ["2002-10-24", 88], ["2002-10-26", 59], ["2002-10-27", 30], ["2002-10-28", 48], ["2002-10-29", 109], ["2002-10-30", 146], ["2002-10-31", 76], ["2002-11-01", 33], ["2002-11-02", 52], ["2002-11-03", 54], ["2002-11-04", 70], ["2002-11-05", 107], ["2002-11-06", 96], ["2002-11-07", 76], ["2002-11-08", 37], ["2002-11-09", 94], ["2002-11-10", 182], ["2002-11-11", 452], ["2002-11-12", 66], ["2002-11-13", 56], ["2002-11-14", 80], ["2002-11-15", 85], ["2002-11-16", 104], ["2002-11-17", 43], ["2002-11-18", 52], ["2002-11-19", 115], ["2002-11-20", 143], ["2002-11-21", 75], ["2002-11-22", 110], ["2002-11-23", 134], ["2002-11-24", 129], ["2002-11-25", 153], ["2002-11-26", 54], ["2002-11-27", 114], ["2002-11-28", 145], ["2002-11-29", 87], ["2002-11-30", 138], ["2002-12-01", 198], ["2002-12-02", 273], ["2002-12-03", 395], ["2002-12-04", 498], ["2002-12-05", 97], ["2002-12-06", 112], ["2002-12-07", 97], ["2002-12-08", 86], ["2002-12-09", 97], ["2002-12-10", 99], ["2002-12-12", 151], ["2002-12-13", 135], ["2002-12-14", 193], ["2002-12-15", 153], ["2002-12-16", 95], ["2002-12-17", 91], ["2002-12-18", 137], ["2002-12-19", 98], ["2002-12-20", 77], ["2002-12-21", 95], ["2002-12-22", 96], ["2002-12-23", 83], ["2002-12-24", 71], ["2002-12-25", 53], ["2002-12-26", 69], ["2002-12-27", 75], ["2002-12-28", 106], ["2002-12-29", 90], ["2002-12-30", 106], ["2002-12-31", 64], ["2003-01-01", 105], ["2003-01-02", 100], ["2003-01-03", 69], ["2003-01-04", 55], ["2003-01-05", 65], ["2003-01-06", 112], ["2003-01-07", 83], ["2003-01-08", 131], ["2003-01-09", 151], ["2003-01-10", 93], ["2003-01-11", 97], ["2003-01-12", 104], ["2003-01-13", 92], ["2003-01-14", 53], ["2003-01-15", 105], ["2003-01-16", 159], ["2003-01-17", 106], ["2003-01-18", 89], ["2003-01-19", 88], ["2003-01-20", 87], ["2003-01-21", 99], ["2003-01-22", 117], ["2003-01-23", 72], ["2003-01-24", 109], ["2003-01-25", 91], ["2003-01-26", 100], ["2003-01-27", 48], ["2003-01-28", 58], ["2003-01-29", 65], ["2003-01-30", 105], ["2003-01-31", 87], ["2003-02-01", 148], ["2003-02-02", 109], ["2003-02-03", 96], ["2003-02-04", 87], ["2003-02-05", 56], ["2003-02-06", 105], ["2003-02-07", 126], ["2003-02-08", 164], ["2003-02-09", 113], ["2003-02-10", 54], ["2003-02-11", 47], ["2003-02-12", 93], ["2003-02-13", 83], ["2003-02-14", 91], ["2003-02-15", 135], ["2003-02-16", 65], ["2003-02-17", 100], ["2003-02-18", 147], ["2003-02-19", 56], ["2003-02-20", 89], ["2003-02-21", 107], ["2003-02-22", 99], ["2003-02-23", 124], ["2003-02-24", 152], ["2003-02-25", 115], ["2003-02-26", 87], ["2003-02-27", 76], ["2003-02-28", 93], ["2003-03-01", 172], ["2003-03-02", 235], ["2003-03-03", 65], ["2003-03-04", 55], ["2003-03-05", 93], ["2003-03-06", 96], ["2003-03-07", 127], ["2003-03-08", 71], ["2003-03-09", 88], ["2003-03-10", 81], ["2003-03-11", 115], ["2003-03-12", 54], ["2003-03-13", 94], ["2003-03-14", 92], ["2003-03-15", 98], ["2003-03-17", 73], ["2003-03-18", 69], ["2003-03-19", 156], ["2003-03-20", 93], ["2003-03-21", 37], ["2003-03-22", 92], ["2003-03-23", 114], ["2003-03-24", 124], ["2003-03-25", 108], ["2003-03-26", 106], ["2003-03-27", 39], ["2003-03-28", 66], ["2003-03-29", 126], ["2003-03-30", 282], ["2003-03-31", 136], ["2003-04-01", 92], ["2003-04-02", 54], ["2003-04-03", 81], ["2003-04-04", 89], ["2003-04-05", 115], ["2003-04-06", 108], ["2003-04-07", 100], ["2003-04-08", 55], ["2003-04-09", 75], ["2003-04-10", 88], ["2003-04-11", 94], ["2003-04-12", 143], ["2003-04-13", 62], ["2003-04-14", 138], ["2003-04-15", 187], ["2003-04-16", 157], ["2003-04-17", 154], ["2003-04-18", 56], ["2003-04-19", 54], ["2003-04-20", 57], ["2003-04-21", 46], ["2003-04-22", 82], ["2003-04-24", 179], ["2003-04-25", 138], ["2003-04-26", 147], ["2003-04-28", 147], ["2003-04-29", 106], ["2003-04-30", 95], ["2003-05-01", 107], ["2003-05-02", 102], ["2003-05-03", 120], ["2003-05-04", 117], ["2003-05-05", 87], ["2003-05-06", 71], ["2003-05-07", 58], ["2003-05-08", 95], ["2003-05-09", 117], ["2003-05-10", 142], ["2003-05-11", 104], ["2003-05-12", 124], ["2003-05-13", 100], ["2003-05-14", 82], ["2003-05-15", 77], ["2003-05-16", 70], ["2003-05-17", 34], ["2003-05-18", 60], ["2003-05-19", 83], ["2003-05-20", 107], ["2003-05-21", 126], ["2003-05-22", 93], ["2003-05-23", 100], ["2003-05-24", 96], ["2003-05-25", 87], ["2003-05-26", 116], ["2003-05-27", 111], ["2003-05-28", 90], ["2003-05-29", 68], ["2003-05-30", 96], ["2003-05-31", 86], ["2003-06-01", 131], ["2003-06-02", 110], ["2003-06-03", 119], ["2003-06-04", 126], ["2003-06-05", 67], ["2003-06-06", 86], ["2003-06-07", 81], ["2003-06-08", 104], ["2003-06-09", 71], ["2003-06-10", 35], ["2003-06-11", 57], ["2003-06-12", 56], ["2003-06-13", 57], ["2003-06-14", 40], ["2003-06-15", 72], ["2003-06-16", 96], ["2003-06-17", 137], ["2003-06-18", 180], ["2003-06-19", 171], ["2003-06-20", 167], ["2003-06-21", 173], ["2003-06-22", 124], ["2003-06-23", 79], ["2003-06-24", 29], ["2003-06-25", 76], ["2003-06-26", 96], ["2003-06-27", 89], ["2003-06-28", 67], ["2003-06-29", 51], ["2003-06-30", 92], ["2003-07-01", 94], ["2003-07-02", 100], ["2003-07-03", 129], ["2003-07-04", 128], ["2003-07-05", 44], ["2003-07-06", 64], ["2003-07-07", 59], ["2003-07-08", 75], ["2003-07-09", 41], ["2003-07-10", 85], ["2003-07-11", 91], ["2003-07-12", 125], ["2003-07-13", 108], ["2003-07-14", 116], ["2003-07-15", 135], ["2003-07-16", 111], ["2003-07-17", 95], ["2003-07-18", 79], ["2003-07-19", 75], ["2003-07-20", 104], ["2003-07-21", 82], ["2003-07-22", 80], ["2003-07-23", 99], ["2003-07-24", 110], ["2003-07-25", 96], ["2003-07-26", 163], ["2003-07-27", 126], ["2003-07-28", 69], ["2003-07-29", 98], ["2003-07-30", 68], ["2003-07-31", 75], ["2003-08-01", 109], ["2003-08-02", 75], ["2003-08-03", 102], ["2003-08-04", 115], ["2003-08-05", 110], ["2003-08-06", 93], ["2003-08-07", 80], ["2003-08-08", 65], ["2003-08-09", 64], ["2003-08-10", 64], ["2003-08-11", 58], ["2003-08-13", 123], ["2003-08-14", 87], ["2003-08-15", 88], ["2003-08-16", 89], ["2003-08-17", 86], ["2003-08-18", 91], ["2003-08-19", 132], ["2003-08-20", 85], ["2003-08-21", 96], ["2003-08-22", 90], ["2003-08-23", 78], ["2003-08-24", 79], ["2003-08-25", 76], ["2003-08-26", 84], ["2003-08-27", 88], ["2003-08-28", 57], ["2003-08-29", 44], ["2003-08-30", 78], ["2003-08-31", 95], ["2003-09-01", 93], ["2003-09-02", 86], ["2003-09-03", 108], ["2003-09-04", 124], ["2003-09-05", 70], ["2003-09-06", 113], ["2003-09-07", 82], ["2003-09-08", 111], ["2003-09-09", 59], ["2003-09-10", 60], ["2003-09-11", 89], ["2003-09-12", 132], ["2003-09-13", 133], ["2003-09-14", 112], ["2003-09-15", 69], ["2003-09-16", 132], ["2003-09-17", 75], ["2003-09-18", 37], ["2003-09-19", 37], ["2003-09-20", 79], ["2003-09-21", 89], ["2003-09-22", 121], ["2003-09-23", 74], ["2003-09-24", 88], ["2003-09-26", 66], ["2003-09-27", 32], ["2003-09-28", 73], ["2003-09-29", 92], ["2003-09-30", 57], ["2003-10-01", 67], ["2003-10-02", 34], ["2003-10-03", 45], ["2003-10-05", 115], ["2003-10-06", 153], ["2003-10-07", 127], ["2003-10-08", 116], ["2003-10-09", 152], ["2003-10-10", 130], ["2003-10-11", 24], ["2003-10-12", 17], ["2003-10-13", 60], ["2003-10-14", 56], ["2003-10-15", 51], ["2003-10-16", 56], ["2003-10-17", 80], ["2003-10-18", 56], ["2003-10-19", 98], ["2003-10-20", 145], ["2003-10-21", 121], ["2003-10-22", 41], ["2003-10-23", 86], ["2003-10-24", 121], ["2003-10-25", 69], ["2003-10-26", 116], ["2003-10-27", 165], ["2003-10-29", 120], ["2003-10-30", 171], ["2003-10-31", 289], ["2003-11-01", 500], ["2003-11-02", 181], ["2003-11-03", 28], ["2003-11-04", 92], ["2003-11-05", 146], ["2003-11-06", 44], ["2003-11-07", 22], ["2003-11-08", 25], ["2003-11-09", 51], ["2003-11-10", 74], ["2003-11-11", 51], ["2003-11-12", 106], ["2003-11-13", 149], ["2003-11-14", 213], ["2003-11-15", 130], ["2003-11-16", 32], ["2003-11-17", 116], ["2003-11-18", 162], ["2003-11-19", 173], ["2003-11-20", 118], ["2003-11-21", 20], ["2003-11-22", 85], ["2003-11-23", 161], ["2003-11-24", 186], ["2003-11-25", 147], ["2003-11-26", 57], ["2003-11-27", 88], ["2003-11-28", 107], ["2003-11-29", 159], ["2003-11-30", 147], ["2003-12-01", 153], ["2003-12-02", 135], ["2003-12-03", 99], ["2003-12-04", 92], ["2003-12-05", 109], ["2003-12-06", 99], ["2003-12-07", 57], ["2003-12-08", 64], ["2003-12-09", 79], ["2003-12-10", 143], ["2003-12-11", 93], ["2003-12-12", 52], ["2003-12-13", 95], ["2003-12-14", 141], ["2003-12-15", 59], ["2003-12-16", 109], ["2003-12-17", 58], ["2003-12-18", 60], ["2003-12-19", 52], ["2003-12-20", 71], ["2003-12-21", 110], ["2003-12-22", 107], ["2003-12-23", 114], ["2003-12-24", 98], ["2003-12-25", 96], ["2003-12-26", 48], ["2003-12-27", 89], ["2003-12-28", 130], ["2003-12-29", 90], ["2003-12-30", 106], ["2003-12-31", 111], ["2004-01-01", 128], ["2004-01-02", 83], ["2004-01-03", 60], ["2004-01-04", 109], ["2004-01-05", 137], ["2004-01-06", 147], ["2004-01-07", 99], ["2004-01-08", 73], ["2004-01-09", 126], ["2004-01-10", 73], ["2004-01-11", 72], ["2004-01-12", 87], ["2004-01-13", 85], ["2004-01-14", 115], ["2004-01-15", 121], ["2004-01-16", 97], ["2004-01-17", 109], ["2004-01-18", 74], ["2004-01-19", 52], ["2004-01-20", 49], ["2004-01-21", 41], ["2004-01-22", 64], ["2004-01-23", 80], ["2004-01-24", 38], ["2004-01-25", 58], ["2004-01-26", 106], ["2004-01-27", 57], ["2004-01-28", 106], ["2004-01-29", 111], ["2004-01-31", 118], ["2004-02-01", 109], ["2004-02-02", 53], ["2004-02-03", 50], ["2004-02-04", 59], ["2004-02-06", 56], ["2004-02-07", 68], ["2004-02-08", 52], ["2004-02-09", 68], ["2004-02-10", 130], ["2004-02-11", 95], ["2004-02-12", 103], ["2004-02-13", 124], ["2004-02-14", 95], ["2004-02-15", 92], ["2004-02-16", 95], ["2004-02-17", 135], ["2004-02-18", 242], ["2004-02-19", 451], ["2004-02-20", 140], ["2004-02-21", 109], ["2004-02-23", 88], ["2004-02-24", 164], ["2004-02-25", 145], ["2004-02-26", 46], ["2004-02-27", 85], ["2004-02-28", 125], ["2004-02-29", 54], ["2004-03-01", 83], ["2004-03-02", 73], ["2004-03-03", 60], ["2004-03-04", 85], ["2004-03-05", 73], ["2004-03-06", 51], ["2004-03-07", 56], ["2004-03-08", 108], ["2004-03-09", 179], ["2004-03-10", 446], ["2004-03-11", 84], ["2004-03-13", 104], ["2004-03-14", 87], ["2004-03-15", 143], ["2004-03-16", 206], ["2004-03-17", 77], ["2004-03-19", 114], ["2004-03-20", 87], ["2004-03-21", 92], ["2004-03-22", 165], ["2004-03-23", 104], ["2004-03-24", 33], ["2004-03-25", 88], ["2004-03-26", 137], ["2004-03-27", 151], ["2004-03-28", 338], ["2004-03-29", 239], ["2004-03-30", 139], ["2004-03-31", 79], ["2004-04-01", 123], ["2004-04-02", 64], ["2004-04-03", 51], ["2004-04-05", 133], ["2004-04-06", 93], ["2004-04-07", 39], ["2004-04-08", 111], ["2004-04-09", 145], ["2004-04-10", 193], ["2004-04-11", 131], ["2004-04-12", 131], ["2004-04-13", 108], ["2004-04-14", 95], ["2004-04-15", 141], ["2004-04-16", 186], ["2004-04-17", 156], ["2004-04-18", 260], ["2004-04-19", 138], ["2004-04-20", 133], ["2004-04-21", 107], ["2004-04-22", 143], ["2004-04-23", 61], ["2004-04-24", 109], ["2004-04-25", 151], ["2004-04-26", 63], ["2004-04-27", 63], ["2004-04-28", 79], ["2004-04-29", 138], ["2004-04-30", 47], ["2004-05-01", 67], ["2004-05-02", 84], ["2004-05-03", 95], ["2004-05-04", 73], ["2004-05-05", 89], ["2004-05-06", 91], ["2004-05-07", 152], ["2004-05-08", 189], ["2004-05-09", 92], ["2004-05-10", 97], ["2004-05-11", 107], ["2004-05-12", 81], ["2004-05-13", 89], ["2004-05-14", 93], ["2004-05-15", 92], ["2004-05-16", 50], ["2004-05-17", 61], ["2004-05-18", 66], ["2004-05-19", 77], ["2004-05-21", 56], ["2004-05-22", 65], ["2004-05-23", 86], ["2004-05-24", 134], ["2004-05-25", 141], ["2004-05-26", 30], ["2004-05-27", 83], ["2004-05-28", 111], ["2004-05-29", 56], ["2004-05-30", 66], ["2004-05-31", 56], ["2004-06-01", 100], ["2004-06-02", 109], ["2004-06-03", 118], ["2004-06-04", 107], ["2004-06-05", 74], ["2004-06-06", 58], ["2004-06-07", 88], ["2004-06-08", 100], ["2004-06-09", 109], ["2004-06-10", 125], ["2004-06-11", 114], ["2004-06-12", 110], ["2004-06-13", 118], ["2004-06-14", 135], ["2004-06-15", 147], ["2004-06-16", 99], ["2004-06-17", 29], ["2004-06-18", 75], ["2004-06-19", 73], ["2004-06-20", 97], ["2004-06-21", 102], ["2004-06-22", 93], ["2004-06-23", 78], ["2004-06-24", 58], ["2004-06-25", 61], ["2004-06-26", 100], ["2004-06-27", 106], ["2004-06-28", 139], ["2004-06-29", 152], ["2004-06-30", 49], ["2004-07-01", 46], ["2004-07-02", 85], ["2004-07-03", 97], ["2004-07-04", 58], ["2004-07-05", 56], ["2004-07-06", 59], ["2004-07-07", 74], ["2004-07-08", 63], ["2004-07-09", 59], ["2004-07-10", 91], ["2004-07-11", 70], ["2004-07-12", 53], ["2004-07-13", 55], ["2004-07-14", 67], ["2004-07-15", 97], ["2004-07-16", 123], ["2004-07-17", 118], ["2004-07-18", 100], ["2004-07-19", 80], ["2004-07-20", 135], ["2004-07-21", 67], ["2004-07-22", 70], ["2004-07-23", 105], ["2004-07-24", 55], ["2004-07-25", 78], ["2004-07-26", 78], ["2004-07-27", 59], ["2004-07-28", 111], ["2004-07-29", 78], ["2004-07-30", 30], ["2004-07-31", 78], ["2004-08-01", 91], ["2004-08-02", 119], ["2004-08-03", 95], ["2004-08-04", 73], ["2004-08-05", 76], ["2004-08-06", 89], ["2004-08-07", 117], ["2004-08-08", 145], ["2004-08-09", 143], ["2004-08-10", 84], ["2004-08-11", 84], ["2004-08-12", 51], ["2004-08-13", 31], ["2004-08-14", 83], ["2004-08-15", 76], ["2004-08-16", 51], ["2004-08-17", 67], ["2004-08-18", 75], ["2004-08-19", 68], ["2004-08-20", 80], ["2004-08-21", 99], ["2004-08-22", 70], ["2004-08-23", 60], ["2004-08-24", 105], ["2004-08-25", 122], ["2004-08-26", 100], ["2004-08-27", 125], ["2004-08-28", 70], ["2004-08-29", 57], ["2004-08-30", 79], ["2004-08-31", 68], ["2004-09-01", 61], ["2004-09-02", 67], ["2004-09-03", 77], ["2004-09-04", 64], ["2004-09-05", 96], ["2004-09-06", 101], ["2004-09-07", 24], ["2004-09-08", 61], ["2004-09-09", 80], ["2004-09-10", 85], ["2004-09-11", 88], ["2004-09-12", 95], ["2004-09-13", 101], ["2004-09-14", 140], ["2004-09-15", 34], ["2004-09-16", 81], ["2004-09-17", 89], ["2004-09-18", 86], ["2004-09-19", 71], ["2004-09-20", 94], ["2004-09-21", 40], ["2004-09-22", 84], ["2004-09-23", 122], ["2004-09-24", 197], ["2004-09-25", 179], ["2004-09-26", 111], ["2004-09-27", 114], ["2004-09-29", 134], ["2004-09-30", 141], ["2004-10-01", 17], ["2004-10-02", 59], ["2004-10-03", 83], ["2004-10-04", 118], ["2004-10-05", 153], ["2004-10-06", 166], ["2004-10-07", 325], ["2004-10-08", 402], ["2004-10-09", 263], ["2004-10-10", 374], ["2004-10-11", 127], ["2004-10-12", 37], ["2004-10-13", 62], ["2004-10-14", 67], ["2004-10-15", 99], ["2004-10-16", 116], ["2004-10-17", 110], ["2004-10-18", 126], ["2004-10-19", 149], ["2004-10-20", 110], ["2004-10-21", 56], ["2004-10-22", 59], ["2004-10-23", 97], ["2004-10-24", 146], ["2004-10-25", 142], ["2004-10-26", 34], ["2004-10-27", 79], ["2004-10-28", 154], ["2004-10-29", 191], ["2004-10-30", 219], ["2004-10-31", 157], ["2004-11-01", 35], ["2004-11-02", 39], ["2004-11-03", 124], ["2004-11-04", 164], ["2004-11-05", 56], ["2004-11-06", 92], ["2004-11-07", 133], ["2004-11-08", 173], ["2004-11-09", 86], ["2004-11-10", 77], ["2004-11-11", 62], ["2004-11-12", 45], ["2004-11-13", 93], ["2004-11-14", 160], ["2004-11-15", 54], ["2004-11-16", 67], ["2004-11-17", 65], ["2004-11-18", 99], ["2004-11-19", 97], ["2004-11-20", 47], ["2004-11-21", 93], ["2004-11-22", 165], ["2004-11-23", 156], ["2004-11-24", 89], ["2004-11-25", 41], ["2004-11-26", 53], ["2004-11-27", 89], ["2004-11-28", 99], ["2004-11-29", 81], ["2004-11-30", 139], ["2004-12-01", 275], ["2004-12-02", 270], ["2004-12-03", 330], ["2004-12-04", 97], ["2004-12-05", 37], ["2004-12-06", 97], ["2004-12-07", 89], ["2004-12-08", 170], ["2004-12-09", 248], ["2004-12-10", 97], ["2004-12-11", 181], ["2004-12-12", 123], ["2004-12-13", 89], ["2004-12-14", 198], ["2004-12-15", 305], ["2004-12-16", 86], ["2004-12-17", 92], ["2004-12-18", 143], ["2004-12-19", 82], ["2004-12-20", 23], ["2004-12-21", 81], ["2004-12-22", 88], ["2004-12-23", 75], ["2004-12-24", 99], ["2004-12-25", 150], ["2004-12-26", 97], ["2004-12-27", 44], ["2004-12-28", 49], ["2004-12-29", 61], ["2004-12-30", 80], ["2004-12-31", 45], ["2005-01-01", 63], ["2005-01-02", 118], ["2005-01-03", 100], ["2005-01-04", 52], ["2005-01-05", 104], ["2005-01-06", 147], ["2005-01-07", 48], ["2005-01-08", 56], ["2005-01-09", 44], ["2005-01-10", 96], ["2005-01-11", 67], ["2005-01-12", 52], ["2005-01-13", 83], ["2005-01-14", 65], ["2005-01-15", 67], ["2005-01-16", 87], ["2005-01-17", 111], ["2005-01-18", 47], ["2005-01-19", 55], ["2005-01-20", 57], ["2005-01-21", 85], ["2005-01-22", 119], ["2005-01-23", 174], ["2005-01-24", 143], ["2005-01-25", 95], ["2005-01-26", 115], ["2005-01-27", 173], ["2005-01-28", 163], ["2005-01-29", 95], ["2005-01-30", 50], ["2005-01-31", 69], ["2005-02-01", 69], ["2005-02-02", 47], ["2005-02-03", 96], ["2005-02-04", 79], ["2005-02-05", 46], ["2005-02-06", 68], ["2005-02-07", 71], ["2005-02-08", 68], ["2005-02-09", 84], ["2005-02-10", 38], ["2005-02-11", 71], ["2005-02-12", 102], ["2005-02-13", 122], ["2005-02-14", 153], ["2005-02-15", 150], ["2005-02-16", 69], ["2005-02-17", 105], ["2005-02-18", 60], ["2005-02-19", 42], ["2005-02-20", 47], ["2005-02-21", 87], ["2005-02-22", 102], ["2005-02-23", 30], ["2005-02-24", 55], ["2005-02-25", 46], ["2005-02-26", 64], ["2005-02-27", 95], ["2005-02-28", 61], ["2005-03-01", 64], ["2005-03-02", 74], ["2005-03-03", 57], ["2005-03-04", 46], ["2005-03-05", 58], ["2005-03-06", 114], ["2005-03-07", 108], ["2005-03-08", 82], ["2005-03-09", 80], ["2005-03-10", 110], ["2005-03-11", 67], ["2005-03-12", 59], ["2005-03-13", 36], ["2005-03-14", 69], ["2005-03-15", 99], ["2005-03-16", 120], ["2005-03-17", 109], ["2005-03-18", 52], ["2005-03-19", 96], ["2005-03-20", 119], ["2005-03-21", 94], ["2005-03-22", 151], ["2005-03-23", 90], ["2005-03-24", 63], ["2005-03-25", 99], ["2005-03-26", 133], ["2005-03-27", 161], ["2005-03-28", 141], ["2005-03-29", 48], ["2005-03-30", 122], ["2005-03-31", 113], ["2005-04-01", 83], ["2005-04-02", 82], ["2005-04-03", 82], ["2005-04-04", 116], ["2005-04-05", 332], ["2005-04-06", 352], ["2005-04-07", 156], ["2005-04-08", 100], ["2005-04-09", 64], ["2005-04-10", 64], ["2005-04-11", 95], ["2005-04-12", 92], ["2005-04-13", 90], ["2005-04-14", 179], ["2005-04-15", 88], ["2005-04-16", 213], ["2005-04-17", 143], ["2005-04-18", 159], ["2005-04-19", 132], ["2005-04-20", 173], ["2005-04-21", 69], ["2005-04-22", 58], ["2005-04-23", 107], ["2005-04-24", 106], ["2005-04-25", 73], ["2005-04-26", 115], ["2005-04-27", 122], ["2005-04-28", 418], ["2005-04-29", 98], ["2005-04-30", 138], ["2005-05-01", 183], ["2005-05-02", 122], ["2005-05-03", 139], ["2005-05-04", 160], ["2005-05-05", 97], ["2005-05-06", 48], ["2005-05-07", 80], ["2005-05-08", 130], ["2005-05-09", 63], ["2005-05-10", 62], ["2005-05-11", 86], ["2005-05-12", 110], ["2005-05-13", 81], ["2005-05-14", 85], ["2005-05-15", 113], ["2005-05-16", 83], ["2005-05-17", 49], ["2005-05-18", 51], ["2005-05-19", 53], ["2005-05-20", 80], ["2005-05-21", 120], ["2005-05-22", 46], ["2005-05-23", 59], ["2005-05-24", 82], ["2005-05-25", 88], ["2005-05-26", 107], ["2005-05-27", 83], ["2005-05-28", 120], ["2005-05-29", 100], ["2005-05-30", 109], ["2005-05-31", 95], ["2005-06-01", 93], ["2005-06-02", 54], ["2005-06-03", 58], ["2005-06-04", 77], ["2005-06-05", 75], ["2005-06-06", 53], ["2005-06-07", 86], ["2005-06-08", 96], ["2005-06-09", 81], ["2005-06-10", 85], ["2005-06-11", 136], ["2005-06-12", 106], ["2005-06-13", 94], ["2005-06-14", 69], ["2005-06-15", 56], ["2005-06-16", 83], ["2005-06-17", 79], ["2005-06-18", 92], ["2005-06-19", 116], ["2005-06-20", 131], ["2005-06-21", 113], ["2005-06-22", 116], ["2005-06-23", 120], ["2005-06-24", 148], ["2005-06-25", 141], ["2005-06-26", 79], ["2005-06-27", 52], ["2005-06-28", 84], ["2005-06-29", 86], ["2005-06-30", 100], ["2005-07-01", 97], ["2005-07-02", 76], ["2005-07-03", 87], ["2005-07-04", 64], ["2005-07-05", 63], ["2005-07-06", 70], ["2005-07-07", 89], ["2005-07-08", 98], ["2005-07-09", 91], ["2005-07-10", 79], ["2005-07-11", 69], ["2005-07-12", 81], ["2005-07-13", 93], ["2005-07-14", 93], ["2005-07-15", 97], ["2005-07-17", 150], ["2005-07-18", 103], ["2005-07-19", 114], ["2005-07-20", 125], ["2005-07-21", 104], ["2005-07-22", 79], ["2005-07-23", 51], ["2005-07-24", 23], ["2005-07-25", 75], ["2005-07-26", 109], ["2005-07-27", 73], ["2005-07-28", 63], ["2005-07-29", 57], ["2005-07-30", 95], ["2005-07-31", 79], ["2005-08-01", 81], ["2005-08-02", 68], ["2005-08-03", 72], ["2005-08-04", 46], ["2005-08-05", 63], ["2005-08-06", 86], ["2005-08-07", 71], ["2005-08-08", 72], ["2005-08-09", 62], ["2005-08-10", 60], ["2005-08-11", 146], ["2005-08-12", 141], ["2005-08-13", 63], ["2005-08-14", 98], ["2005-08-15", 100], ["2005-08-16", 46], ["2005-08-17", 26], ["2005-08-18", 53], ["2005-08-19", 59], ["2005-08-20", 79], ["2005-08-21", 110], ["2005-08-22", 91], ["2005-08-23", 97], ["2005-08-24", 90], ["2005-08-25", 85], ["2005-08-26", 110], ["2005-08-27", 94], ["2005-08-28", 154], ["2005-08-29", 136], ["2005-08-30", 113], ["2005-08-31", 152], ["2005-09-01", 118], ["2005-09-02", 42], ["2005-09-03", 68], ["2005-09-04", 80], ["2005-09-05", 90], ["2005-09-06", 99], ["2005-09-07", 98], ["2005-09-08", 83], ["2005-09-09", 141], ["2005-09-10", 164], ["2005-09-11", 182], ["2005-09-12", 107], ["2005-09-13", 76], ["2005-09-14", 62], ["2005-09-15", 104], ["2005-09-16", 78], ["2005-09-17", 73], ["2005-09-18", 66], ["2005-09-19", 99], ["2005-09-20", 92], ["2005-09-21", 71], ["2005-09-22", 60], ["2005-09-23", 110], ["2005-09-24", 112], ["2005-09-25", 134], ["2005-09-26", 168], ["2005-09-27", 97], ["2005-09-28", 115], ["2005-09-29", 100], ["2005-09-30", 47], ["2005-10-01", 88], ["2005-10-02", 72], ["2005-10-03", 70], ["2005-10-04", 77], ["2005-10-05", 103], ["2005-10-06", 136], ["2005-10-07", 82], ["2005-10-08", 42], ["2005-10-09", 93], ["2005-10-10", 167], ["2005-10-11", 152], ["2005-10-12", 183], ["2005-10-13", 155], ["2005-10-14", 50], ["2005-10-15", 73], ["2005-10-16", 120], ["2005-10-17", 57], ["2005-10-18", 96], ["2005-10-19", 94], ["2005-10-20", 151], ["2005-10-21", 96], ["2005-10-22", 92], ["2005-10-23", 135], ["2005-10-24", 139], ["2005-10-25", 99], ["2005-10-26", 176], ["2005-10-27", 156], ["2005-10-28", 24], ["2005-10-29", 48], ["2005-10-30", 54], ["2005-10-31", 97], ["2005-11-01", 134], ["2005-11-02", 252], ["2005-11-03", 334], ["2005-11-04", 330], ["2005-11-05", 472], ["2005-11-06", 191], ["2005-11-07", 141], ["2005-11-08", 45], ["2005-11-09", 104], ["2005-11-10", 156], ["2005-11-11", 79], ["2005-11-12", 95], ["2005-11-13", 70], ["2005-11-14", 80], ["2005-11-15", 60], ["2005-11-16", 104], ["2005-11-17", 160], ["2005-11-18", 184], ["2005-11-19", 126], ["2005-11-20", 91], ["2005-11-21", 73], ["2005-11-22", 134], ["2005-11-23", 76], ["2005-11-24", 108], ["2005-11-25", 127], ["2005-11-26", 131], ["2005-11-27", 163], ["2005-11-28", 220], ["2005-11-29", 73], ["2005-11-30", 154], ["2005-12-01", 97], ["2005-12-02", 58], ["2005-12-03", 99], ["2005-12-04", 61], ["2005-12-05", 60], ["2005-12-06", 37], ["2005-12-07", 39], ["2005-12-08", 72], ["2005-12-09", 121], ["2005-12-10", 99], ["2005-12-11", 44], ["2005-12-12", 49], ["2005-12-13", 40], ["2005-12-14", 53], ["2005-12-15", 50], ["2005-12-16", 49], ["2005-12-17", 44], ["2005-12-18", 77], ["2005-12-19", 129], ["2005-12-20", 114], ["2005-12-21", 57], ["2005-12-22", 86], ["2005-12-23", 120], ["2005-12-24", 102], ["2005-12-25", 146], ["2005-12-26", 61], ["2005-12-27", 57], ["2005-12-28", 122], ["2005-12-29", 113], ["2005-12-30", 157], ["2005-12-31", 76], ["2006-01-01", 108], ["2006-01-02", 100], ["2006-01-03", 119], ["2006-01-04", 69], ["2006-01-05", 53], ["2006-01-06", 54], ["2006-01-07", 62], ["2006-01-08", 100], ["2006-01-09", 103], ["2006-01-10", 147], ["2006-01-11", 100], ["2006-01-12", 110], ["2006-01-13", 98], ["2006-01-14", 107], ["2006-01-15", 252], ["2006-01-16", 243], ["2006-01-17", 116], ["2006-01-18", 110], ["2006-01-19", 181], ["2006-01-20", 273], ["2006-01-21", 310], ["2006-01-22", 136], ["2006-01-23", 110], ["2006-01-24", 146], ["2006-01-25", 119], ["2006-01-26", 157], ["2006-01-27", 153], ["2006-01-28", 69], ["2006-01-29", 143], ["2006-01-30", 54], ["2006-01-31", 65], ["2006-02-01", 83], ["2006-02-02", 75], ["2006-02-03", 56], ["2006-02-04", 73], ["2006-02-05", 114], ["2006-02-06", 138], ["2006-02-07", 61], ["2006-02-08", 34], ["2006-02-09", 70], ["2006-02-10", 93], ["2006-02-11", 99], ["2006-02-12", 110], ["2006-02-13", 228], ["2006-02-14", 178], ["2006-02-15", 64], ["2006-02-16", 93], ["2006-02-17", 59], ["2006-02-18", 87], ["2006-02-19", 95], ["2006-02-20", 133], ["2006-02-21", 215], ["2006-02-22", 75], ["2006-02-23", 93], ["2006-02-24", 74], ["2006-02-25", 112], ["2006-02-26", 54], ["2006-02-27", 81], ["2006-02-28", 83], ["2006-03-01", 62], ["2006-03-02", 49], ["2006-03-03", 89], ["2006-03-04", 154], ["2006-03-05", 99], ["2006-03-06", 80], ["2006-03-07", 90], ["2006-03-08", 98], ["2006-03-09", 91], ["2006-03-10", 408], ["2006-03-11", 95], ["2006-03-12", 85], ["2006-03-13", 90], ["2006-03-14", 88], ["2006-03-15", 109], ["2006-03-16", 91], ["2006-03-17", 135], ["2006-03-18", 256], ["2006-03-19", 84], ["2006-03-20", 226], ["2006-03-21", 197], ["2006-03-22", 181], ["2006-03-23", 66], ["2006-03-24", 97], ["2006-03-25", 206], ["2006-03-26", 99], ["2006-03-27", 347], ["2006-03-28", 98], ["2006-03-29", 124], ["2006-03-30", 92], ["2006-03-31", 96], ["2006-04-01", 183], ["2006-04-02", 122], ["2006-04-03", 187], ["2006-04-04", 162], ["2006-04-05", 99], ["2006-04-06", 78], ["2006-04-07", 158], ["2006-04-08", 186], ["2006-04-09", 500], ["2006-04-10", 500], ["2006-04-11", 166], ["2006-04-12", 95], ["2006-04-13", 60], ["2006-04-14", 149], ["2006-04-15", 128], ["2006-04-16", 84], ["2006-04-17", 500], ["2006-04-18", 168], ["2006-04-19", 319], ["2006-04-20", 79], ["2006-04-21", 123], ["2006-04-22", 145], ["2006-04-23", 203], ["2006-04-24", 94], ["2006-04-25", 128], ["2006-04-26", 210], ["2006-04-27", 98], ["2006-04-28", 99], ["2006-04-29", 131], ["2006-04-30", 165], ["2006-05-01", 432], ["2006-05-02", 94], ["2006-05-03", 92], ["2006-05-04", 147], ["2006-05-05", 95], ["2006-05-06", 93], ["2006-05-07", 138], ["2006-05-08", 123], ["2006-05-09", 79], ["2006-05-10", 71], ["2006-05-11", 61], ["2006-05-12", 63], ["2006-05-13", 44], ["2006-05-14", 93], ["2006-05-15", 95], ["2006-05-16", 98], ["2006-05-17", 500], ["2006-05-18", 168], ["2006-05-19", 240], ["2006-05-20", 82], ["2006-05-21", 96], ["2006-05-22", 96], ["2006-05-23", 95], ["2006-05-24", 84], ["2006-05-25", 91], ["2006-05-26", 78], ["2006-05-27", 32], ["2006-05-28", 51], ["2006-05-29", 84], ["2006-05-30", 98], ["2006-05-31", 118], ["2006-06-01", 96], ["2006-06-02", 112], ["2006-06-03", 69], ["2006-06-04", 100], ["2006-06-05", 137], ["2006-06-06", 147], ["2006-06-07", 86], ["2006-06-08", 65], ["2006-06-09", 92], ["2006-06-10", 39], ["2006-06-11", 61], ["2006-06-12", 96], ["2006-06-13", 77], ["2006-06-14", 43], ["2006-06-15", 78], ["2006-06-16", 86], ["2006-06-17", 50], ["2006-06-18", 68], ["2006-06-19", 97], ["2006-06-20", 84], ["2006-06-21", 152], ["2006-06-22", 118], ["2006-06-23", 123], ["2006-06-24", 76], ["2006-06-25", 68], ["2006-06-26", 84], ["2006-06-27", 75], ["2006-06-28", 90], ["2006-06-29", 66], ["2006-06-30", 42], ["2006-07-01", 57], ["2006-07-02", 52], ["2006-07-03", 81], ["2006-07-04", 75], ["2006-07-05", 97], ["2006-07-06", 60], ["2006-07-07", 65], ["2006-07-08", 67], ["2006-07-09", 82], ["2006-07-10", 99], ["2006-07-11", 66], ["2006-07-12", 72], ["2006-07-13", 44], ["2006-07-14", 78], ["2006-07-15", 70], ["2006-07-16", 69], ["2006-07-17", 58], ["2006-07-18", 43], ["2006-07-19", 55], ["2006-07-20", 74], ["2006-07-21", 76], ["2006-07-22", 36], ["2006-07-23", 72], ["2006-07-24", 61], ["2006-07-25", 46], ["2006-07-26", 50], ["2006-07-27", 65], ["2006-07-28", 98], ["2006-07-29", 115], ["2006-07-30", 138], ["2006-07-31", 88], ["2006-08-01", 47], ["2006-08-02", 39], ["2006-08-03", 61], ["2006-08-04", 64], ["2006-08-05", 74], ["2006-08-06", 100], ["2006-08-07", 82], ["2006-08-08", 84], ["2006-08-09", 64], ["2006-08-10", 89], ["2006-08-11", 75], ["2006-08-12", 98], ["2006-08-13", 69], ["2006-08-14", 27], ["2006-08-15", 70], ["2006-08-16", 84], ["2006-08-17", 91], ["2006-08-18", 85], ["2006-08-19", 97], ["2006-08-20", 77], ["2006-08-21", 45], ["2006-08-22", 69], ["2006-08-23", 67], ["2006-08-24", 99], ["2006-08-25", 131], ["2006-08-26", 69], ["2006-08-27", 66], ["2006-08-28", 93], ["2006-08-29", 62], ["2006-08-30", 59], ["2006-08-31", 64], ["2006-09-01", 89], ["2006-09-02", 100], ["2006-09-03", 109], ["2006-09-04", 28], ["2006-09-05", 71], ["2006-09-06", 87], ["2006-09-07", 112], ["2006-09-08", 71], ["2006-09-09", 37], ["2006-09-10", 67], ["2006-09-11", 86], ["2006-09-12", 89], ["2006-09-13", 100], ["2006-09-14", 107], ["2006-09-15", 109], ["2006-09-16", 116], ["2006-09-17", 134], ["2006-09-18", 100], ["2006-09-19", 132], ["2006-09-20", 151], ["2006-09-21", 99], ["2006-09-22", 95], ["2006-09-23", 118], ["2006-09-24", 121], ["2006-09-25", 119], ["2006-09-26", 49], ["2006-09-27", 91], ["2006-09-28", 98], ["2006-09-29", 70], ["2006-09-30", 100], ["2006-10-01", 139], ["2006-10-02", 152], ["2006-10-03", 143], ["2006-10-04", 73], ["2006-10-05", 99], ["2006-10-06", 194], ["2006-10-07", 100], ["2006-10-08", 91], ["2006-10-09", 77], ["2006-10-10", 131], ["2006-10-11", 65], ["2006-10-12", 73], ["2006-10-13", 121], ["2006-10-14", 135], ["2006-10-15", 100], ["2006-10-16", 146], ["2006-10-17", 59], ["2006-10-18", 86], ["2006-10-19", 121], ["2006-10-20", 71], ["2006-10-21", 97], ["2006-10-22", 87], ["2006-10-23", 46], ["2006-10-24", 99], ["2006-10-25", 139], ["2006-10-26", 64], ["2006-10-27", 99], ["2006-10-28", 163], ["2006-10-29", 77], ["2006-10-30", 130], ["2006-10-31", 154], ["2006-11-01", 96], ["2006-11-02", 98], ["2006-11-03", 140], ["2006-11-04", 180], ["2006-11-05", 48], ["2006-11-06", 93], ["2006-11-07", 94], ["2006-11-08", 148], ["2006-11-09", 61], ["2006-11-10", 89], ["2006-11-11", 72], ["2006-11-12", 136], ["2006-11-13", 98], ["2006-11-14", 37], ["2006-11-15", 78], ["2006-11-16", 99], ["2006-11-17", 100], ["2006-11-18", 129], ["2006-11-19", 147], ["2006-11-20", 249], ["2006-11-21", 414], ["2006-11-22", 97], ["2006-11-23", 74], ["2006-11-24", 153], ["2006-11-25", 124], ["2006-11-26", 129], ["2006-11-27", 47], ["2006-11-28", 58], ["2006-11-29", 61], ["2006-11-30", 96], ["2006-12-01", 88], ["2006-12-02", 49], ["2006-12-03", 66], ["2006-12-04", 111], ["2006-12-05", 94], ["2006-12-06", 78], ["2006-12-07", 86], ["2006-12-08", 97], ["2006-12-09", 81], ["2006-12-10", 105], ["2006-12-11", 256], ["2006-12-12", 500], ["2006-12-13", 88], ["2006-12-14", 118], ["2006-12-15", 94], ["2006-12-16", 76], ["2006-12-17", 52], ["2006-12-18", 100], ["2006-12-19", 140], ["2006-12-20", 180], ["2006-12-21", 180], ["2006-12-22", 88], ["2006-12-23", 95], ["2006-12-24", 85], ["2006-12-25", 136], ["2006-12-26", 160], ["2006-12-27", 80], ["2006-12-28", 47], ["2006-12-29", 90], ["2006-12-30", 157], ["2006-12-31", 139], ["2007-01-01", 158], ["2007-01-02", 150], ["2007-01-03", 133], ["2007-01-04", 170], ["2007-01-05", 322], ["2007-01-06", 73], ["2007-01-07", 47], ["2007-01-08", 45], ["2007-01-09", 86], ["2007-01-10", 98], ["2007-01-11", 75], ["2007-01-12", 56], ["2007-01-13", 74], ["2007-01-14", 102], ["2007-01-15", 170], ["2007-01-16", 64], ["2007-01-17", 84], ["2007-01-18", 52], ["2007-01-19", 93], ["2007-01-20", 147], ["2007-01-21", 98], ["2007-01-22", 58], ["2007-01-23", 96], ["2007-01-24", 118], ["2007-01-25", 140], ["2007-01-26", 68], ["2007-01-27", 55], ["2007-01-28", 55], ["2007-01-29", 114], ["2007-01-30", 85], ["2007-01-31", 76], ["2007-02-01", 50], ["2007-02-02", 100], ["2007-02-03", 115], ["2007-02-04", 93], ["2007-02-05", 175], ["2007-02-06", 67], ["2007-02-07", 110], ["2007-02-08", 99], ["2007-02-09", 67], ["2007-02-10", 61], ["2007-02-11", 55], ["2007-02-12", 103], ["2007-02-13", 181], ["2007-02-14", 74], ["2007-02-15", 75], ["2007-02-16", 97], ["2007-02-17", 98], ["2007-02-18", 115], ["2007-02-19", 99], ["2007-02-20", 160], ["2007-02-21", 200], ["2007-02-22", 173], ["2007-02-23", 78], ["2007-02-24", 75], ["2007-02-25", 123], ["2007-02-26", 169], ["2007-02-27", 172], ["2007-02-28", 108], ["2007-03-01", 98], ["2007-03-02", 85], ["2007-03-03", 87], ["2007-03-04", 28], ["2007-03-05", 34], ["2007-03-06", 35], ["2007-03-07", 51], ["2007-03-08", 54], ["2007-03-09", 105], ["2007-03-10", 75], ["2007-03-11", 34], ["2007-03-12", 68], ["2007-03-13", 133], ["2007-03-14", 157], ["2007-03-15", 106], ["2007-03-16", 78], ["2007-03-17", 100], ["2007-03-18", 121], ["2007-03-19", 119], ["2007-03-21", 138], ["2007-03-22", 145], ["2007-03-23", 202], ["2007-03-24", 192], ["2007-03-25", 79], ["2007-03-26", 78], ["2007-03-27", 84], ["2007-03-28", 98], ["2007-03-29", 99], ["2007-03-30", 66], ["2007-03-31", 103], ["2007-04-01", 63], ["2007-04-02", 48], ["2007-04-03", 40], ["2007-04-04", 95], ["2007-04-05", 110], ["2007-04-06", 148], ["2007-04-07", 46], ["2007-04-08", 43], ["2007-04-09", 96], ["2007-04-10", 133], ["2007-04-11", 88], ["2007-04-12", 107], ["2007-04-13", 55], ["2007-04-14", 74], ["2007-04-15", 72], ["2007-04-16", 81], ["2007-04-17", 74], ["2007-04-18", 100], ["2007-04-19", 173], ["2007-04-20", 155], ["2007-04-21", 62], ["2007-04-22", 58], ["2007-04-23", 81], ["2007-04-24", 78], ["2007-04-25", 72], ["2007-04-26", 90], ["2007-04-27", 113], ["2007-04-28", 115], ["2007-04-29", 190], ["2007-04-30", 151], ["2007-05-01", 61], ["2007-05-02", 87], ["2007-05-03", 96], ["2007-05-04", 97], ["2007-05-05", 123], ["2007-05-06", 91], ["2007-05-07", 139], ["2007-05-08", 147], ["2007-05-09", 98], ["2007-05-10", 116], ["2007-05-11", 116], ["2007-05-12", 99], ["2007-05-13", 100], ["2007-05-14", 72], ["2007-05-15", 97], ["2007-05-16", 100], ["2007-05-17", 84], ["2007-05-18", 58], ["2007-05-19", 60], ["2007-05-20", 98], ["2007-05-21", 82], ["2007-05-22", 116], ["2007-05-23", 60], ["2007-05-24", 169], ["2007-05-25", 250], ["2007-05-26", 98], ["2007-05-27", 118], ["2007-05-28", 96], ["2007-05-29", 98], ["2007-05-30", 126], ["2007-05-31", 119], ["2007-06-01", 75], ["2007-06-02", 107], ["2007-06-03", 99], ["2007-06-04", 84], ["2007-06-05", 99], ["2007-06-06", 146], ["2007-06-07", 195], ["2007-06-08", 194], ["2007-06-09", 134], ["2007-06-10", 97], ["2007-06-11", 136], ["2007-06-12", 168], ["2007-06-13", 142], ["2007-06-14", 52], ["2007-06-15", 91], ["2007-06-16", 98], ["2007-06-17", 123], ["2007-06-18", 138], ["2007-06-19", 202], ["2007-06-20", 151], ["2007-06-21", 123], ["2007-06-22", 85], ["2007-06-23", 121], ["2007-06-24", 97], ["2007-06-25", 72], ["2007-06-26", 98], ["2007-06-27", 135], ["2007-06-28", 52], ["2007-06-29", 95], ["2007-06-30", 87], ["2007-07-01", 28], ["2007-07-02", 77], ["2007-07-03", 99], ["2007-07-04", 82], ["2007-07-06", 145], ["2007-07-07", 80], ["2007-07-08", 75], ["2007-07-09", 115], ["2007-07-10", 58], ["2007-07-11", 65], ["2007-07-12", 78], ["2007-07-13", 74], ["2007-07-14", 83], ["2007-07-15", 93], ["2007-07-16", 96], ["2007-07-17", 169], ["2007-07-18", 98], ["2007-07-19", 47], ["2007-07-20", 76], ["2007-07-21", 98], ["2007-07-22", 99], ["2007-07-23", 117], ["2007-07-24", 99], ["2007-07-25", 119], ["2007-07-26", 151], ["2007-07-27", 150], ["2007-07-28", 98], ["2007-07-29", 80], ["2007-07-30", 138], ["2007-07-31", 26], ["2007-08-01", 52], ["2007-08-02", 42], ["2007-08-03", 70], ["2007-08-04", 85], ["2007-08-05", 98], ["2007-08-06", 107], ["2007-08-07", 93], ["2007-08-08", 88], ["2007-08-09", 86], ["2007-08-10", 79], ["2007-08-11", 74], ["2007-08-12", 66], ["2007-08-13", 56], ["2007-08-14", 76], ["2007-08-15", 86], ["2007-08-16", 115], ["2007-08-17", 91], ["2007-08-18", 93], ["2007-08-19", 95], ["2007-08-20", 95], ["2007-08-21", 116], ["2007-08-22", 88], ["2007-08-23", 77], ["2007-08-24", 83], ["2007-08-25", 95], ["2007-08-26", 78], ["2007-08-27", 49], ["2007-08-28", 78], ["2007-08-29", 64], ["2007-08-30", 75], ["2007-08-31", 98], ["2007-09-01", 108], ["2007-09-02", 95], ["2007-09-03", 73], ["2007-09-04", 77], ["2007-09-05", 94], ["2007-09-06", 100], ["2007-09-07", 98], ["2007-09-08", 94], ["2007-09-09", 98], ["2007-09-10", 142], ["2007-09-11", 171], ["2007-09-12", 133], ["2007-09-13", 97], ["2007-09-14", 58], ["2007-09-15", 66], ["2007-09-16", 99], ["2007-09-17", 138], ["2007-09-18", 60], ["2007-09-19", 24], ["2007-09-20", 62], ["2007-09-21", 79], ["2007-09-22", 99], ["2007-09-23", 97], ["2007-09-24", 98], ["2007-09-25", 95], ["2007-09-26", 80], ["2007-09-27", 40], ["2007-09-28", 63], ["2007-09-29", 80], ["2007-09-30", 64], ["2007-10-01", 75], ["2007-10-02", 52], ["2007-10-03", 78], ["2007-10-04", 94], ["2007-10-05", 34], ["2007-10-06", 48], ["2007-10-07", 28], ["2007-10-08", 22], ["2007-10-09", 44], ["2007-10-10", 69], ["2007-10-11", 88], ["2007-10-12", 119], ["2007-10-13", 95], ["2007-10-14", 35], ["2007-10-15", 53], ["2007-10-16", 66], ["2007-10-17", 95], ["2007-10-18", 82], ["2007-10-19", 49], ["2007-10-20", 60], ["2007-10-21", 98], ["2007-10-22", 100], ["2007-10-23", 89], ["2007-10-24", 96], ["2007-10-25", 143], ["2007-10-26", 184], ["2007-10-27", 179], ["2007-10-28", 27], ["2007-10-29", 47], ["2007-10-30", 121], ["2007-10-31", 95], ["2007-11-01", 18], ["2007-11-02", 83], ["2007-11-03", 57], ["2007-11-04", 76], ["2007-11-05", 119], ["2007-11-06", 172], ["2007-11-07", 253], ["2007-11-08", 186], ["2007-11-09", 35], ["2007-11-10", 54], ["2007-11-11", 90], ["2007-11-12", 159], ["2007-11-13", 153], ["2007-11-14", 58], ["2007-11-15", 31], ["2007-11-16", 76], ["2007-11-17", 112], ["2007-11-18", 45], ["2007-11-19", 88], ["2007-11-20", 83], ["2007-11-21", 98], ["2007-11-22", 144], ["2007-11-23", 119], ["2007-11-24", 117], ["2007-11-25", 269], ["2007-11-26", 55], ["2007-11-27", 85], ["2007-11-28", 100], ["2007-11-29", 81], ["2007-11-30", 78], ["2007-12-01", 136], ["2007-12-02", 96], ["2007-12-03", 71], ["2007-12-04", 68], ["2007-12-05", 88], ["2007-12-06", 129], ["2007-12-07", 54], ["2007-12-08", 77], ["2007-12-09", 118], ["2007-12-11", 110], ["2007-12-12", 46], ["2007-12-13", 56], ["2007-12-14", 91], ["2007-12-15", 59], ["2007-12-16", 89], ["2007-12-17", 78], ["2007-12-18", 104], ["2007-12-19", 155], ["2007-12-20", 153], ["2007-12-21", 114], ["2007-12-22", 166], ["2007-12-23", 98], ["2007-12-24", 124], ["2007-12-25", 280], ["2007-12-26", 269], ["2007-12-27", 421], ["2007-12-28", 500], ["2007-12-29", 156], ["2007-12-30", 72], ["2007-12-31", 58], ["2008-01-01", 32], ["2008-01-02", 57], ["2008-01-03", 75], ["2008-01-04", 90], ["2008-01-05", 147], ["2008-01-06", 146], ["2008-01-07", 115], ["2008-01-08", 121], ["2008-01-09", 94], ["2008-01-10", 95], ["2008-01-11", 113], ["2008-01-12", 46], ["2008-01-13", 39], ["2008-01-14", 87], ["2008-01-15", 119], ["2008-01-16", 72], ["2008-01-17", 80], ["2008-01-18", 122], ["2008-01-19", 149], ["2008-01-20", 134], ["2008-01-21", 66], ["2008-01-22", 79], ["2008-01-23", 51], ["2008-01-24", 50], ["2008-01-25", 54], ["2008-01-26", 67], ["2008-01-27", 70], ["2008-01-28", 77], ["2008-01-29", 48], ["2008-01-30", 44], ["2008-01-31", 45], ["2008-02-01", 57], ["2008-02-02", 64], ["2008-02-03", 52], ["2008-02-04", 65], ["2008-02-05", 83], ["2008-02-06", 35], ["2008-02-08", 37], ["2008-02-09", 38], ["2008-02-10", 64], ["2008-02-11", 61], ["2008-02-12", 64], ["2008-02-13", 55], ["2008-02-14", 55], ["2008-02-15", 68], ["2008-02-16", 69], ["2008-02-17", 70], ["2008-02-18", 72], ["2008-02-19", 111], ["2008-02-20", 88], ["2008-02-21", 152], ["2008-02-22", 160], ["2008-02-23", 85], ["2008-02-25", 65], ["2008-02-26", 78], ["2008-02-27", 75], ["2008-02-28", 84], ["2008-02-29", 82], ["2008-03-01", 82], ["2008-03-02", 126], ["2008-03-03", 46], ["2008-03-04", 55], ["2008-03-05", 86], ["2008-03-06", 80], ["2008-03-08", 129], ["2008-03-09", 158], ["2008-03-10", 238], ["2008-03-11", 174], ["2008-03-12", 128], ["2008-03-13", 99], ["2008-03-14", 82], ["2008-03-15", 110], ["2008-03-16", 72], ["2008-03-17", 126], ["2008-03-18", 304], ["2008-03-19", 286], ["2008-03-20", 147], ["2008-03-21", 98], ["2008-03-22", 120], ["2008-03-23", 69], ["2008-03-24", 76], ["2008-03-25", 52], ["2008-03-26", 46], ["2008-03-27", 55], ["2008-03-28", 74], ["2008-03-29", 59], ["2008-03-30", 81], ["2008-03-31", 53], ["2008-04-01", 90], ["2008-04-02", 63], ["2008-04-03", 55], ["2008-04-04", 88], ["2008-04-05", 145], ["2008-04-06", 161], ["2008-04-07", 131], ["2008-04-08", 177], ["2008-04-09", 93], ["2008-04-10", 94], ["2008-04-11", 65], ["2008-04-12", 79], ["2008-04-13", 71], ["2008-04-14", 98], ["2008-04-15", 129], ["2008-04-16", 173], ["2008-04-17", 159], ["2008-04-18", 139], ["2008-04-19", 138], ["2008-04-20", 97], ["2008-04-21", 19], ["2008-04-22", 32], ["2008-04-23", 43], ["2008-04-24", 76], ["2008-04-25", 100], ["2008-04-26", 72], ["2008-04-27", 79], ["2008-04-28", 94], ["2008-04-29", 176], ["2008-04-30", 155], ["2008-05-01", 140], ["2008-05-02", 144], ["2008-05-03", 185], ["2008-05-04", 32], ["2008-05-05", 81], ["2008-05-06", 134], ["2008-05-07", 138], ["2008-05-08", 95], ["2008-05-09", 89], ["2008-05-10", 62], ["2008-05-11", 54], ["2008-05-12", 24], ["2008-05-13", 57], ["2008-05-14", 87], ["2008-05-15", 77], ["2008-05-16", 107], ["2008-05-17", 117], ["2008-05-18", 91], ["2008-05-19", 83], ["2008-05-20", 112], ["2008-05-21", 408], ["2008-05-22", 153], ["2008-05-23", 186], ["2008-05-24", 161], ["2008-05-25", 121], ["2008-05-26", 138], ["2008-05-27", 463], ["2008-05-28", 253], ["2008-05-29", 395], ["2008-05-30", 95], ["2008-05-31", 115], ["2008-06-01", 92], ["2008-06-02", 50], ["2008-06-03", 74], ["2008-06-05", 78], ["2008-06-06", 94], ["2008-06-07", 81], ["2008-06-08", 126], ["2008-06-09", 97], ["2008-06-10", 100], ["2008-06-11", 80], ["2008-06-12", 89], ["2008-06-13", 105], ["2008-06-14", 96], ["2008-06-15", 93], ["2008-06-16", 84], ["2008-06-17", 55], ["2008-06-18", 61], ["2008-06-19", 120], ["2008-06-20", 165], ["2008-06-21", 81], ["2008-06-22", 125], ["2008-06-23", 81], ["2008-06-24", 75], ["2008-06-25", 109], ["2008-06-26", 87], ["2008-06-27", 88], ["2008-06-28", 89], ["2008-06-30", 98], ["2008-07-01", 72], ["2008-07-02", 61], ["2008-07-03", 92], ["2008-07-04", 100], ["2008-07-05", 66], ["2008-07-06", 39], ["2008-07-07", 69], ["2008-07-08", 98], ["2008-07-09", 62], ["2008-07-10", 85], ["2008-07-11", 112], ["2008-07-12", 74], ["2008-07-13", 59], ["2008-07-14", 84], ["2008-07-15", 31], ["2008-07-16", 66], ["2008-07-17", 77], ["2008-07-18", 66], ["2008-07-19", 64], ["2008-07-20", 55], ["2008-07-21", 64], ["2008-07-22", 66], ["2008-07-23", 89], ["2008-07-24", 113], ["2008-07-25", 109], ["2008-07-26", 118], ["2008-07-27", 113], ["2008-07-28", 96], ["2008-07-29", 90], ["2008-07-30", 43], ["2008-07-31", 69], ["2008-08-01", 27], ["2008-08-02", 34], ["2008-08-03", 35], ["2008-08-04", 83], ["2008-08-05", 88], ["2008-08-06", 85], ["2008-08-07", 95], ["2008-08-08", 94], ["2008-08-09", 78], ["2008-08-10", 82], ["2008-08-11", 37], ["2008-08-12", 32], ["2008-08-13", 60], ["2008-08-14", 61], ["2008-08-15", 17], ["2008-08-16", 23], ["2008-08-16", 84], ["2008-08-17", 42], ["2008-08-18", 25], ["2008-08-19", 42], ["2008-08-20", 53], ["2008-08-21", 60], ["2008-08-22", 36], ["2008-08-23", 41], ["2008-08-24", 45], ["2008-08-25", 67], ["2008-08-26", 64], ["2008-08-27", 56], ["2008-08-28", 79], ["2008-08-29", 110], ["2008-08-30", 64], ["2008-08-31", 24], ["2008-09-01", 25], ["2008-09-02", 37], ["2008-09-03", 72], ["2008-09-04", 57], ["2008-09-05", 58], ["2008-09-06", 59], ["2008-09-07", 86], ["2008-09-08", 49], ["2008-09-09", 64], ["2008-09-10", 51], ["2008-09-11", 46], ["2008-09-12", 58], ["2008-09-13", 57], ["2008-09-14", 56], ["2008-09-15", 58], ["2008-09-16", 63], ["2008-09-17", 62], ["2008-09-19", 66], ["2008-09-20", 59], ["2008-09-21", 88], ["2008-09-22", 59], ["2008-09-23", 12], ["2008-09-24", 26], ["2008-09-25", 30], ["2008-09-26", 17], ["2008-09-28", 71], ["2008-09-29", 83], ["2008-09-30", 106], ["2008-10-01", 104], ["2008-10-02", 126], ["2008-10-03", 108], ["2008-10-04", 63], ["2008-10-05", 49], ["2008-10-06", 25], ["2008-10-07", 58], ["2008-10-08", 75], ["2008-10-09", 47], ["2008-10-10", 58], ["2008-10-11", 44], ["2008-10-12", 59], ["2008-10-13", 92], ["2008-10-14", 114], ["2008-10-15", 85], ["2008-10-16", 61], ["2008-10-17", 93], ["2008-10-18", 174], ["2008-10-19", 86], ["2008-10-20", 86], ["2008-10-21", 134], ["2008-10-22", 111], ["2008-10-23", 43], ["2008-10-24", 14], ["2008-10-25", 58], ["2008-10-26", 32], ["2008-10-27", 32], ["2008-10-28", 67], ["2008-10-29", 80], ["2008-10-30", 58], ["2008-10-31", 79], ["2008-11-01", 71], ["2008-11-02", 60], ["2008-11-03", 54], ["2008-11-04", 68], ["2008-11-05", 109], ["2008-11-06", 97], ["2008-11-07", 55], ["2008-11-08", 65], ["2008-11-09", 86], ["2008-11-10", 94], ["2008-11-11", 131], ["2008-11-12", 186], ["2008-11-13", 161], ["2008-11-14", 34], ["2008-11-15", 120], ["2008-11-16", 54], ["2008-11-17", 46], ["2008-11-18", 28], ["2008-11-19", 40], ["2008-11-20", 103], ["2008-11-21", 52], ["2008-11-22", 91], ["2008-11-23", 95], ["2008-11-24", 97], ["2008-11-25", 59], ["2008-11-26", 89], ["2008-11-27", 40], ["2008-11-28", 77], ["2008-11-29", 53], ["2008-11-30", 84], ["2008-12-01", 146], ["2008-12-02", 87], ["2008-12-03", 144], ["2008-12-04", 51], ["2008-12-05", 59], ["2008-12-06", 51], ["2008-12-07", 112], ["2008-12-08", 169], ["2008-12-09", 246], ["2008-12-10", 162], ["2008-12-11", 96], ["2008-12-12", 154], ["2008-12-13", 57], ["2008-12-14", 86], ["2008-12-15", 109], ["2008-12-16", 135], ["2008-12-17", 134], ["2008-12-18", 46], ["2008-12-19", 98], ["2008-12-20", 45], ["2008-12-21", 67], ["2008-12-22", 49], ["2008-12-23", 89], ["2008-12-24", 115], ["2008-12-25", 55], ["2008-12-26", 66], ["2008-12-27", 129], ["2008-12-28", 134], ["2008-12-29", 69], ["2008-12-30", 36], ["2008-12-31", 29], ["2009-01-01", 42], ["2009-01-02", 79], ["2009-01-03", 90], ["2009-01-04", 69], ["2009-01-05", 64], ["2009-01-06", 71], ["2009-01-07", 56], ["2009-01-08", 100], ["2009-01-09", 32], ["2009-01-10", 54], ["2009-01-11", 51], ["2009-01-12", 36], ["2009-01-13", 59], ["2009-01-14", 43], ["2009-01-15", 72], ["2009-01-16", 90], ["2009-01-17", 74], ["2009-01-18", 97], ["2009-01-19", 76], ["2009-01-20", 137], ["2009-01-21", 109], ["2009-01-22", 117], ["2009-01-23", 97], ["2009-01-24", 67], ["2009-01-25", 48], ["2009-01-26", 88], ["2009-01-27", 95], ["2009-01-28", 129], ["2009-01-29", 135], ["2009-01-30", 131], ["2009-01-31", 133], ["2009-02-01", 91], ["2009-02-02", 107], ["2009-02-03", 87], ["2009-02-04", 80], ["2009-02-05", 98], ["2009-02-06", 78], ["2009-02-07", 90], ["2009-02-08", 71], ["2009-02-09", 112], ["2009-02-10", 307], ["2009-02-11", 89], ["2009-02-12", 139], ["2009-02-13", 82], ["2009-02-14", 72], ["2009-02-15", 53], ["2009-02-16", 55], ["2009-02-17", 56], ["2009-02-19", 64], ["2009-02-20", 99], ["2009-02-21", 86], ["2009-02-22", 80], ["2009-02-23", 59], ["2009-02-24", 84], ["2009-02-25", 36], ["2009-02-26", 68], ["2009-02-27", 96], ["2009-02-28", 67], ["2009-03-01", 93], ["2009-03-02", 59], ["2009-03-03", 98], ["2009-03-04", 161], ["2009-03-05", 96], ["2009-03-06", 19], ["2009-03-07", 73], ["2009-03-08", 119], ["2009-03-09", 64], ["2009-03-10", 74], ["2009-03-11", 85], ["2009-03-12", 88], ["2009-03-13", 99], ["2009-03-14", 81], ["2009-03-15", 119], ["2009-03-16", 100], ["2009-03-17", 169], ["2009-03-18", 268], ["2009-03-19", 195], ["2009-03-20", 80], ["2009-03-21", 82], ["2009-03-22", 77], ["2009-03-23", 64], ["2009-03-24", 59], ["2009-03-25", 44], ["2009-03-26", 58], ["2009-03-27", 79], ["2009-03-28", 69], ["2009-03-29", 69], ["2009-03-30", 71], ["2009-03-31", 51], ["2009-04-01", 27], ["2009-04-02", 72], ["2009-04-03", 91], ["2009-04-04", 96], ["2009-04-05", 72], ["2009-04-06", 53], ["2009-04-07", 94], ["2009-04-08", 140], ["2009-04-09", 117], ["2009-04-10", 115], ["2009-04-11", 113], ["2009-04-12", 122], ["2009-04-13", 148], ["2009-04-14", 75], ["2009-04-15", 81], ["2009-04-16", 69], ["2009-04-17", 84], ["2009-04-18", 116], ["2009-04-19", 97], ["2009-04-20", 63], ["2009-04-21", 34], ["2009-04-22", 59], ["2009-04-23", 70], ["2009-04-24", 77], ["2009-04-25", 54], ["2009-04-26", 34], ["2009-04-27", 57], ["2009-04-28", 78], ["2009-04-29", 73], ["2009-04-30", 95], ["2009-05-01", 95], ["2009-05-02", 54], ["2009-05-03", 82], ["2009-05-04", 96], ["2009-05-05", 106], ["2009-05-06", 100], ["2009-05-07", 109], ["2009-05-08", 125], ["2009-05-09", 106], ["2009-05-10", 57], ["2009-05-11", 72], ["2009-05-12", 75], ["2009-05-13", 63], ["2009-05-14", 91], ["2009-05-15", 64], ["2009-05-16", 81], ["2009-05-17", 78], ["2009-05-18", 90], ["2009-05-19", 97], ["2009-05-20", 98], ["2009-05-21", 85], ["2009-05-22", 27], ["2009-05-23", 65], ["2009-05-24", 95], ["2009-05-25", 128], ["2009-05-26", 81], ["2009-05-27", 105], ["2009-05-28", 94], ["2009-05-29", 59], ["2009-05-30", 45], ["2009-05-31", 56], ["2009-06-01", 79], ["2009-06-02", 55], ["2009-06-03", 61], ["2009-06-04", 71], ["2009-06-05", 68], ["2009-06-06", 67], ["2009-06-07", 63], ["2009-06-08", 77], ["2009-06-09", 34], ["2009-06-10", 21], ["2009-06-11", 66], ["2009-06-12", 60], ["2009-06-13", 58], ["2009-06-14", 61], ["2009-06-15", 70], ["2009-06-16", 89], ["2009-06-17", 75], ["2009-06-18", 104], ["2009-06-20", 165], ["2009-06-21", 98], ["2009-06-22", 42], ["2009-06-23", 60], ["2009-06-24", 67], ["2009-06-25", 81], ["2009-06-26", 104], ["2009-06-27", 116], ["2009-06-28", 96], ["2009-06-29", 90], ["2009-06-30", 48], ["2009-07-01", 30], ["2009-07-02", 51], ["2009-07-03", 73], ["2009-07-04", 103], ["2009-07-05", 110], ["2009-07-06", 70], ["2009-07-07", 93], ["2009-07-08", 85], ["2009-07-09", 48], ["2009-07-10", 79], ["2009-07-11", 94], ["2009-07-12", 72], ["2009-07-13", 104], ["2009-07-14", 57], ["2009-07-15", 71], ["2009-07-16", 100], ["2009-07-17", 60], ["2009-07-18", 45], ["2009-07-19", 74], ["2009-07-20", 69], ["2009-07-21", 60], ["2009-07-22", 101], ["2009-07-23", 64], ["2009-07-24", 36], ["2009-07-25", 29], ["2009-07-26", 59], ["2009-07-27", 81], ["2009-07-28", 79], ["2009-07-29", 107], ["2009-07-30", 109], ["2009-07-31", 71], ["2009-08-01", 89], ["2009-08-02", 59], ["2009-08-03", 75], ["2009-08-04", 97], ["2009-08-05", 74], ["2009-08-06", 58], ["2009-08-07", 74], ["2009-08-08", 75], ["2009-08-09", 81], ["2009-08-10", 60], ["2009-08-11", 75], ["2009-08-12", 68], ["2009-08-13", 82], ["2009-08-14", 123], ["2009-08-15", 115], ["2009-08-16", 113], ["2009-08-17", 63], ["2009-08-18", 76], ["2009-08-19", 77], ["2009-08-20", 38], ["2009-08-21", 62], ["2009-08-22", 58], ["2009-08-23", 71], ["2009-08-24", 97], ["2009-08-25", 90], ["2009-08-26", 97], ["2009-08-27", 69], ["2009-08-28", 36], ["2009-08-29", 61], ["2009-08-30", 69], ["2009-08-31", 78], ["2009-09-01", 88], ["2009-09-02", 98], ["2009-09-03", 109], ["2009-09-04", 99], ["2009-09-05", 92], ["2009-09-06", 32], ["2009-09-07", 20], ["2009-09-08", 51], ["2009-09-09", 66], ["2009-09-10", 77], ["2009-09-11", 69], ["2009-09-12", 42], ["2009-09-13", 65], ["2009-09-14", 91], ["2009-09-15", 72], ["2009-09-16", 93], ["2009-09-17", 117], ["2009-09-18", 121], ["2009-09-19", 75], ["2009-09-20", 101], ["2009-09-21", 111], ["2009-09-22", 79], ["2009-09-23", 90], ["2009-09-24", 108], ["2009-09-25", 130], ["2009-09-26", 98], ["2009-09-27", 66], ["2009-09-28", 74], ["2009-09-29", 97], ["2009-09-30", 112], ["2009-10-01", 88], ["2009-10-02", 22], ["2009-10-03", 29], ["2009-10-04", 44], ["2009-10-05", 69], ["2009-10-06", 83], ["2009-10-07", 74], ["2009-10-08", 72], ["2009-10-09", 88], ["2009-10-10", 73], ["2009-10-11", 94], ["2009-10-12", 108], ["2009-10-13", 37], ["2009-10-14", 42], ["2009-10-15", 72], ["2009-10-16", 114], ["2009-10-17", 57], ["2009-10-18", 92], ["2009-10-19", 90], ["2009-10-20", 77], ["2009-10-21", 76], ["2009-10-22", 100], ["2009-10-23", 111], ["2009-10-24", 141], ["2009-10-25", 147], ["2009-10-26", 77], ["2009-10-27", 68], ["2009-10-28", 100], ["2009-10-29", 137], ["2009-10-30", 120], ["2009-10-31", 51], ["2009-11-01", 48], ["2009-11-02", 12], ["2009-11-03", 66], ["2009-11-04", 111], ["2009-11-05", 136], ["2009-11-06", 186], ["2009-11-07", 276], ["2009-11-08", 259], ["2009-11-09", 84], ["2009-11-10", 20], ["2009-11-11", 34], ["2009-11-12", 53], ["2009-11-13", 59], ["2009-11-14", 53], ["2009-11-15", 26], ["2009-11-16", 29], ["2009-11-17", 35], ["2009-11-18", 66], ["2009-11-19", 47], ["2009-11-20", 74], ["2009-11-21", 63], ["2009-11-22", 121], ["2009-11-23", 149], ["2009-11-24", 184], ["2009-11-25", 79], ["2009-11-26", 107], ["2009-11-27", 132], ["2009-11-28", 99], ["2009-11-29", 167], ["2009-11-30", 117], ["2009-12-01", 86], ["2009-12-02", 133], ["2009-12-03", 36], ["2009-12-04", 99], ["2009-12-05", 62], ["2009-12-06", 94], ["2009-12-07", 141], ["2009-12-08", 186], ["2009-12-10", 167], ["2009-12-11", 147], ["2009-12-12", 31], ["2009-12-13", 80], ["2009-12-14", 96], ["2009-12-15", 49], ["2009-12-16", 55], ["2009-12-17", 45], ["2009-12-18", 42], ["2009-12-19", 44], ["2009-12-20", 48], ["2009-12-21", 63], ["2009-12-22", 94], ["2009-12-23", 93], ["2009-12-24", 133], ["2009-12-25", 500], ["2009-12-26", 96], ["2009-12-27", 94], ["2009-12-28", 89], ["2009-12-29", 160], ["2009-12-30", 55], ["2009-12-31", 55], ["2010-01-01", 91], ["2010-01-02", 105], ["2010-01-03", 90], ["2010-01-04", 49], ["2010-01-05", 47], ["2010-01-06", 59], ["2010-01-07", 64], ["2010-01-08", 80], ["2010-01-09", 100], ["2010-01-10", 60], ["2010-01-11", 52], ["2010-01-12", 30], ["2010-01-13", 54], ["2010-01-14", 76], ["2010-01-15", 58], ["2010-01-16", 85], ["2010-01-17", 124], ["2010-01-18", 143], ["2010-01-19", 183], ["2010-01-20", 140], ["2010-01-21", 24], ["2010-01-22", 57], ["2010-01-23", 78], ["2010-01-24", 66], ["2010-01-25", 99], ["2010-01-26", 76], ["2010-01-27", 128], ["2010-01-28", 63], ["2010-01-29", 43], ["2010-01-30", 58], ["2010-01-31", 56], ["2010-02-01", 65], ["2010-02-02", 61], ["2010-02-03", 54], ["2010-02-04", 54], ["2010-02-05", 63], ["2010-02-06", 70], ["2010-02-07", 61], ["2010-02-08", 87], ["2010-02-09", 109], ["2010-02-10", 50], ["2010-02-11", 23], ["2010-02-12", 31], ["2010-02-14", 137], ["2010-02-15", 38], ["2010-02-16", 52], ["2010-02-17", 94], ["2010-02-18", 58], ["2010-02-19", 98], ["2010-02-20", 87], ["2010-02-21", 118], ["2010-02-22", 82], ["2010-02-23", 92], ["2010-02-24", 152], ["2010-02-25", 153], ["2010-02-26", 76], ["2010-02-27", 65], ["2010-02-28", 80], ["2010-03-01", 56], ["2010-03-02", 72], ["2010-03-03", 113], ["2010-03-04", 140], ["2010-03-05", 97], ["2010-03-06", 27], ["2010-03-07", 71], ["2010-03-08", 68], ["2010-03-09", 25], ["2010-03-10", 56], ["2010-03-11", 89], ["2010-03-12", 98], ["2010-03-13", 76], ["2010-03-14", 90], ["2010-03-16", 77], ["2010-03-17", 66], ["2010-03-18", 76], ["2010-03-19", 145], ["2010-03-20", 500], ["2010-03-21", 136], ["2010-03-22", 245], ["2010-03-23", 157], ["2010-03-24", 92], ["2010-03-25", 60], ["2010-03-26", 83], ["2010-03-27", 110], ["2010-03-28", 82], ["2010-03-29", 100], ["2010-03-30", 159], ["2010-03-31", 94], ["2010-04-01", 99], ["2010-04-02", 63], ["2010-04-03", 73], ["2010-04-04", 147], ["2010-04-05", 125], ["2010-04-06", 56], ["2010-04-07", 77], ["2010-04-08", 147], ["2010-04-09", 163], ["2010-04-10", 69], ["2010-04-11", 77], ["2010-04-12", 66], ["2010-04-13", 61], ["2010-04-14", 59], ["2010-04-15", 93], ["2010-04-16", 147], ["2010-04-17", 94], ["2010-04-18", 109], ["2010-04-19", 150], ["2010-04-20", 74], ["2010-04-21", 60], ["2010-04-22", 31], ["2010-04-23", 40], ["2010-04-24", 72], ["2010-04-25", 100], ["2010-04-26", 45], ["2010-04-27", 50], ["2010-04-28", 52], ["2010-04-29", 46], ["2010-04-30", 54], ["2010-05-01", 90], ["2010-05-02", 116], ["2010-05-03", 97], ["2010-05-04", 149], ["2010-05-05", 119], ["2010-05-06", 17], ["2010-05-07", 86], ["2010-05-08", 145], ["2010-05-09", 144], ["2010-05-10", 146], ["2010-05-11", 58], ["2010-05-12", 59], ["2010-05-13", 78], ["2010-05-14", 95], ["2010-05-15", 133], ["2010-05-16", 121], ["2010-05-17", 52], ["2010-05-18", 53], ["2010-05-19", 61], ["2010-05-20", 75], ["2010-05-22", 127], ["2010-05-23", 122], ["2010-05-24", 91], ["2010-05-25", 46], ["2010-05-26", 76], ["2010-05-27", 82], ["2010-05-28", 63], ["2010-05-29", 84], ["2010-05-30", 39], ["2010-05-31", 58], ["2010-06-01", 69], ["2010-06-02", 68], ["2010-06-03", 83], ["2010-06-04", 88], ["2010-06-05", 96], ["2010-06-06", 114], ["2010-06-07", 118], ["2010-06-08", 98], ["2010-06-09", 86], ["2010-06-10", 64], ["2010-06-11", 58], ["2010-06-12", 81], ["2010-06-13", 82], ["2010-06-14", 66], ["2010-06-15", 95], ["2010-06-16", 77], ["2010-06-17", 56], ["2010-06-18", 47], ["2010-06-19", 77], ["2010-06-20", 71], ["2010-06-21", 71], ["2010-06-23", 77], ["2010-06-24", 83], ["2010-06-25", 99], ["2010-06-26", 112], ["2010-06-27", 93], ["2010-06-28", 94], ["2010-06-29", 123], ["2010-06-30", 100], ["2010-07-01", 118], ["2010-07-02", 40], ["2010-07-03", 63], ["2010-07-04", 86], ["2010-07-05", 66], ["2010-07-06", 54], ["2010-07-07", 73], ["2010-07-08", 80], ["2010-07-09", 74], ["2010-07-10", 59], ["2010-07-11", 68], ["2010-07-12", 73], ["2010-07-13", 84], ["2010-07-14", 78], ["2010-07-15", 89], ["2010-07-16", 115], ["2010-07-17", 84], ["2010-07-18", 87], ["2010-07-19", 121], ["2010-07-21", 63], ["2010-07-22", 90], ["2010-07-23", 123], ["2010-07-24", 88], ["2010-07-25", 100], ["2010-07-26", 121], ["2010-07-27", 139], ["2010-07-28", 100], ["2010-07-29", 119], ["2010-07-30", 113], ["2010-07-31", 92], ["2010-08-01", 48], ["2010-08-02", 68], ["2010-08-03", 83], ["2010-08-04", 98], ["2010-08-05", 26], ["2010-08-06", 31], ["2010-08-07", 71], ["2010-08-08", 57], ["2010-08-09", 94], ["2010-08-10", 90], ["2010-08-11", 94], ["2010-08-12", 64], ["2010-08-13", 83], ["2010-08-13", 83], ["2010-08-14", 84], ["2010-08-15", 57], ["2010-08-16", 66], ["2010-08-17", 94], ["2010-08-18", 137], ["2010-08-19", 73], ["2010-08-20", 76], ["2010-08-21", 56], ["2010-08-22", 23], ["2010-08-23", 54], ["2010-08-24", 87], ["2010-08-25", 65], ["2010-08-26", 66], ["2010-08-27", 52], ["2010-08-28", 55], ["2010-08-29", 76], ["2010-08-30", 79], ["2010-08-31", 78], ["2010-09-01", 67], ["2010-09-02", 54], ["2010-09-03", 73], ["2010-09-04", 64], ["2010-09-05", 80], ["2010-09-06", 87], ["2010-09-07", 95], ["2010-09-08", 67], ["2010-09-09", 89], ["2010-09-10", 75], ["2010-09-11", 49], ["2010-09-12", 67], ["2010-09-13", 84], ["2010-09-14", 97], ["2010-09-15", 134], ["2010-09-16", 122], ["2010-09-17", 62], ["2010-09-18", 19], ["2010-09-19", 50], ["2010-09-20", 60], ["2010-09-21", 23], ["2010-09-22", 24], ["2010-09-23", 52], ["2010-09-24", 72], ["2010-09-25", 93], ["2010-09-26", 84], ["2010-09-27", 57], ["2010-09-28", 32], ["2010-09-29", 65], ["2010-09-30", 92], ["2010-10-01", 125], ["2010-10-02", 88], ["2010-10-03", 17], ["2010-10-04", 36], ["2010-10-05", 63], ["2010-10-06", 95], ["2010-10-07", 186], ["2010-10-08", 192], ["2010-10-09", 177], ["2010-10-10", 202], ["2010-10-11", 70], ["2010-10-12", 27], ["2010-10-13", 65], ["2010-10-14", 58], ["2010-10-15", 30], ["2010-10-16", 80], ["2010-10-17", 65], ["2010-10-18", 80], ["2010-10-19", 50], ["2010-10-20", 66], ["2010-10-21", 83], ["2010-10-22", 95], ["2010-10-23", 103], ["2010-10-24", 96], ["2010-10-25", 17], ["2010-10-26", 15], ["2010-10-27", 63], ["2010-10-28", 92], ["2010-10-29", 67], ["2010-10-30", 62], ["2010-10-31", 70], ["2010-11-01", 65], ["2010-11-02", 36], ["2010-11-03", 86], ["2010-11-04", 81], ["2010-11-05", 86], ["2010-11-06", 107], ["2010-11-07", 142], ["2010-11-08", 34], ["2010-11-09", 34], ["2010-11-10", 85], ["2010-11-11", 139], ["2010-11-12", 51], ["2010-11-13", 66], ["2010-11-14", 39], ["2010-11-15", 34], ["2010-11-16", 96], ["2010-11-17", 122], ["2010-11-18", 243], ["2010-11-19", 313], ["2010-11-20", 165], ["2010-11-21", 192], ["2010-11-22", 37], ["2010-11-23", 100], ["2010-11-24", 141], ["2010-11-25", 42], ["2010-11-26", 88], ["2010-11-27", 130], ["2010-11-28", 72], ["2010-11-29", 143], ["2010-11-30", 132], ["2010-12-01", 177], ["2010-12-02", 199], ["2010-12-03", 52], ["2010-12-04", 97], ["2010-12-05", 125], ["2010-12-06", 37], ["2010-12-07", 65], ["2010-12-08", 81], ["2010-12-09", 97], ["2010-12-10", 176], ["2010-12-11", 50], ["2010-12-12", 85], ["2010-12-13", 72], ["2010-12-14", 31], ["2010-12-15", 53], ["2010-12-16", 92], ["2010-12-17", 105], ["2010-12-18", 156], ["2010-12-19", 182], ["2010-12-20", 100], ["2010-12-21", 165], ["2010-12-22", 222], ["2010-12-23", 30], ["2010-12-24", 40], ["2010-12-25", 57], ["2010-12-26", 66], ["2010-12-27", 82], ["2010-12-28", 70], ["2010-12-29", 63], ["2010-12-30", 67], ["2010-12-31", 47], ["2011-01-01", 34], ["2011-01-02", 41], ["2011-01-03", 82], ["2011-01-04", 96], ["2011-01-05", 55], ["2011-01-06", 35], ["2011-01-07", 36], ["2011-01-08", 78], ["2011-01-09", 35], ["2011-01-10", 34], ["2011-01-11", 67], ["2011-01-12", 49], ["2011-01-13", 90], ["2011-01-14", 73], ["2011-01-16", 35], ["2011-01-17", 62], ["2011-01-18", 30], ["2011-01-19", 39], ["2011-01-20", 36], ["2011-01-21", 61], ["2011-01-22", 76], ["2011-01-23", 50], ["2011-01-24", 35], ["2011-01-25", 61], ["2011-01-26", 41], ["2011-01-27", 59], ["2011-01-28", 41], ["2011-01-29", 30], ["2011-01-30", 25], ["2011-01-31", 48], ["2011-02-01", 53], ["2011-02-02", 58], ["2011-02-03", 83], ["2011-02-04", 111], ["2011-02-05", 75], ["2011-02-06", 84], ["2011-02-07", 77], ["2011-02-09", 83], ["2011-02-10", 58], ["2011-02-11", 58], ["2011-02-12", 21], ["2011-02-13", 53], ["2011-02-14", 41], ["2011-02-15", 74], ["2011-02-16", 146], ["2011-02-17", 132], ["2011-02-18", 115], ["2011-02-19", 112], ["2011-02-20", 100], ["2011-02-21", 333], ["2011-02-22", 270], ["2011-02-23", 208], ["2011-02-25", 56], ["2011-02-26", 56], ["2011-02-27", 60], ["2011-02-28", 30], ["2011-03-01", 21], ["2011-03-02", 33], ["2011-03-03", 34], ["2011-03-04", 59], ["2011-03-05", 77], ["2011-03-06", 65], ["2011-03-07", 26], ["2011-03-08", 41], ["2011-03-09", 33], ["2011-03-10", 64], ["2011-03-11", 58], ["2011-03-12", 135], ["2011-03-13", 197], ["2011-03-14", 54], ["2011-03-15", 56], ["2011-03-16", 72], ["2011-03-17", 98], ["2011-03-18", 161], ["2011-03-19", 123], ["2011-03-20", 250], ["2011-03-21", 121], ["2011-03-22", 67], ["2011-03-23", 51], ["2011-03-24", 51], ["2011-03-25", 48], ["2011-03-26", 78], ["2011-03-27", 41], ["2011-03-28", 71], ["2011-03-29", 86], ["2011-03-30", 98], ["2011-03-31", 140], ["2011-04-01", 137], ["2011-04-02", 38], ["2011-04-03", 59], ["2011-04-04", 67], ["2011-04-05", 88], ["2011-04-06", 95], ["2011-04-07", 96], ["2011-04-08", 70], ["2011-04-09", 108], ["2011-04-10", 142], ["2011-04-11", 53], ["2011-04-12", 88], ["2011-04-13", 157], ["2011-04-14", 138], ["2011-04-15", 98], ["2011-04-16", 128], ["2011-04-17", 164], ["2011-04-18", 99], ["2011-04-19", 83], ["2011-04-20", 127], ["2011-04-21", 154], ["2011-04-22", 44], ["2011-04-23", 49], ["2011-04-24", 26], ["2011-04-25", 76], ["2011-04-26", 111], ["2011-04-27", 60], ["2011-04-28", 76], ["2011-04-29", 119], ["2011-04-30", 141], ["2011-05-01", 500], ["2011-05-02", 85], ["2011-05-03", 60], ["2011-05-04", 79], ["2011-05-05", 87], ["2011-05-06", 99], ["2011-05-07", 57], ["2011-05-08", 74], ["2011-05-09", 53], ["2011-05-10", 50], ["2011-05-11", 80], ["2011-05-12", 197], ["2011-05-13", 52], ["2011-05-14", 70], ["2011-05-15", 76], ["2011-05-16", 90], ["2011-05-17", 91], ["2011-05-18", 155], ["2011-05-19", 64], ["2011-05-20", 59], ["2011-05-21", 54], ["2011-05-22", 83], ["2011-05-23", 98], ["2011-05-24", 94], ["2011-05-25", 75], ["2011-05-26", 86], ["2011-05-27", 65], ["2011-05-28", 102], ["2011-05-29", 98], ["2011-05-30", 75], ["2011-05-31", 47], ["2011-06-01", 28], ["2011-06-02", 75], ["2011-06-03", 75], ["2011-06-04", 66], ["2011-06-05", 79], ["2011-06-06", 83], ["2011-06-07", 98], ["2011-06-08", 51], ["2011-06-10", 105], ["2011-06-11", 75], ["2011-06-12", 28], ["2011-06-13", 71], ["2011-06-14", 99], ["2011-06-15", 107], ["2011-06-16", 77], ["2011-06-17", 81], ["2011-06-18", 97], ["2011-06-19", 119], ["2011-06-20", 122], ["2011-06-21", 130], ["2011-06-22", 128], ["2011-06-23", 123], ["2011-06-24", 24], ["2011-06-25", 38], ["2011-06-26", 57], ["2011-06-27", 56], ["2011-06-28", 90], ["2011-06-29", 129], ["2011-06-30", 99], ["2011-07-01", 94], ["2011-07-02", 71], ["2011-07-03", 71], ["2011-07-04", 55], ["2011-07-05", 80], ["2011-07-06", 115], ["2011-07-07", 73], ["2011-07-08", 42], ["2011-07-09", 37], ["2011-07-10", 75], ["2011-07-11", 112], ["2011-07-12", 88], ["2011-07-13", 83], ["2011-07-14", 83], ["2011-07-15", 65], ["2011-07-16", 65], ["2011-07-17", 67], ["2011-07-18", 65], ["2011-07-19", 83], ["2011-07-20", 42], ["2011-07-21", 53], ["2011-07-22", 71], ["2011-07-23", 148], ["2011-07-24", 159], ["2011-07-25", 19], ["2011-07-26", 28], ["2011-07-27", 52], ["2011-07-28", 92], ["2011-07-29", 113], ["2011-07-30", 21], ["2011-07-31", 54], ["2011-08-01", 78], ["2011-08-02", 94], ["2011-08-03", 69], ["2011-08-04", 82], ["2011-08-05", 98], ["2011-08-06", 91], ["2011-08-07", 74], ["2011-08-08", 77], ["2011-08-09", 108], ["2011-08-10", 58], ["2011-08-11", 68], ["2011-08-12", 90], ["2011-08-13", 93], ["2011-08-14", 78], ["2011-08-15", 73], ["2011-08-16", 29], ["2011-08-17", 58], ["2011-08-18", 28], ["2011-08-19", 65], ["2011-08-20", 72], ["2011-08-21", 80], ["2011-08-22", 78], ["2011-08-23", 88], ["2011-08-24", 95], ["2011-08-25", 80], ["2011-08-26", 61], ["2011-08-27", 63], ["2011-08-28", 65], ["2011-08-29", 80], ["2011-08-30", 99], ["2011-08-31", 117], ["2011-09-01", 89], ["2011-09-02", 54], ["2011-09-03", 69], ["2011-09-04", 77], ["2011-09-05", 76], ["2011-09-06", 76], ["2011-09-07", 126], ["2011-09-08", 48], ["2011-09-09", 39], ["2011-09-10", 35], ["2011-09-11", 24], ["2011-09-12", 61], ["2011-09-13", 81], ["2011-09-14", 87], ["2011-09-15", 93], ["2011-09-16", 52], ["2011-09-17", 22], ["2011-09-18", 35], ["2011-09-19", 45], ["2011-09-20", 50], ["2011-09-21", 52], ["2011-09-22", 58], ["2011-09-24", 96], ["2011-09-25", 125], ["2011-09-26", 160], ["2011-09-27", 121], ["2011-09-28", 128], ["2011-09-29", 94], ["2011-09-30", 30], ["2011-10-01", 56], ["2011-10-02", 33], ["2011-10-03", 47], ["2011-10-04", 79], ["2011-10-05", 157], ["2011-10-06", 61], ["2011-10-07", 84], ["2011-10-08", 106], ["2011-10-09", 159], ["2011-10-10", 137], ["2011-10-11", 87], ["2011-10-12", 130], ["2011-10-13", 98], ["2011-10-14", 32], ["2011-10-15", 33], ["2011-10-16", 31], ["2011-10-17", 35], ["2011-10-18", 72], ["2011-10-19", 87], ["2011-10-20", 149], ["2011-10-21", 146], ["2011-10-22", 139], ["2011-10-23", 155], ["2011-10-24", 19], ["2011-10-25", 28], ["2011-10-26", 78], ["2011-10-27", 129], ["2011-10-29", 97], ["2011-10-30", 147], ["2011-10-31", 131], ["2011-11-01", 128], ["2011-11-02", 53], ["2011-11-03", 68], ["2011-11-04", 82], ["2011-11-05", 60], ["2011-11-06", 52], ["2011-11-07", 63], ["2011-11-08", 73], ["2011-11-09", 49], ["2011-11-10", 60], ["2011-11-11", 84], ["2011-11-12", 99], ["2011-11-13", 65], ["2011-11-14", 73], ["2011-11-15", 124], ["2011-11-16", 128], ["2011-11-17", 97], ["2011-11-18", 62], ["2011-11-19", 36], ["2011-11-20", 27], ["2011-11-21", 80], ["2011-11-22", 131], ["2011-11-23", 40], ["2011-11-24", 68], ["2011-11-25", 120], ["2011-11-26", 142], ["2011-11-27", 135], ["2011-11-28", 109], ["2011-11-29", 66], ["2011-11-30", 81], ["2011-12-01", 71], ["2011-12-02", 144], ["2011-12-03", 97], ["2011-12-04", 80], ["2011-12-05", 193], ["2011-12-06", 131], ["2011-12-07", 111], ["2011-12-08", 17], ["2011-12-09", 19], ["2011-12-10", 23], ["2011-12-11", 77], ["2011-12-12", 56], ["2011-12-13", 76], ["2011-12-14", 84], ["2011-12-15", 19], ["2011-12-16", 27], ["2011-12-17", 63], ["2011-12-18", 63], ["2011-12-19", 53], ["2011-12-20", 70], ["2011-12-21", 67], ["2011-12-22", 31], ["2011-12-23", 61], ["2011-12-24", 27], ["2011-12-25", 59], ["2011-12-26", 69], ["2011-12-27", 100], ["2011-12-28", 114], ["2011-12-29", 81], ["2011-12-30", 75], ["2011-12-31", 109], ["2012-01-01", 81], ["2012-01-02", 74], ["2012-01-03", 35], ["2012-01-04", 30], ["2012-01-05", 63], ["2012-01-06", 95], ["2012-01-07", 65], ["2012-01-08", 89], ["2012-01-09", 102], ["2012-01-10", 161], ["2012-01-11", 25], ["2012-01-12", 86], ["2012-01-13", 79], ["2012-01-14", 60], ["2012-01-15", 70], ["2012-01-16", 106], ["2012-01-17", 111], ["2012-01-18", 193], ["2012-01-19", 269], ["2012-01-20", 131], ["2012-01-21", 21], ["2012-01-22", 23], ["2012-01-23", 149], ["2012-01-24", 49], ["2012-01-25", 45], ["2012-01-26", 78], ["2012-01-27", 67], ["2012-01-28", 74], ["2012-01-29", 62], ["2012-01-30", 66], ["2012-01-31", 92], ["2012-02-01", 30], ["2012-02-02", 26], ["2012-02-03", 60], ["2012-02-04", 52], ["2012-02-05", 84], ["2012-02-06", 112], ["2012-02-07", 64], ["2012-02-08", 34], ["2012-02-09", 58], ["2012-02-10", 49], ["2012-02-11", 73], ["2012-02-12", 75], ["2012-02-13", 100], ["2012-02-14", 125], ["2012-02-15", 62], ["2012-02-16", 61], ["2012-02-17", 34], ["2012-02-18", 29], ["2012-02-19", 68], ["2012-02-20", 73], ["2012-02-21", 118], ["2012-02-22", 118], ["2012-02-23", 73], ["2012-02-24", 73], ["2012-02-25", 57], ["2012-02-26", 57], ["2012-02-27", 95], ["2012-02-28", 152], ["2012-02-29", 118], ["2012-03-01", 142], ["2012-03-02", 111], ["2012-03-03", 68], ["2012-03-04", 90], ["2012-03-05", 97], ["2012-03-06", 63], ["2012-03-07", 38], ["2012-03-08", 31], ["2012-03-09", 65], ["2012-03-10", 78], ["2012-03-11", 36], ["2012-03-12", 62], ["2012-03-13", 104], ["2012-03-14", 57], ["2012-03-15", 64], ["2012-03-16", 109], ["2012-03-17", 144], ["2012-03-18", 61], ["2012-03-19", 57], ["2012-03-20", 81], ["2012-03-21", 105], ["2012-03-22", 146], ["2012-03-23", 55], ["2012-03-24", 56], ["2012-03-25", 30], ["2012-03-26", 90], ["2012-03-27", 112], ["2012-03-28", 65], ["2012-03-29", 90], ["2012-03-30", 76], ["2012-03-31", 159], ["2012-04-01", 78], ["2012-04-02", 103], ["2012-04-03", 73], ["2012-04-03", 73], ["2012-04-04", 73], ["2012-04-05", 64], ["2012-04-06", 70], ["2012-04-07", 71], ["2012-04-08", 119], ["2012-04-09", 118], ["2012-04-10", 138], ["2012-04-11", 41], ["2012-04-12", 69], ["2012-04-13", 81], ["2012-04-14", 100], ["2012-04-15", 109], ["2012-04-16", 84], ["2012-04-17", 100], ["2012-04-18", 140], ["2012-04-19", 98], ["2012-04-20", 133], ["2012-04-21", 81], ["2012-04-22", 102], ["2012-04-23", 140], ["2012-04-24", 133], ["2012-04-25", 32], ["2012-04-26", 60], ["2012-04-27", 147], ["2012-04-28", 164], ["2012-04-29", 473], ["2012-04-30", 268], ["2012-05-01", 208], ["2012-05-02", 111], ["2012-05-03", 106], ["2012-05-04", 100], ["2012-05-05", 99], ["2012-05-06", 100], ["2012-05-07", 100], ["2012-05-08", 111], ["2012-05-09", 107], ["2012-05-10", 129], ["2012-05-11", 133], ["2012-05-12", 90], ["2012-05-13", 96], ["2012-05-14", 64], ["2012-05-15", 58], ["2012-05-16", 58], ["2012-05-17", 78], ["2012-05-18", 84], ["2012-05-19", 143], ["2012-05-20", 85], ["2012-05-21", 97], ["2012-05-22", 109], ["2012-05-23", 64], ["2012-05-24", 69], ["2012-05-25", 63], ["2012-05-26", 90], ["2012-05-27", 88], ["2012-05-28", 133], ["2012-05-29", 116], ["2012-05-30", 29], ["2012-05-31", 64], ["2012-06-01", 54], ["2012-06-02", 90], ["2012-06-03", 112], ["2012-06-04", 80], ["2012-06-05", 65], ["2012-06-06", 98], ["2012-06-07", 71], ["2012-06-08", 77], ["2012-06-09", 91], ["2012-06-10", 32], ["2012-06-11", 50], ["2012-06-12", 58], ["2012-06-13", 62], ["2012-06-14", 50], ["2012-06-15", 22], ["2012-06-16", 33], ["2012-06-17", 69], ["2012-06-18", 137], ["2012-06-19", 132], ["2012-06-20", 105], ["2012-06-21", 112], ["2012-06-22", 84], ["2012-06-23", 81], ["2012-06-24", 95], ["2012-06-25", 49], ["2012-06-26", 65], ["2012-06-27", 55], ["2012-06-28", 54], ["2012-06-29", 60], ["2012-06-30", 46], ["2012-07-01", 70], ["2012-07-02", 69], ["2012-07-03", 59], ["2012-07-04", 71], ["2012-07-05", 70], ["2012-07-06", 59], ["2012-07-07", 86], ["2012-07-08", 84], ["2012-07-09", 64], ["2012-07-10", 50], ["2012-07-11", 44], ["2012-07-12", 46], ["2012-07-13", 31], ["2012-07-14", 48], ["2012-07-15", 53], ["2012-07-16", 70], ["2012-07-17", 78], ["2012-07-18", 71], ["2012-07-19", 82], ["2012-07-20", 111], ["2012-07-21", 131], ["2012-07-22", 15], ["2012-07-24", 60], ["2012-07-25", 72], ["2012-07-26", 55], ["2012-07-26", 55], ["2012-07-27", 50], ["2012-07-28", 56], ["2012-07-29", 57], ["2012-07-30", 30], ["2012-07-31", 28], ["2012-08-01", 20], ["2012-08-02", 17], ["2012-08-03", 53], ["2012-08-04", 40], ["2012-08-05", 48], ["2012-08-06", 60], ["2012-08-07", 59], ["2012-08-08", 68], ["2012-08-09", 43], ["2012-08-10", 72], ["2012-08-11", 80], ["2012-08-12", 41], ["2012-08-13", 36], ["2012-08-14", 62], ["2012-08-15", 60], ["2012-08-16", 68], ["2012-08-17", 83], ["2012-08-18", 110], ["2012-08-19", 84], ["2012-08-20", 92], ["2012-08-21", 25], ["2012-08-22", 40], ["2012-08-23", 74], ["2012-08-24", 94], ["2012-08-25", 92], ["2012-08-26", 117], ["2012-08-27", 100], ["2012-08-28", 59], ["2012-08-29", 84], ["2012-08-30", 135], ["2012-08-31", 150], ["2012-09-01", 128], ["2012-09-02", 52], ["2012-09-03", 15], ["2012-09-04", 22], ["2012-09-05", 50], ["2012-09-06", 70], ["2012-09-07", 77], ["2012-09-08", 40], ["2012-09-09", 79], ["2012-09-10", 96], ["2012-09-11", 93], ["2012-09-12", 44], ["2012-09-13", 28], ["2012-09-14", 31], ["2012-09-15", 50], ["2012-09-16", 65], ["2012-09-17", 63], ["2012-09-18", 61], ["2012-09-19", 56], ["2012-09-21", 128], ["2012-09-22", 93], ["2012-09-23", 85], ["2012-09-24", 74], ["2012-09-25", 78], ["2012-09-26", 26], ["2012-09-27", 65], ["2012-09-28", 15], ["2012-09-29", 24], ["2012-09-30", 38], ["2012-10-01", 52], ["2012-10-02", 78], ["2012-10-03", 108], ["2012-10-04", 28], ["2012-10-05", 41], ["2012-10-06", 74], ["2012-10-07", 83], ["2012-10-08", 123], ["2012-10-09", 140], ["2012-10-10", 18], ["2012-10-11", 73], ["2012-10-12", 121], ["2012-10-13", 97], ["2012-10-14", 40], ["2012-10-15", 83], ["2012-10-16", 78], ["2012-10-17", 23], ["2012-10-18", 65], ["2012-10-19", 79], ["2012-10-20", 139], ["2012-10-21", 81], ["2012-10-22", 26], ["2012-10-23", 54], ["2012-10-24", 89], ["2012-10-25", 90], ["2012-10-26", 163], ["2012-10-27", 154], ["2012-10-28", 22], ["2012-10-29", 59], ["2012-10-30", 36], ["2012-10-31", 51], ["2012-11-01", 67], ["2012-11-02", 103], ["2012-11-03", 135], ["2012-11-04", 20], ["2012-11-05", 16], ["2012-11-06", 48], ["2012-11-07", 80], ["2012-11-08", 62], ["2012-11-09", 93], ["2012-11-10", 82], ["2012-11-11", 17], ["2012-11-12", 27], ["2012-11-13", 30], ["2012-11-14", 26], ["2012-11-15", 71], ["2012-11-16", 92], ["2012-11-17", 47], ["2012-11-18", 96], ["2012-11-19", 55], ["2012-11-20", 74], ["2012-11-21", 123], ["2012-11-22", 156], ["2012-11-23", 22], ["2012-11-24", 80], ["2012-11-25", 133], ["2012-11-26", 44], ["2012-11-27", 105], ["2012-11-28", 151], ["2012-11-29", 54], ["2012-12-01", 50], ["2012-12-02", 96], ["2012-12-03", 123], ["2012-12-04", 50], ["2012-12-05", 64], ["2012-12-06", 50], ["2012-12-07", 73], ["2012-12-08", 53], ["2012-12-09", 38], ["2012-12-10", 53], ["2012-12-11", 86], ["2012-12-12", 103], ["2012-12-13", 130], ["2012-12-14", 107], ["2012-12-15", 114], ["2012-12-16", 108], ["2012-12-17", 45], ["2012-12-18", 22], ["2012-12-19", 72], ["2012-12-20", 121], ["2012-12-21", 120], ["2012-12-22", 24], ["2012-12-23", 36], ["2012-12-24", 53], ["2012-12-25", 58], ["2012-12-26", 67], ["2012-12-28", 137], ["2012-12-29", 94], ["2012-12-30", 38], ["2012-12-31", 57], ["2013-01-01", 71], ["2013-01-02", 27], ["2013-01-03", 35], ["2013-01-04", 57], ["2013-01-05", 79], ["2013-01-06", 58], ["2013-01-07", 105], ["2013-01-08", 124], ["2013-01-09", 32], ["2013-01-10", 87], ["2013-01-11", 232], ["2013-01-12", 174], ["2013-01-13", 498], ["2013-01-14", 184], ["2014-01-01", 85], ["2014-01-02", 158], ["2014-01-03", 74], ["2014-01-04", 165], ["2014-01-05", 113], ["2014-01-06", 190], ["2014-01-07", 122], ["2014-01-10", 95], ["2014-01-11", 159], ["2014-01-12", 52], ["2014-01-13", 117], ["2014-01-14", 113], ["2014-01-15", 180], ["2014-01-16", 403], ["2014-01-17", 209], ["2014-01-18", 113], ["2014-01-19", 149], ["2014-01-21", 68], ["2014-01-22", 162], ["2014-01-23", 276], ["2014-01-24", 195], ["2014-01-26", 77], ["2014-01-27", 114], ["2014-01-28", 67], ["2014-01-29", 165], ["2014-01-30", 93], ["2014-01-31", 188], ["2014-02-01", 178], ["2014-02-02", 85], ["2014-02-05", 119], ["2014-02-06", 158], ["2014-02-07", 124], ["2014-02-08", 84], ["2014-02-10", 53], ["2014-02-11", 142], ["2014-02-12", 150], ["2014-02-13", 242], ["2014-02-14", 329], ["2014-02-15", 429], ["2014-02-16", 348], ["2014-02-17", 118], ["2014-02-18", 98], ["2014-02-19", 92], ["2014-02-20", 270], ["2014-02-21", 311], ["2014-02-22", 311], ["2014-02-23", 255], ["2014-02-24", 313], ["2014-02-25", 404], ["2014-02-28", 113], ["2014-03-01", 68], ["2014-03-02", 189], ["2014-03-03", 268], ["2014-03-04", 67], ["2014-03-07", 70], ["2014-03-08", 179], ["2014-03-09", 127], ["2014-03-10", 110], ["2014-03-11", 195], ["2014-03-13", 69], ["2014-03-14", 64], ["2014-03-15", 133], ["2014-03-16", 145], ["2014-03-17", 142], ["2014-03-18", 85], ["2014-03-19", 73], ["2014-03-21", 62], ["2014-03-22", 86], ["2014-03-23", 186], ["2014-03-24", 271], ["2014-03-25", 255], ["2014-03-26", 331], ["2014-03-27", 285], ["2014-03-28", 169], ["2014-03-29", 63], ["2014-03-30", 77], ["2014-03-31", 183], ["2014-04-01", 147], ["2014-04-02", 133], ["2014-04-03", 66], ["2014-04-04", 91], ["2014-04-05", 68], ["2014-04-06", 98], ["2014-04-07", 135], ["2014-04-08", 223], ["2014-04-09", 156], ["2014-04-10", 246], ["2014-04-11", 83], ["2014-04-12", 133], ["2014-04-13", 212], ["2014-04-14", 270], ["2014-04-15", 109], ["2014-04-16", 90], ["2014-04-17", 124], ["2014-04-18", 182], ["2014-04-19", 84], ["2014-04-20", 84], ["2014-04-21", 73], ["2014-04-22", 85], ["2014-04-23", 156], ["2014-04-24", 156], ["2014-04-25", 163], ["2014-04-26", 69], ["2014-04-27", 74], ["2014-04-28", 83], ["2014-04-29", 122], ["2014-04-30", 139], ["2014-05-01", 156], ["2014-05-03", 93], ["2014-05-04", 57], ["2014-05-05", 54], ["2014-05-06", 105], ["2014-05-07", 82], ["2014-05-08", 104], ["2014-05-09", 84], ["2014-05-10", 69], ["2014-05-12", 74], ["2014-05-13", 86], ["2014-05-14", 59], ["2014-05-15", 122], ["2014-05-16", 92], ["2014-05-17", 124], ["2014-05-18", 171], ["2014-05-19", 146], ["2014-05-20", 113], ["2014-05-21", 170], ["2014-05-22", 183], ["2014-05-23", 140], ["2014-05-24", 104], ["2014-05-25", 91], ["2014-05-26", 77], ["2014-05-27", 107], ["2014-05-28", 121], ["2014-05-29", 120], ["2014-05-30", 192], ["2014-05-31", 177], ["2014-06-01", 130], ["2014-06-02", 90], ["2014-06-03", 117], ["2014-06-04", 124], ["2014-06-05", 157], ["2014-06-06", 103], ["2014-06-07", 51], ["2014-06-08", 70], ["2014-06-09", 87], ["2014-06-10", 95], ["2014-06-11", 74], ["2014-06-12", 90], ["2014-06-13", 116], ["2014-06-14", 165], ["2014-06-15", 178], ["2014-06-16", 178], ["2014-06-17", 104], ["2014-06-18", 116], ["2014-06-19", 116], ["2014-06-20", 84], ["2014-06-21", 96], ["2014-06-22", 91], ["2014-06-23", 115], ["2014-06-24", 161], ["2014-06-25", 138], ["2014-06-26", 163], ["2014-06-27", 68], ["2014-06-28", 77], ["2014-06-29", 161], ["2014-06-30", 185], ["2014-07-01", 172], ["2014-07-02", 80], ["2014-07-03", 248], ["2014-07-04", 237], ["2014-07-05", 165], ["2014-07-06", 256], ["2014-07-07", 216], ["2014-07-08", 134], ["2014-07-09", 63], ["2014-07-10", 114], ["2014-07-11", 77], ["2014-07-12", 80], ["2014-07-13", 64], ["2014-07-14", 156], ["2014-07-15", 140], ["2014-07-16", 133], ["2014-07-17", 186], ["2014-07-18", 182], ["2014-07-19", 106], ["2014-07-20", 119], ["2014-07-21", 68], ["2014-07-22", 54], ["2014-07-23", 82], ["2014-07-24", 90], ["2014-07-25", 134], ["2014-07-26", 188], ["2014-07-27", 194], ["2014-07-28", 159], ["2014-07-29", 159], ["2014-07-30", 169], ["2014-07-31", 244], ["2014-08-01", 199], ["2014-08-02", 163], ["2014-08-03", 149], ["2014-08-05", 80], ["2014-08-06", 67], ["2014-08-07", 162], ["2014-08-08", 140], ["2014-08-09", 143], ["2014-08-10", 125], ["2014-08-11", 76], ["2014-08-12", 119], ["2014-08-13", 70], ["2014-08-14", 104], ["2014-08-15", 109], ["2014-08-16", 159], ["2014-08-17", 124], ["2014-08-18", 135], ["2014-08-19", 150], ["2014-08-20", 164], ["2014-08-21", 169], ["2014-08-22", 83], ["2014-08-23", 155], ["2014-08-24", 75], ["2014-08-25", 59], ["2014-08-26", 78], ["2014-08-27", 136], ["2014-08-28", 103], ["2014-08-29", 104], ["2014-08-30", 176], ["2014-08-31", 89], ["2014-09-01", 127], ["2014-09-03", 54], ["2014-09-04", 100], ["2014-09-05", 140], ["2014-09-06", 186], ["2014-09-07", 200], ["2014-09-08", 61], ["2014-09-09", 109], ["2014-09-10", 111], ["2014-09-11", 114], ["2014-09-12", 97], ["2014-09-13", 94], ["2014-09-14", 66], ["2014-09-15", 54], ["2014-09-16", 87], ["2014-09-17", 80], ["2014-09-18", 84], ["2014-09-19", 117], ["2014-09-20", 168], ["2014-09-21", 129], ["2014-09-22", 127], ["2014-09-23", 64], ["2014-09-24", 60], ["2014-09-25", 144], ["2014-09-26", 170], ["2014-09-27", 58], ["2014-09-28", 87], ["2014-09-29", 70], ["2014-09-30", 53], ["2014-10-01", 92], ["2014-10-02", 78], ["2014-10-03", 123], ["2014-10-04", 95], ["2014-10-05", 54], ["2014-10-06", 68], ["2014-10-07", 200], ["2014-10-08", 314], ["2014-10-09", 379], ["2014-10-10", 346], ["2014-10-11", 233], ["2014-10-14", 80], ["2014-10-15", 73], ["2014-10-16", 76], ["2014-10-17", 132], ["2014-10-18", 211], ["2014-10-19", 289], ["2014-10-20", 250], ["2014-10-21", 82], ["2014-10-22", 99], ["2014-10-23", 163], ["2014-10-24", 267], ["2014-10-25", 353], ["2014-10-26", 78], ["2014-10-27", 72], ["2014-10-28", 88], ["2014-10-29", 140], ["2014-10-30", 206], ["2014-10-31", 204], ["2014-11-01", 65], ["2014-11-03", 59], ["2014-11-04", 150], ["2014-11-05", 79], ["2014-11-07", 63], ["2014-11-08", 93], ["2014-11-09", 80], ["2014-11-10", 95], ["2014-11-11", 59], ["2014-11-13", 65], ["2014-11-14", 77], ["2014-11-15", 143], ["2014-11-16", 98], ["2014-11-17", 64], ["2014-11-18", 93], ["2014-11-19", 282], ["2014-11-23", 155], ["2014-11-24", 94], ["2014-11-25", 196], ["2014-11-26", 293], ["2014-11-27", 83], ["2014-11-28", 114], ["2014-11-29", 276], ["2014-12-01", 54], ["2014-12-02", 65], ["2014-12-03", 51], ["2014-12-05", 62], ["2014-12-06", 89], ["2014-12-07", 65], ["2014-12-08", 82], ["2014-12-09", 276], ["2014-12-10", 153], ["2014-12-11", 52], ["2014-12-13", 69], ["2014-12-14", 113], ["2014-12-15", 82], ["2014-12-17", 99], ["2014-12-19", 53], ["2014-12-22", 103], ["2014-12-23", 100], ["2014-12-25", 73], ["2014-12-26", 155], ["2014-12-27", 243], ["2014-12-28", 155], ["2014-12-29", 125], ["2014-12-30", 65], ["2015-01-01", 65], ["2015-01-02", 79], ["2015-01-03", 200], ["2015-01-04", 226], ["2015-01-05", 122], ["2015-01-06", 60], ["2015-01-07", 85], ["2015-01-08", 190], ["2015-01-09", 105], ["2015-01-10", 208], ["2015-01-11", 59], ["2015-01-12", 160], ["2015-01-13", 211], ["2015-01-14", 265], ["2015-01-15", 386], ["2015-01-16", 118], ["2015-01-17", 89], ["2015-01-18", 94], ["2015-01-19", 77], ["2015-01-20", 113], ["2015-01-22", 143], ["2015-01-23", 257], ["2015-01-24", 117], ["2015-01-25", 185], ["2015-01-26", 119], ["2015-01-28", 65], ["2015-01-29", 87], ["2015-01-31", 60], ["2015-02-01", 108], ["2015-02-02", 188], ["2015-02-03", 143], ["2015-02-05", 62], ["2015-02-06", 100], ["2015-02-09", 152], ["2015-02-10", 166], ["2015-02-11", 55], ["2015-02-12", 59], ["2015-02-13", 175], ["2015-02-14", 293], ["2015-02-15", 326], ["2015-02-16", 153], ["2015-02-18", 73], ["2015-02-19", 267], ["2015-02-20", 183], ["2015-02-21", 394], ["2015-02-22", 158], ["2015-02-23", 86], ["2015-02-24", 207], ] ( Line() .add_xaxis(xaxis_data=[item[0] for item in all_data]) .add_yaxis( series_name="", y_axis=[item[1] for item in all_data], yaxis_index=0, is_smooth=True, is_symbol_show=False, ) .set_global_opts( title_opts=opts.TitleOpts(title="Beijing AQI"), tooltip_opts=opts.TooltipOpts(trigger="axis"), datazoom_opts=[ opts.DataZoomOpts(yaxis_index=0), opts.DataZoomOpts(type_="inside", yaxis_index=0), ], visualmap_opts=opts.VisualMapOpts( pos_top="10", pos_right="10", is_piecewise=True, pieces=[ {"gt": 0, "lte": 50, "color": "#096"}, {"gt": 50, "lte": 100, "color": "#ffde33"}, {"gt": 100, "lte": 150, "color": "#ff9933"}, {"gt": 150, "lte": 200, "color": "#cc0033"}, {"gt": 200, "lte": 300, "color": "#660099"}, {"gt": 300, "color": "#7e0023"}, ], out_of_range={"color": "#999"}, ), xaxis_opts=opts.AxisOpts(type_="category"), yaxis_opts=opts.AxisOpts( type_="value", name_location="start", min_=0, max_=500, is_scale=True, axistick_opts=opts.AxisTickOpts(is_inside=False), ), ) .set_series_opts( markline_opts=opts.MarkLineOpts( data=[ {"yAxis": 50}, {"yAxis": 100}, {"yAxis": 150}, {"yAxis": 200}, {"yAxis": 300}, ], label_opts=opts.LabelOpts(position="end"), ) ) .render("beijing_aqi.html") )
122,030
91,257
""" Stacking and Visualizing ------------------------ This script does the following: 1. Input LSST images, warp to sparse matrix, store as scidb arrays. This tests the warping of a single LSST exposure into a sparse matrix representation of a HEALPix grid. """ import os import sys import glob import matplotlib.pyplot as plt import numpy as np sys.path.append(os.path.abspath('..')) from spheredb.scidb_tools import HPXPixels3D, find_index_bounds filenames = glob.glob("/home/jakevdp/research/LSST_IMGS/*/R*/S*.fits") print "total number of files:", len(filenames) HPX_data = HPXPixels3D(input_files=filenames[:20], name='LSSTdata', force_reload=False) times = HPX_data.unique_times() xlim, ylim, tlim = HPX_data.index_bounds() for time in times[:2]: tslice = HPX_data.time_slice(time) tslice_arr = tslice.arr[xlim[0]:xlim[1], ylim[0]:ylim[1]].toarray() fig, ax = plt.subplots() im = ax.imshow(np.log(tslice_arr), cmap=plt.cm.binary) ax.set_xlim(400, 440) ax.set_ylim(860, 820) fig.colorbar(im, ax=ax) ax.set_title("time = {0}".format(time)) coadd = HPX_data.coadd().arr[xlim[0]:xlim[1], ylim[0]:ylim[1]].toarray() fig, ax = plt.subplots() im = ax.imshow(np.log(coadd), cmap=plt.cm.binary) ax.set_xlim(400, 440) ax.set_ylim(860, 820) fig.colorbar(im, ax=ax) ax.set_title("coadd") plt.show()
1,416
580
import numpy as np sizes = list(map(int, input().split())) arr1 = list() arr2 = list() for _ in range(sizes[0]): arr1.append(list(map(int, input().split()))) for _ in range(sizes[1]): arr2.append(list(map(int, input().split()))) print(np.concatenate((np.array(arr1), np.array(arr2)), axis=0))
307
126
#!/usr/bin/env python # -*- coding: utf-8 -*- import unittest from bibdeskparser.customization import InvalidName, splitname class TestSplitnameMethod(unittest.TestCase): def test_splitname_basic(self): """Basic tests of customization.splitname() """ # Empty input. result = splitname("") expected = {} self.assertEqual(result, expected, msg="Invalid output for empty name") # Non-whitespace names. result = splitname(" ") expected = {} self.assertEqual( result, expected, msg="Invalid output for space-only name" ) result = splitname(" \t~~") expected = {} self.assertEqual( result, expected, msg="Invalid output for whitespace name" ) # Test strict mode. with self.assertRaises(InvalidName): # Trailing comma (4 cases). splitname("BB,", strict_mode=True) with self.assertRaises(InvalidName): splitname("BB, ", strict_mode=True) with self.assertRaises(InvalidName): splitname("BB, ~\t", strict_mode=True) with self.assertRaises(InvalidName): splitname(", ~\t", strict_mode=True) with self.assertRaises(InvalidName): # Too many sections. splitname("AA, BB, CC, DD", strict_mode=True) with self.assertRaises( InvalidName ): # Unterminated opening brace (x3). splitname("AA {BB CC", strict_mode=True) with self.assertRaises(InvalidName): splitname("AA {{{BB CC", strict_mode=True) with self.assertRaises(InvalidName): splitname("AA {{{BB} CC}", strict_mode=True) with self.assertRaises(InvalidName): # Unmatched closing brace (x3). splitname("AA BB CC}", strict_mode=True) with self.assertRaises(InvalidName): splitname("AA BB CC}}}", strict_mode=True) with self.assertRaises(InvalidName): splitname("{AA {BB CC}}}", strict_mode=True) # Test strict mode off for trailing comma. expected = {'first': [], 'von': [], 'last': ["BB"], 'jr': []} result = splitname("BB,", strict_mode=False) self.assertEqual( result, expected, msg="Invalid output for trailing comma with strict mode off", ) result = splitname("BB, ", strict_mode=False) self.assertEqual( result, expected, msg="Invalid output for trailing comma with strict mode off", ) result = splitname("BB, ~\t ", strict_mode=False) self.assertEqual( result, expected, msg="Invalid output for trailing comma with strict mode off", ) expected = {} result = splitname(", ~\t", strict_mode=False) self.assertEqual( result, expected, msg="Invalid output for trailing comma with strict mode off", ) # Test strict mode off for too many sections. expected = { 'first': ["CC", "DD"], 'von': [], 'last': ["AA"], 'jr': ["BB"], } result = splitname("AA, BB, CC, DD", strict_mode=False) self.assertEqual( result, expected, msg="Invalid output for too many sections with strict mode off", ) # Test strict mode off for an unterminated opening brace. result = splitname("AA {BB CC", strict_mode=False) expected = {'first': ["AA"], 'von': [], 'last': ["{BB CC}"], 'jr': []} self.assertEqual( result, expected, msg="Invalid output for unterminated opening brace with strict mode off", ) result = splitname("AA {{{BB CC", strict_mode=False) expected = { 'first': ["AA"], 'von': [], 'last': ["{{{BB CC}}}"], 'jr': [], } self.assertEqual( result, expected, msg="Invalid output for unterminated opening brace with strict mode off", ) result = splitname("AA {{{BB} CC}", strict_mode=False) expected = { 'first': ["AA"], 'von': [], 'last': ["{{{BB} CC}}"], 'jr': [], } self.assertEqual( result, expected, msg="Invalid output for unterminated opening brace with strict mode off", ) # Test strict mode off for an unmatched closing brace. result = splitname("AA BB CC}", strict_mode=False) expected = { 'first': ["AA", "BB"], 'von': [], 'last': ["{CC}"], 'jr': [], } self.assertEqual( result, expected, msg="Invalid output for unmatched closing brace with strict mode off", ) result = splitname("AA BB CC}}}", strict_mode=False) expected = { 'first': ["AA", "BB"], 'von': [], 'last': ["{{{CC}}}"], 'jr': [], } self.assertEqual( result, expected, msg="Invalid output for unmatched closing brace with strict mode off", ) result = splitname("{AA {BB CC}}}", strict_mode=False) expected = { 'first': [], 'von': [], 'last': ["{{AA {BB CC}}}"], 'jr': [], } self.assertEqual( result, expected, msg="Invalid output for unmatched closing brace with strict mode off", ) # Test it handles commas at higher brace levels. result = splitname("CC, dd, {AA, BB}") expected = { 'first': ["{AA, BB}"], 'von': [], 'last': ["CC"], 'jr': ["dd"], } self.assertEqual( result, expected, msg="Invalid output for braced commas" ) def test_splitname_cases(self): """Test customization.splitname() vs output from BibTeX """ for name, expected in splitname_test_cases: result = splitname(name) self.assertEqual( result, expected, msg="Input name: {0}".format(name) ) splitname_test_cases = ( ( r'Per Brinch Hansen', {'first': ['Per', 'Brinch'], 'von': [], 'last': ['Hansen'], 'jr': []}, ), ( r'Brinch Hansen, Per', {'first': ['Per'], 'von': [], 'last': ['Brinch', 'Hansen'], 'jr': []}, ), ( r'Brinch Hansen,, Per', {'first': ['Per'], 'von': [], 'last': ['Brinch', 'Hansen'], 'jr': []}, ), ( r"Charles Louis Xavier Joseph de la Vall{\'e}e Poussin", { 'first': ['Charles', 'Louis', 'Xavier', 'Joseph'], 'von': ['de', 'la'], 'last': [r'Vall{\'e}e', 'Poussin'], 'jr': [], }, ), ( r'D[onald] E. Knuth', {'first': ['D[onald]', 'E.'], 'von': [], 'last': ['Knuth'], 'jr': []}, ), ( r'A. {Delgado de Molina}', { 'first': ['A.'], 'von': [], 'last': ['{Delgado de Molina}'], 'jr': [], }, ), ( r"M. Vign{\'e}", {'first': ['M.'], 'von': [], 'last': [r"Vign{\'e}"], 'jr': []}, ), ############################################################################### # # Test cases from # http://maverick.inria.fr/~Xavier.Decoret/resources/xdkbibtex/bibtex_summary.html # ############################################################################### (r'AA BB', {'first': ['AA'], 'von': [], 'last': ['BB'], 'jr': []}), (r'AA', {'first': [], 'von': [], 'last': ['AA'], 'jr': []}), (r'AA bb', {'first': ['AA'], 'von': [], 'last': ['bb'], 'jr': []}), (r'aa', {'first': [], 'von': [], 'last': ['aa'], 'jr': []}), (r'AA bb CC', {'first': ['AA'], 'von': ['bb'], 'last': ['CC'], 'jr': []}), ( r'AA bb CC dd EE', {'first': ['AA'], 'von': ['bb', 'CC', 'dd'], 'last': ['EE'], 'jr': []}, ), ( r'AA 1B cc dd', {'first': ['AA', '1B'], 'von': ['cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA 1b cc dd', {'first': ['AA'], 'von': ['1b', 'cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA {b}B cc dd', {'first': ['AA', '{b}B'], 'von': ['cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA {b}b cc dd', {'first': ['AA'], 'von': ['{b}b', 'cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA {B}b cc dd', {'first': ['AA'], 'von': ['{B}b', 'cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA {B}B cc dd', {'first': ['AA', '{B}B'], 'von': ['cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA \BB{b} cc dd', {'first': ['AA', r'\BB{b}'], 'von': ['cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA \bb{b} cc dd', {'first': ['AA'], 'von': [r'\bb{b}', 'cc'], 'last': ['dd'], 'jr': []}, ), ( r'AA {bb} cc DD', {'first': ['AA', '{bb}'], 'von': ['cc'], 'last': ['DD'], 'jr': []}, ), ( r'AA bb {cc} DD', {'first': ['AA'], 'von': ['bb'], 'last': ['{cc}', 'DD'], 'jr': []}, ), ( r'AA {bb} CC', {'first': ['AA', '{bb}'], 'von': [], 'last': ['CC'], 'jr': []}, ), (r'bb CC, AA', {'first': ['AA'], 'von': ['bb'], 'last': ['CC'], 'jr': []}), (r'bb CC, aa', {'first': ['aa'], 'von': ['bb'], 'last': ['CC'], 'jr': []}), ( r'bb CC dd EE, AA', {'first': ['AA'], 'von': ['bb', 'CC', 'dd'], 'last': ['EE'], 'jr': []}, ), (r'bb, AA', {'first': ['AA'], 'von': [], 'last': ['bb'], 'jr': []}), ( r'bb CC,XX, AA', {'first': ['AA'], 'von': ['bb'], 'last': ['CC'], 'jr': ['XX']}, ), ( r'bb CC,xx, AA', {'first': ['AA'], 'von': ['bb'], 'last': ['CC'], 'jr': ['xx']}, ), (r'BB,, AA', {'first': ['AA'], 'von': [], 'last': ['BB'], 'jr': []}), ( r"Paul \'Emile Victor", { 'first': ['Paul', r"\'Emile"], 'von': [], 'last': ['Victor'], 'jr': [], }, ), ( r"Paul {\'E}mile Victor", { 'first': ['Paul', r"{\'E}mile"], 'von': [], 'last': ['Victor'], 'jr': [], }, ), ( r"Paul \'emile Victor", {'first': ['Paul'], 'von': [r"\'emile"], 'last': ['Victor'], 'jr': []}, ), ( r"Paul {\'e}mile Victor", { 'first': ['Paul'], 'von': [r"{\'e}mile"], 'last': ['Victor'], 'jr': [], }, ), ( r"Victor, Paul \'Emile", { 'first': ['Paul', r"\'Emile"], 'von': [], 'last': ['Victor'], 'jr': [], }, ), ( r"Victor, Paul {\'E}mile", { 'first': ['Paul', r"{\'E}mile"], 'von': [], 'last': ['Victor'], 'jr': [], }, ), ( r"Victor, Paul \'emile", { 'first': ['Paul', r"\'emile"], 'von': [], 'last': ['Victor'], 'jr': [], }, ), ( r"Victor, Paul {\'e}mile", { 'first': ['Paul', r"{\'e}mile"], 'von': [], 'last': ['Victor'], 'jr': [], }, ), ( r'Dominique Galouzeau de Villepin', { 'first': ['Dominique', 'Galouzeau'], 'von': ['de'], 'last': ['Villepin'], 'jr': [], }, ), ( r'Dominique {G}alouzeau de Villepin', { 'first': ['Dominique'], 'von': ['{G}alouzeau', 'de'], 'last': ['Villepin'], 'jr': [], }, ), ( r'Galouzeau de Villepin, Dominique', { 'first': ['Dominique'], 'von': ['Galouzeau', 'de'], 'last': ['Villepin'], 'jr': [], }, ), ############################################################################### # # Test cases from pybtex # See file /pybtex/tests/parse_name_test.py in the pybtex source. # ############################################################################### ( r'A. E. Siegman', {'first': ['A.', 'E.'], 'von': [], 'last': ['Siegman'], 'jr': []}, ), ( r'A. G. W. Cameron', { 'first': ['A.', 'G.', 'W.'], 'von': [], 'last': ['Cameron'], 'jr': [], }, ), (r'A. Hoenig', {'first': ['A.'], 'von': [], 'last': ['Hoenig'], 'jr': []}), ( r'A. J. Van Haagen', { 'first': ['A.', 'J.', 'Van'], 'von': [], 'last': ['Haagen'], 'jr': [], }, ), ( r'A. S. Berdnikov', {'first': ['A.', 'S.'], 'von': [], 'last': ['Berdnikov'], 'jr': []}, ), ( r'A. Trevorrow', {'first': ['A.'], 'von': [], 'last': ['Trevorrow'], 'jr': []}, ), ( r'Adam H. Lewenberg', {'first': ['Adam', 'H.'], 'von': [], 'last': ['Lewenberg'], 'jr': []}, ), ( r'Addison-Wesley Publishing Company', { 'first': ['Addison-Wesley', 'Publishing'], 'von': [], 'last': ['Company'], 'jr': [], }, ), ( r'Advogato (Raph Levien)', { 'first': ['Advogato', '(Raph'], 'von': [], 'last': ['Levien)'], 'jr': [], }, ), ( r'Andrea de Leeuw van Weenen', { 'first': ['Andrea'], 'von': ['de', 'Leeuw', 'van'], 'last': ['Weenen'], 'jr': [], }, ), ( r'Andreas Geyer-Schulz', {'first': ['Andreas'], 'von': [], 'last': ['Geyer-Schulz'], 'jr': []}, ), ( r'Andr{\'e} Heck', {'first': [r'Andr{\'e}'], 'von': [], 'last': ['Heck'], 'jr': []}, ), ( r'Anne Br{\"u}ggemann-Klein', { 'first': ['Anne'], 'von': [], 'last': [r'Br{\"u}ggemann-Klein'], 'jr': [], }, ), (r'Anonymous', {'first': [], 'von': [], 'last': ['Anonymous'], 'jr': []}), (r'B. Beeton', {'first': ['B.'], 'von': [], 'last': ['Beeton'], 'jr': []}), ( r'B. Hamilton Kelly', {'first': ['B.', 'Hamilton'], 'von': [], 'last': ['Kelly'], 'jr': []}, ), ( r'B. V. Venkata Krishna Sastry', { 'first': ['B.', 'V.', 'Venkata', 'Krishna'], 'von': [], 'last': ['Sastry'], 'jr': [], }, ), ( r'Benedict L{\o}fstedt', {'first': ['Benedict'], 'von': [], 'last': [r'L{\o}fstedt'], 'jr': []}, ), ( r'Bogus{\l}aw Jackowski', {'first': ['Bogus{\l}aw'], 'von': [], 'last': ['Jackowski'], 'jr': []}, ), ( r'Christina A. L.\ Thiele', { 'first': ['Christina', 'A.', 'L.\\'], 'von': [], 'last': ['Thiele'], 'jr': [], }, ), ( r"D. Men'shikov", {'first': ['D.'], 'von': [], 'last': ["Men'shikov"], 'jr': []}, ), ( r'Darko \v{Z}ubrini{\'c}', { 'first': ['Darko'], 'von': [], 'last': [r'\v{Z}ubrini{\'c}'], 'jr': [], }, ), ( r'Dunja Mladeni{\'c}', {'first': ['Dunja'], 'von': [], 'last': [r'Mladeni{\'c}'], 'jr': []}, ), ( r'Edwin V. {Bell, II}', { 'first': ['Edwin', 'V.'], 'von': [], 'last': ['{Bell, II}'], 'jr': [], }, ), ( r'Frank G. {Bennett, Jr.}', { 'first': ['Frank', 'G.'], 'von': [], 'last': ['{Bennett, Jr.}'], 'jr': [], }, ), ( r'Fr{\'e}d{\'e}ric Boulanger', { 'first': [r'Fr{\'e}d{\'e}ric'], 'von': [], 'last': ['Boulanger'], 'jr': [], }, ), ( r'Ford, Jr., Henry', {'first': ['Henry'], 'von': [], 'last': ['Ford'], 'jr': ['Jr.']}, ), ( r'mr Ford, Jr., Henry', {'first': ['Henry'], 'von': ['mr'], 'last': ['Ford'], 'jr': ['Jr.']}, ), (r'Fukui Rei', {'first': ['Fukui'], 'von': [], 'last': ['Rei'], 'jr': []}), ( r'G. Gr{\"a}tzer', {'first': ['G.'], 'von': [], 'last': [r'Gr{\"a}tzer'], 'jr': []}, ), ( r'George Gr{\"a}tzer', {'first': ['George'], 'von': [], 'last': [r'Gr{\"a}tzer'], 'jr': []}, ), ( r'Georgia K. M. Tobin', { 'first': ['Georgia', 'K.', 'M.'], 'von': [], 'last': ['Tobin'], 'jr': [], }, ), ( r'Gilbert van den Dobbelsteen', { 'first': ['Gilbert'], 'von': ['van', 'den'], 'last': ['Dobbelsteen'], 'jr': [], }, ), ( r'Gy{\"o}ngyi Bujdos{\'o}', { 'first': [r'Gy{\"o}ngyi'], 'von': [], 'last': [r'Bujdos{\'o}'], 'jr': [], }, ), ( r'Helmut J{\"u}rgensen', {'first': ['Helmut'], 'von': [], 'last': [r'J{\"u}rgensen'], 'jr': []}, ), ( r'Herbert Vo{\ss}', {'first': ['Herbert'], 'von': [], 'last': ['Vo{\ss}'], 'jr': []}, ), ( r"H{\'a}n Th{\^e}\llap{\raise 0.5ex\hbox{\'{\relax}}} Th{\'a}nh", { 'first': [ r'H{\'a}n', r"Th{\^e}\llap{\raise 0.5ex\hbox{\'{\relax}}}", ], 'von': [], 'last': [r"Th{\'a}nh"], 'jr': [], }, ), ( r"H{\`a}n Th\^e\llap{\raise0.5ex\hbox{\'{\relax}}} Th{\`a}nh", { 'first': [r'H{\`a}n', r"Th\^e\llap{\raise0.5ex\hbox{\'{\relax}}}"], 'von': [], 'last': [r"Th{\`a}nh"], 'jr': [], }, ), ( r'J. Vesel{\'y}', {'first': ['J.'], 'von': [], 'last': [r'Vesel{\'y}'], 'jr': []}, ), ( r'Javier Rodr\'{\i}guez Laguna', { 'first': ['Javier', r'Rodr\'{\i}guez'], 'von': [], 'last': ['Laguna'], 'jr': [], }, ), ( r'Ji\v{r}\'{\i} Vesel{\'y}', { 'first': [r'Ji\v{r}\'{\i}'], 'von': [], 'last': [r'Vesel{\'y}'], 'jr': [], }, ), ( r'Ji\v{r}\'{\i} Zlatu{\v{s}}ka', { 'first': [r'Ji\v{r}\'{\i}'], 'von': [], 'last': [r'Zlatu{\v{s}}ka'], 'jr': [], }, ), ( r'Ji\v{r}{\'\i} Vesel{\'y}', { 'first': [r'Ji\v{r}{\'\i}'], 'von': [], 'last': [r'Vesel{\'y}'], 'jr': [], }, ), ( r'Ji\v{r}{\'{\i}}Zlatu{\v{s}}ka', { 'first': [], 'von': [], 'last': [r'Ji\v{r}{\'{\i}}Zlatu{\v{s}}ka'], 'jr': [], }, ), ( r'Jim Hef{}feron', {'first': ['Jim'], 'von': [], 'last': ['Hef{}feron'], 'jr': []}, ), ( r'J{\"o}rg Knappen', {'first': [r'J{\"o}rg'], 'von': [], 'last': ['Knappen'], 'jr': []}, ), ( r'J{\"o}rgen L. Pind', { 'first': [r'J{\"o}rgen', 'L.'], 'von': [], 'last': ['Pind'], 'jr': [], }, ), ( r'J{\'e}r\^ome Laurens', {'first': [r'J{\'e}r\^ome'], 'von': [], 'last': ['Laurens'], 'jr': []}, ), ( r'J{{\"o}}rg Knappen', {'first': [r'J{{\"o}}rg'], 'von': [], 'last': ['Knappen'], 'jr': []}, ), ( r'K. Anil Kumar', {'first': ['K.', 'Anil'], 'von': [], 'last': ['Kumar'], 'jr': []}, ), ( r'Karel Hor{\'a}k', {'first': ['Karel'], 'von': [], 'last': [r'Hor{\'a}k'], 'jr': []}, ), ( r'Karel P\'{\i}{\v{s}}ka', { 'first': ['Karel'], 'von': [], 'last': [r'P\'{\i}{\v{s}}ka'], 'jr': [], }, ), ( r'Karel P{\'\i}{\v{s}}ka', { 'first': ['Karel'], 'von': [], 'last': [r'P{\'\i}{\v{s}}ka'], 'jr': [], }, ), ( r'Karel Skoup\'{y}', {'first': ['Karel'], 'von': [], 'last': [r'Skoup\'{y}'], 'jr': []}, ), ( r'Karel Skoup{\'y}', {'first': ['Karel'], 'von': [], 'last': [r'Skoup{\'y}'], 'jr': []}, ), ( r'Kent McPherson', {'first': ['Kent'], 'von': [], 'last': ['McPherson'], 'jr': []}, ), ( r'Klaus H{\"o}ppner', {'first': ['Klaus'], 'von': [], 'last': [r'H{\"o}ppner'], 'jr': []}, ), ( r'Lars Hellstr{\"o}m', {'first': ['Lars'], 'von': [], 'last': [r'Hellstr{\"o}m'], 'jr': []}, ), ( r'Laura Elizabeth Jackson', { 'first': ['Laura', 'Elizabeth'], 'von': [], 'last': ['Jackson'], 'jr': [], }, ), ( r'M. D{\'{\i}}az', {'first': ['M.'], 'von': [], 'last': [r'D{\'{\i}}az'], 'jr': []}, ), ( r'M/iche/al /O Searc/oid', { 'first': [r'M/iche/al', r'/O'], 'von': [], 'last': [r'Searc/oid'], 'jr': [], }, ), ( r'Marek Ry{\'c}ko', {'first': ['Marek'], 'von': [], 'last': [r'Ry{\'c}ko'], 'jr': []}, ), ( r'Marina Yu. Nikulina', { 'first': ['Marina', 'Yu.'], 'von': [], 'last': ['Nikulina'], 'jr': [], }, ), ( r'Max D{\'{\i}}az', {'first': ['Max'], 'von': [], 'last': [r'D{\'{\i}}az'], 'jr': []}, ), ( r'Merry Obrecht Sawdey', { 'first': ['Merry', 'Obrecht'], 'von': [], 'last': ['Sawdey'], 'jr': [], }, ), ( r'Miroslava Mis{\'a}kov{\'a}', { 'first': ['Miroslava'], 'von': [], 'last': [r'Mis{\'a}kov{\'a}'], 'jr': [], }, ), ( r'N. A. F. M. Poppelier', { 'first': ['N.', 'A.', 'F.', 'M.'], 'von': [], 'last': ['Poppelier'], 'jr': [], }, ), ( r'Nico A. F. M. Poppelier', { 'first': ['Nico', 'A.', 'F.', 'M.'], 'von': [], 'last': ['Poppelier'], 'jr': [], }, ), ( r'Onofrio de Bari', {'first': ['Onofrio'], 'von': ['de'], 'last': ['Bari'], 'jr': []}, ), ( r'Pablo Rosell-Gonz{\'a}lez', { 'first': ['Pablo'], 'von': [], 'last': [r'Rosell-Gonz{\'a}lez'], 'jr': [], }, ), ( r'Paco La Bruna', {'first': ['Paco', 'La'], 'von': [], 'last': ['Bruna'], 'jr': []}, ), ( r'Paul Franchi-Zannettacci', { 'first': ['Paul'], 'von': [], 'last': ['Franchi-Zannettacci'], 'jr': [], }, ), ( r'Pavel \v{S}eve\v{c}ek', { 'first': ['Pavel'], 'von': [], 'last': [r'\v{S}eve\v{c}ek'], 'jr': [], }, ), ( r'Petr Ol{\v{s}}ak', {'first': ['Petr'], 'von': [], 'last': [r'Ol{\v{s}}ak'], 'jr': []}, ), ( r'Petr Ol{\v{s}}{\'a}k', {'first': ['Petr'], 'von': [], 'last': [r'Ol{\v{s}}{\'a}k'], 'jr': []}, ), ( r'Primo\v{z} Peterlin', {'first': [r'Primo\v{z}'], 'von': [], 'last': ['Peterlin'], 'jr': []}, ), ( r'Prof. Alban Grimm', {'first': ['Prof.', 'Alban'], 'von': [], 'last': ['Grimm'], 'jr': []}, ), ( r'P{\'e}ter Husz{\'a}r', { 'first': [r'P{\'e}ter'], 'von': [], 'last': [r'Husz{\'a}r'], 'jr': [], }, ), ( r'P{\'e}ter Szab{\'o}', {'first': [r'P{\'e}ter'], 'von': [], 'last': [r'Szab{\'o}'], 'jr': []}, ), ( r'Rafa{\l}\.Zbikowski', {'first': [], 'von': [], 'last': [r'Rafa{\l}\.Zbikowski'], 'jr': []}, ), ( r'Rainer Sch{\"o}pf', {'first': ['Rainer'], 'von': [], 'last': [r'Sch{\"o}pf'], 'jr': []}, ), ( r'T. L. (Frank) Pappas', { 'first': ['T.', 'L.', '(Frank)'], 'von': [], 'last': ['Pappas'], 'jr': [], }, ), ( r'TUG 2004 conference', { 'first': ['TUG', '2004'], 'von': [], 'last': ['conference'], 'jr': [], }, ), ( r'TUG {\sltt DVI} Driver Standards Committee', { 'first': ['TUG', '{\sltt DVI}', 'Driver', 'Standards'], 'von': [], 'last': ['Committee'], 'jr': [], }, ), ( r'TUG {\sltt xDVIx} Driver Standards Committee', { 'first': ['TUG'], 'von': ['{\sltt xDVIx}'], 'last': ['Driver', 'Standards', 'Committee'], 'jr': [], }, ), ( r'University of M{\"u}nster', { 'first': ['University'], 'von': ['of'], 'last': [r'M{\"u}nster'], 'jr': [], }, ), ( r'Walter van der Laan', { 'first': ['Walter'], 'von': ['van', 'der'], 'last': ['Laan'], 'jr': [], }, ), ( r'Wendy G. McKay', {'first': ['Wendy', 'G.'], 'von': [], 'last': ['McKay'], 'jr': []}, ), ( r'Wendy McKay', {'first': ['Wendy'], 'von': [], 'last': ['McKay'], 'jr': []}, ), ( r'W{\l}odek Bzyl', {'first': [r'W{\l}odek'], 'von': [], 'last': ['Bzyl'], 'jr': []}, ), ( r'\LaTeX Project Team', { 'first': [r'\LaTeX', 'Project'], 'von': [], 'last': ['Team'], 'jr': [], }, ), ( r'\rlap{Lutz Birkhahn}', {'first': [], 'von': [], 'last': [r'\rlap{Lutz Birkhahn}'], 'jr': []}, ), ( r'{Jim Hef{}feron}', {'first': [], 'von': [], 'last': ['{Jim Hef{}feron}'], 'jr': []}, ), ( r'{Kristoffer H\o{}gsbro Rose}', { 'first': [], 'von': [], 'last': ['{Kristoffer H\o{}gsbro Rose}'], 'jr': [], }, ), ( r'{TUG} {Working} {Group} on a {\TeX} {Directory} {Structure}', { 'first': ['{TUG}', '{Working}', '{Group}'], 'von': ['on', 'a'], 'last': [r'{\TeX}', '{Directory}', '{Structure}'], 'jr': [], }, ), ( r'{The \TUB{} Team}', {'first': [], 'von': [], 'last': [r'{The \TUB{} Team}'], 'jr': []}, ), ( r'{\LaTeX} project team', { 'first': [r'{\LaTeX}'], 'von': ['project'], 'last': ['team'], 'jr': [], }, ), ( r'{\NTG{} \TeX{} future working group}', { 'first': [], 'von': [], 'last': [r'{\NTG{} \TeX{} future working group}'], 'jr': [], }, ), ( r'{{\LaTeX\,3} Project Team}', { 'first': [], 'von': [], 'last': [r'{{\LaTeX\,3} Project Team}'], 'jr': [], }, ), ( r'Johansen Kyle, Derik Mamania M.', { 'first': ['Derik', 'Mamania', 'M.'], 'von': [], 'last': ['Johansen', 'Kyle'], 'jr': [], }, ), ( r"Johannes Adam Ferdinand Alois Josef Maria Marko d'Aviano Pius von und zu Liechtenstein", { 'first': [ 'Johannes', 'Adam', 'Ferdinand', 'Alois', 'Josef', 'Maria', 'Marko', ], 'von': ["d'Aviano", 'Pius', 'von', 'und', 'zu'], 'last': ['Liechtenstein'], 'jr': [], }, ), ( r"Brand\~{a}o, F", {'first': ['F'], 'von': [], 'last': ['Brand\\', '{a}o'], 'jr': []}, ), ) if __name__ == '__main__': unittest.main()
28,941
10,270
# schema for SQL database from data import app, db class HNuser(db.Model): """ SQL database class """ username = db.Column(db.String(100), primary_key=True) post_id = db.Column(db.Integer) salty_rank = db.Column(db.Float, nullable=False) salty_comments = db.Column(db.Integer, nullable=False) # comments_total = db.Column(db.Integer, nullable=False) def __repr__(self) return f"User {self.username} -- Salty Ranking: {self.salty_rank}" def salty_hackers(self): """return user information in Json format """ return { "username" : self.username, "date" : self.date, "salty_rank" : self.salty_rank, "salty_comments" : self.salty_comments, } class Comments(db.Model): comment_id = db.Column(db.BigInteger, primary_key=True) username = db.Column(db.String(100), db.ForeignKey('user.username')) text = db.Column(db.String(3000)) date = db.Column(db.BigInteger) def __repr__(self): return f"User {self.username} -- Comment: {self.text}" def salty_comments(self): """ returns comments in JSON format """ return { "comment_id" : self.comment_id, "username" : self.username, "text" : self.text, "date" : self.date "" }
1,349
436
import re import logging logger = logging.getLogger(__name__) def get_sorted_pair(a, b): # ensure citation pair is always in same order if a > b: return (a, b) else: return (b, a) def to_label(t, labels): if t in labels: return t else: return 'other' def normalize_title(t): if t: t = t.replace('.', ' ').replace('-', ' ').strip().lower() #t = re.sub(r'\W+', '', t) return t def normalize_section(title): if title: return re.sub(r'[\.0-9]', '', title. strip() \ .lower() \ .replace('conclusions', 'conclusion') \ .replace('methodology', 'method') \ .replace('methods', 'method') \ .replace('related works', 'related work') \ .replace('models', 'model') \ .replace('datasets', 'dataset') \ .replace('our ', '') \ .replace('evaluations', 'evaluation') \ .replace('experiments', 'experiment') ).strip() # .replace('conclusion and future perspectives', 'conclusion')\ # .replace('materials and methods', 'methods') def get_text_from_doc(doc) -> str: """ Build document text from title + abstract :param doc: S2 paper :return: Document text """ text = '' if 'title' in doc: text += doc['title'] if doc['abstract']: text += '\n' + doc['abstract'] return text def get_text_from_doc_id(doc_id: str, doc_index) -> str: """ Build document text from title + abstract :param doc_id: S2-id :param doc_index: S2-id to S2-paper data :return: Document text """ if doc_id in doc_index: return get_text_from_doc(doc_index[doc_id]) else: raise ValueError(f'Document not found in index: {doc_id}') # resolve 'and' titles and filter for out-of-index docs def resolve_and_sect_titles(items, doc_index=None): for from_s2_id, to_s2_id, sect_generic, sect_title, sect_marker in items: if doc_index and (from_s2_id not in doc_index or to_s2_id not in doc_index): # One of the IDs does not exist in document index continue sect_title = normalize_section(sect_title) if sect_title: # Resolve combined sections for t in sect_title.split(' and '): if t: yield (from_s2_id, to_s2_id, t, sect_marker)
2,608
798
from django.test import TestCase from django.utils import timezone from .models import Episode class PodCastsTests(TestCase): def setUp(self): self.episode = Episode.objects.create( title="My Awesome Podcast Episode", description="Look mom, I made it!", pub_date=timezone.now(), link="https://myawesomeshow.com", image="https://image.myawesomeshow.com", podcast_name="My Python Podcast", guid="de194720-7b4c-49e2-a05f-432436d3fetr", ) def test_episode_content(self): self.assertEqual(self.episode.description, "Look mom, I made it!") self.assertEqual(self.episode.link, "https://myawesomeshow.com") self.assertEqual( self.episode.guid, "de194720-7b4c-49e2-a05f-432436d3fetr" ) def test_episode_str_representation(self): self.assertEqual( str(self.episode), "My Python Podcast: My Awesome Podcast Episode" )
995
350
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. """add granularity to charts where missing Revision ID: 070c043f2fdb Revises: 41ce8799acc3 Create Date: 2021-02-04 09:34:13.608891 """ # revision identifiers, used by Alembic. revision = "070c043f2fdb" down_revision = "41ce8799acc3" import json from alembic import op from sqlalchemy import and_, Boolean, Column, Integer, String, Text from sqlalchemy.ext.declarative import declarative_base from superset import db Base = declarative_base() class Slice(Base): __tablename__ = "slices" id = Column(Integer, primary_key=True) params = Column(Text) datasource_id = Column(Integer) datasource_type = Column(String(200)) class SqlaTable(Base): __tablename__ = "tables" id = Column(Integer, primary_key=True) main_dttm_col = Column(String(250)) class TableColumn(Base): __tablename__ = "table_columns" id = Column(Integer, primary_key=True) table_id = Column(Integer) is_dttm = Column(Boolean) column_name = Column(String(255)) def upgrade(): """ Adds the granularity param to charts without it populated. This is required for time range filtering to work properly. Uses the following approach: - Find all charts without a granularity or granularity_sqla param. - Get the dataset that backs the chart. - If the dataset has the main dttm column set, use it. - Otherwise, find all the dttm columns in the dataset and use the first one (this matches the behavior of Explore view on the frontend) - If no dttm columns exist in the dataset, don't change the chart. """ bind = op.get_bind() session = db.Session(bind=bind) slices_changed = 0 for slc in ( session.query(Slice) .filter( and_( Slice.datasource_type == "table", Slice.params.notlike('%"granularity%') ) ) .all() ): try: params = json.loads(slc.params) if "granularity" in params or "granularity_sqla" in params: continue table = session.query(SqlaTable).get(slc.datasource_id) if not table: continue if table.main_dttm_col: params["granularity"] = table.main_dttm_col slc.params = json.dumps(params, sort_keys=True) print(f"Set granularity for slice {slc.id} to {table.main_dttm_col}") slices_changed += 1 continue table_columns = ( session.query(TableColumn) .filter(TableColumn.table_id == table.id) .filter(TableColumn.is_dttm == True) .all() ) if len(table_columns): params["granularity"] = table_columns[0].column_name slc.params = json.dumps(params, sort_keys=True) print( f"Set granularity for slice {slc.id} to {table_columns[0].column_name}" ) slices_changed += 1 except Exception as e: print(e) print(f"Parsing params for slice {slc.id} failed.") pass print(f"{slices_changed} slices altered") session.commit() session.close() def downgrade(): """ It's impossible to downgrade this migration. """ pass
4,114
1,273
# A 以上 B 以下の整数のうち、回文数となるものの個数を求めてください。 # ただし、回文数とは、先頭に 0 をつけない 10 進表記を文字列として見たとき、 # 前から読んでも後ろから読んでも同じ文字列となるような正の整数のことを指します。 # 文字列を逆順にして、同じ桁数の値をみる a, b = map(int, input().split()) cnt = 0 for i in range(a, b+1): s = str(i) s_r = s[::-1] n = int(len(str(s))/2) if s[: n] == s_r[:n]: cnt += 1 print(cnt)
325
254
bl_info = { "name": "Import Fromsoft FLVER models", "description": "Import models from various Fromsoft games such as Dark Souls", "author": "Eliza Velasquez", "version": (0, 1, 0), "blender": (2, 80, 0), "category": "Import-Export", "location": "File > Import", "warning": "", "support": "COMMUNITY", "wiki_url": "", # TODO: wiki url "tracker_url": "", # TODO: tracker url } _submodules = { "importer", "flver", "reader", } # Reload submodules on addon reload if "bpy" in locals(): import importlib for submodule in _submodules: if submodule in locals(): importlib.reload(locals()[submodule]) import bpy from . import importer from bpy_extras.io_utils import ImportHelper from bpy.props import StringProperty, BoolProperty class FlverImporter(bpy.types.Operator, ImportHelper): bl_idname = "import_scene.flver" bl_label = "Fromsoft (.flver)" filter_glob = StringProperty(default="*.flver", options={"HIDDEN"}) transpose_y_and_z = BoolProperty( name="Transpose Y and Z axes", description=("This will correct the orientation of the model. " + "Rarely necessary to disable."), default=True) import_skeleton = BoolProperty( name="Import skeleton", description=("Disable to prevent the creation of an Armature " + "and corresponding vertex groups."), default=True) connect_bones = BoolProperty( name="Connect bones", description=( "Disable to import disjointed bones rotated about their " + "original Euler angles. This may be potentially desireable " "for authoring derivative FLVER files."), default=True) def execute(self, context): importer.run(context=context, path=self.filepath, transpose_y_and_z=self.transpose_y_and_z, import_skeleton=self.import_skeleton, connect_bones=self.connect_bones) return {"FINISHED"} def menu_import(self, context): self.layout.operator(FlverImporter.bl_idname) def register(): bpy.utils.register_class(FlverImporter) bpy.types.TOPBAR_MT_file_import.append(menu_import) def unregister(): bpy.types.TOPBAR_MT_file_import.remove(menu_import) bpy.utils.unregister_class(FlverImporter)
2,499
805
# -*- coding: utf-8 -*- # https://github.com/Kodi-vStream/venom-xbmc-addons from resources.lib.gui.contextElement import cContextElement from resources.lib.gui.guiElement import cGuiElement from resources.lib.db import cDb from resources.lib.handler.outputParameterHandler import cOutputParameterHandler from resources.lib.handler.inputParameterHandler import cInputParameterHandler from resources.lib.handler.pluginHandler import cPluginHandler from resources.lib.parser import cParser from resources.lib.util import cUtil, QuotePlus from resources.lib.comaddon import listitem, addon, dialog, isKrypton, window, xbmc import re, xbmcplugin class cGui(): SITE_NAME = 'cGui' CONTENT = 'files' searchResults = [] # modif 22/06 listing = [] ADDON = addon() if isKrypton(): CONTENT = 'addons' def addMovie(self, sId, sFunction, sLabel, sIcon, sThumbnail, sDesc, oOutputParameterHandler = ''): cGui.CONTENT = 'movies' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setPoster(sThumbnail) oGuiElement.setMeta(1) oGuiElement.setDescription(sDesc) #oGuiElement.setMovieFanart() oGuiElement.setCat(1) if oOutputParameterHandler.getValue('sMovieTitle'): sTitle = oOutputParameterHandler.getValue('sMovieTitle') oGuiElement.setFileName(sTitle) try: self.addFolder(oGuiElement, oOutputParameterHandler) except: pass # Coffret et integrale de films def addMoviePack(self, sId, sFunction, sLabel, sIcon, sThumbnail, sDesc, oOutputParameterHandler = ''): cGui.CONTENT = 'movies' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setPoster(sThumbnail) oGuiElement.setMeta(3) oGuiElement.setDescription(sDesc) #oGuiElement.setMovieFanart() oGuiElement.setCat(1) if oOutputParameterHandler.getValue('sMovieTitle'): sTitle = oOutputParameterHandler.getValue('sMovieTitle') oGuiElement.setFileName(sTitle) try: self.addFolder(oGuiElement, oOutputParameterHandler) except: pass def addTV(self, sId, sFunction, sLabel, sIcon, sThumbnail, sDesc, oOutputParameterHandler = ''): cGui.CONTENT = 'tvshows' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setPoster(sThumbnail) oGuiElement.setMeta(2) oGuiElement.setDescription(sDesc) #oGuiElement.setTvFanart() oGuiElement.setCat(2) if oOutputParameterHandler.getValue('sMovieTitle'): sTitle = oOutputParameterHandler.getValue('sMovieTitle') oGuiElement.setFileName(sTitle) try: self.addFolder(oGuiElement, oOutputParameterHandler) except: pass def addMisc(self, sId, sFunction, sLabel, sIcon, sThumbnail, sDesc, oOutputParameterHandler = ''): #cGui.CONTENT = 'movies' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setDescription(sDesc) #oGuiElement.setPoster(sThumbnail) oGuiElement.setMeta(0) #oGuiElement.setDirFanart(sIcon) oGuiElement.setCat(5) if oOutputParameterHandler.getValue('sMovieTitle'): sTitle = oOutputParameterHandler.getValue('sMovieTitle') oGuiElement.setFileName(sTitle) self.createContexMenuWatch(oGuiElement, oOutputParameterHandler) #self.createContexMenuinfo(oGuiElement, oOutputParameterHandler) self.createContexMenuFav(oGuiElement, oOutputParameterHandler) try: self.addFolder(oGuiElement, oOutputParameterHandler) except: pass #non utiliser le 18/04 #def addFav(self, sId, sFunction, sLabel, sIcon, sThumbnail, fanart, oOutputParameterHandler = ''): #cGui.CONTENT = 'files' #oGuiElement = cGuiElement() #oGuiElement.setSiteName(sId) #oGuiElement.setFunction(sFunction) #oGuiElement.setTitle(sLabel) #oGuiElement.setIcon(sIcon) #oGuiElement.setMeta(0) #oGuiElement.setThumbnail(sThumbnail) #oGuiElement.setFanart(fanart) #self.createContexMenuDelFav(oGuiElement, oOutputParameterHandler) #self.addFolder(oGuiElement, oOutputParameterHandler) def addLink(self, sId, sFunction, sLabel, sThumbnail, sDesc, oOutputParameterHandler = ''): cGui.CONTENT = 'files' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) #oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setPoster(sThumbnail) oGuiElement.setDescription(sDesc) oGuiElement.setMeta(0) #oGuiElement.setDirFanart('') oInputParameterHandler = cInputParameterHandler() sCat = oInputParameterHandler.getValue('sCat') if sCat: oGuiElement.setCat(sCat) try: self.addFolder(oGuiElement, oOutputParameterHandler) except: pass def addDir(self, sId, sFunction, sLabel, sIcon, oOutputParameterHandler = ''): oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(oGuiElement.getIcon()) oGuiElement.setMeta(0) #oGuiElement.setDirFanart(sIcon) oOutputParameterHandler.addParameter('sFav', sFunction) # context parametre if isKrypton(): self.createContexMenuSettings(oGuiElement, oOutputParameterHandler) try: self.addFolder(oGuiElement, oOutputParameterHandler) except: pass def addNext(self, sId, sFunction, sLabel, oOutputParameterHandler): oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon('next.png') oGuiElement.setThumbnail(oGuiElement.getIcon()) oGuiElement.setMeta(0) #oGuiElement.setDirFanart('next.png') oGuiElement.setCat(5) self.createContexMenuPageSelect(oGuiElement, oOutputParameterHandler) self.createContexMenuFav(oGuiElement, oOutputParameterHandler) self.addFolder(oGuiElement, oOutputParameterHandler) # utiliser oGui.addText(SITE_IDENTIFIER) def addNone(self, sId): return self.addText(sId) def addText(self, sId, sLabel = '', sIcon = 'none.png'): # Pas de texte lors des recherches globales if window(10101).getProperty('search') == 'true': return oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction('DoNothing') if not sLabel: sLabel = self.ADDON.VSlang(30204) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setThumbnail(oGuiElement.getIcon()) oGuiElement.setMeta(0) oOutputParameterHandler = cOutputParameterHandler() self.addFolder(oGuiElement, oOutputParameterHandler) # non utiliser depuis le 22/04 def addMovieDB(self, sId, sFunction, sLabel, sIcon, sThumbnail, sFanart, oOutputParameterHandler = ''): cGui.CONTENT = 'movies' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setMeta(1) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setFanart(sFanart) oGuiElement.setCat(7) if oOutputParameterHandler.getValue('sMovieTitle'): sTitle = oOutputParameterHandler.getValue('sMovieTitle') oGuiElement.setFileName(sTitle) self.addFolder(oGuiElement, oOutputParameterHandler) # non utiliser 22/04 def addTVDB(self, sId, sFunction, sLabel, sIcon, sThumbnail, sFanart, oOutputParameterHandler = ''): cGui.CONTENT = 'tvshows' oGuiElement = cGuiElement() oGuiElement.setSiteName(sId) oGuiElement.setFunction(sFunction) oGuiElement.setTitle(sLabel) oGuiElement.setIcon(sIcon) oGuiElement.setMeta(2) oGuiElement.setThumbnail(sThumbnail) oGuiElement.setFanart(sFanart) oGuiElement.setCat(7) if oOutputParameterHandler.getValue('sMovieTitle'): sTitle = oOutputParameterHandler.getValue('sMovieTitle') oGuiElement.setFileName(sTitle) self.addFolder(oGuiElement, oOutputParameterHandler) # afficher les liens non playable def addFolder(self, oGuiElement, oOutputParameterHandler = '', _isFolder = True): # recherche append les reponses if window(10101).getProperty('search') == 'true': import copy cGui.searchResults.append({'guiElement': oGuiElement, 'params': copy.deepcopy(oOutputParameterHandler)}) return # Des infos a rajouter ? params = { 'siteUrl': oGuiElement.setSiteUrl, # indispensable 'sTmdbId': oGuiElement.setTmdbId, 'sImbdId': oGuiElement.setImdbId, # inutile ? 'sYear': oGuiElement.setYear, } for sParam, callback in params.iteritems(): value = oOutputParameterHandler.getValue(sParam) if value: callback(value) oListItem = self.createListItem(oGuiElement) oListItem.setProperty('IsPlayable', 'false') # affiche tag HD # if '1080' in oGuiElement.getTitle(): # oListItem.addStreamInfo('video', {'aspect': '1.78', 'width': 1920, 'height': 1080}) # elif '720' in oGuiElement.getTitle(): # oListItem.addStreamInfo('video', {'aspect': '1.50', 'width': 1280, 'height': 720}) # elif '2160'in oGuiElement.getTitle(): # oListItem.addStreamInfo('video', {'aspect': '1.78', 'width': 3840, 'height': 2160}) # oListItem.addStreamInfo('audio', {'language': 'fr'}) # if oGuiElement.getMeta(): # oOutputParameterHandler.addParameter('sMeta', oGuiElement.getMeta()) if oGuiElement.getCat(): oOutputParameterHandler.addParameter('sCat', oGuiElement.getCat()) sItemUrl = self.__createItemUrl(oGuiElement, oOutputParameterHandler) oOutputParameterHandler.addParameter('sTitleWatched', oGuiElement.getTitleWatched()) # new context prend en charge les metas if oGuiElement.getMeta() > 0: if cGui.CONTENT == 'movies': self.createContexMenuWatch(oGuiElement, oOutputParameterHandler) self.createContexMenuFav(oGuiElement, oOutputParameterHandler) self.createContexMenuinfo(oGuiElement, oOutputParameterHandler) self.createContexMenuba(oGuiElement, oOutputParameterHandler) if self.ADDON.getSetting('bstoken') != '': self.createContexMenuTrakt(oGuiElement, oOutputParameterHandler) if self.ADDON.getSetting('tmdb_account') != '': self.createContexMenuTMDB(oGuiElement, oOutputParameterHandler) self.createContexMenuSimil(oGuiElement, oOutputParameterHandler) elif cGui.CONTENT == 'tvshows': self.createContexMenuWatch(oGuiElement, oOutputParameterHandler) self.createContexMenuFav(oGuiElement, oOutputParameterHandler) self.createContexMenuinfo(oGuiElement, oOutputParameterHandler) self.createContexMenuba(oGuiElement, oOutputParameterHandler) if self.ADDON.getSetting('bstoken') != '': self.createContexMenuTrakt(oGuiElement, oOutputParameterHandler) if self.ADDON.getSetting('tmdb_account') != '': self.createContexMenuTMDB(oGuiElement, oOutputParameterHandler) self.createContexMenuSimil(oGuiElement, oOutputParameterHandler) oListItem = self.__createContextMenu(oGuiElement, oListItem) #sPluginHandle = cPluginHandler().getPluginHandle() # modif 22/06 #xbmcplugin.addDirectoryItem(sPluginHandle, sItemUrl, oListItem, isFolder=_isFolder) self.listing.append((sItemUrl, oListItem, _isFolder)) def createListItem(self, oGuiElement): oListItem = listitem(oGuiElement.getTitle()) oListItem.setInfo(oGuiElement.getType(), oGuiElement.getItemValues()) #oListItem.setThumbnailImage(oGuiElement.getThumbnail()) #oListItem.setIconImage(oGuiElement.getIcon()) # krypton et sont comportement oListItem.setArt({'poster': oGuiElement.getPoster(), 'thumb': oGuiElement.getThumbnail(), 'icon': oGuiElement.getIcon(), 'fanart': oGuiElement.getFanart()}) aProperties = oGuiElement.getItemProperties() for sPropertyKey in aProperties.keys(): oListItem.setProperty(sPropertyKey, aProperties[sPropertyKey]) return oListItem # affiche les liens playable def addHost(self, oGuiElement, oOutputParameterHandler = ''): if isKrypton(): cGui.CONTENT = 'movies' if oOutputParameterHandler.getValue('siteUrl'): sSiteUrl = oOutputParameterHandler.getValue('siteUrl') oGuiElement.setSiteUrl(sSiteUrl) oListItem = self.createListItem(oGuiElement) oListItem.setProperty('IsPlayable', 'true') oListItem.setProperty('Video', 'true') oListItem.addStreamInfo('video', {}) sItemUrl = self.__createItemUrl(oGuiElement, oOutputParameterHandler) oOutputParameterHandler.addParameter('sTitleWatched', oGuiElement.getTitleWatched()) self.createContexMenuWatch(oGuiElement, oOutputParameterHandler) oListItem = self.__createContextMenu(oGuiElement, oListItem) # sPluginHandle = cPluginHandler().getPluginHandle() # modif 13/09 #xbmcplugin.addDirectoryItem(sPluginHandle, sItemUrl, oListItem, isFolder=False) self.listing.append((sItemUrl, oListItem, False)) #Marquer vu/Non vu def createContexMenuWatch(self, oGuiElement, oOutputParameterHandler= ''): self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cGui', oGuiElement.getSiteName(), 'setWatched', self.ADDON.VSlang(30206)) def createContexMenuPageSelect(self, oGuiElement, oOutputParameterHandler): #sSiteUrl = oGuiElement.getSiteName() oContext = cContextElement() oContext.setFile('cGui') oContext.setSiteName('cGui') oContext.setFunction('selectpage') oContext.setTitle(self.ADDON.VSlang(30017)) oOutputParameterHandler.addParameter('OldFunction', oGuiElement.getFunction()) oOutputParameterHandler.addParameter('sId', oGuiElement.getSiteName()) oContext.setOutputParameterHandler(oOutputParameterHandler) oGuiElement.addContextItem(oContext) oContext = cContextElement() oContext.setFile('cGui') oContext.setSiteName('cGui') oContext.setFunction('viewback') oContext.setTitle(self.ADDON.VSlang(30018)) oOutputParameterHandler.addParameter('sId', oGuiElement.getSiteName()) oContext.setOutputParameterHandler(oOutputParameterHandler) oGuiElement.addContextItem(oContext) #marque page def createContexMenuFav(self, oGuiElement, oOutputParameterHandler = ''): oOutputParameterHandler.addParameter('sId', oGuiElement.getSiteName()) oOutputParameterHandler.addParameter('sFav', oGuiElement.getFunction()) oOutputParameterHandler.addParameter('sCat', oGuiElement.getCat()) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cFav', 'cFav', 'setBookmark', self.ADDON.VSlang(30210)) def createContexMenuTrakt(self, oGuiElement, oOutputParameterHandler= ''): oOutputParameterHandler.addParameter('sImdbId', oGuiElement.getImdbId()) oOutputParameterHandler.addParameter('sTmdbId', oGuiElement.getTmdbId()) oOutputParameterHandler.addParameter('sFileName', oGuiElement.getFileName()) sType = cGui.CONTENT.replace('tvshows', 'shows') oOutputParameterHandler.addParameter('sType', sType) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cTrakt', 'cTrakt', 'getAction', self.ADDON.VSlang(30214)) def createContexMenuTMDB(self, oGuiElement, oOutputParameterHandler = ''): oOutputParameterHandler.addParameter('sImdbId', oGuiElement.getImdbId()) oOutputParameterHandler.addParameter('sTmdbId', oGuiElement.getTmdbId()) oOutputParameterHandler.addParameter('sFileName', oGuiElement.getFileName()) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'themoviedb_org', 'themoviedb_org', 'getAction', 'TMDB') def createContexMenuDownload(self, oGuiElement, oOutputParameterHandler = '', status = '0'): if status == '0': self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cDownload', 'cDownload', 'StartDownloadOneFile', self.ADDON.VSlang(30215)) if status == '0' or status == '2': self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cDownload', 'cDownload', 'delDownload', self.ADDON.VSlang(30216)) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cDownload', 'cDownload', 'DelFile', self.ADDON.VSlang(30217)) if status == '1': self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cDownload', 'cDownload', 'StopDownloadList', self.ADDON.VSlang(30218)) if status == '2': self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cDownload', 'cDownload', 'ReadDownload', self.ADDON.VSlang(30219)) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cDownload', 'cDownload', 'ResetDownload', self.ADDON.VSlang(30220)) # Information def createContexMenuinfo(self, oGuiElement, oOutputParameterHandler = ''): oOutputParameterHandler = cOutputParameterHandler() oOutputParameterHandler.addParameter('sTitle', oGuiElement.getTitle()) oOutputParameterHandler.addParameter('sFileName', oGuiElement.getFileName()) oOutputParameterHandler.addParameter('sId', oGuiElement.getSiteName()) oOutputParameterHandler.addParameter('sMeta', oGuiElement.getMeta()) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cGui', oGuiElement.getSiteName(), 'viewinfo', self.ADDON.VSlang(30208)) # Bande annonce def createContexMenuba(self, oGuiElement, oOutputParameterHandler = ''): oOutputParameterHandler = cOutputParameterHandler() oOutputParameterHandler.addParameter('sTitle', oGuiElement.getTitle()) oOutputParameterHandler.addParameter('sFileName', oGuiElement.getFileName()) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cGui', oGuiElement.getSiteName(), 'viewBA', self.ADDON.VSlang(30212)) # Recherche similaire def createContexMenuSimil(self, oGuiElement, oOutputParameterHandler = ''): oOutputParameterHandler = cOutputParameterHandler() oOutputParameterHandler.addParameter('sFileName', oGuiElement.getFileName()) oOutputParameterHandler.addParameter('sTitle', oGuiElement.getTitle()) oOutputParameterHandler.addParameter('sCat', oGuiElement.getCat()) self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cGui', oGuiElement.getSiteName(), 'viewsimil', self.ADDON.VSlang(30213)) def CreateSimpleMenu(self, oGuiElement, oOutputParameterHandler, sFile, sName, sFunction, sTitle): oContext = cContextElement() oContext.setFile(sFile) oContext.setSiteName(sName) oContext.setFunction(sFunction) oContext.setTitle(sTitle) oContext.setOutputParameterHandler(oOutputParameterHandler) oGuiElement.addContextItem(oContext) def createContexMenuDelFav(self, oGuiElement, oOutputParameterHandler = ''): self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'cFav', 'cFav', 'delBookmarksMenu', self.ADDON.VSlang(30209)) def createContexMenuSettings(self, oGuiElement, oOutputParameterHandler = ''): self.CreateSimpleMenu(oGuiElement, oOutputParameterHandler, 'globalParametre', 'globalParametre', 'opensetting', self.ADDON.VSlang(30023)) def __createContextMenu(self, oGuiElement, oListItem): sPluginPath = cPluginHandler().getPluginPath() aContextMenus = [] # Menus classiques reglés a la base if len(oGuiElement.getContextItems()) > 0: for oContextItem in oGuiElement.getContextItems(): oOutputParameterHandler = oContextItem.getOutputParameterHandler() sParams = oOutputParameterHandler.getParameterAsUri() sTest = '%s?site=%s&function=%s&%s' % (sPluginPath, oContextItem.getFile(), oContextItem.getFunction(), sParams) aContextMenus += [(oContextItem.getTitle(), 'XBMC.RunPlugin(%s)' % (sTest, ), )] oListItem.addContextMenuItems(aContextMenus, True) return oListItem def __ContextMenu(self, oGuiElement, oListItem): sPluginPath = cPluginHandler().getPluginPath() aContextMenus = [] if len(oGuiElement.getContextItems()) > 0: for oContextItem in oGuiElement.getContextItems(): oOutputParameterHandler = oContextItem.getOutputParameterHandler() sParams = oOutputParameterHandler.getParameterAsUri() sTest = '%s?site=%s&function=%s&%s' % (sPluginPath, oContextItem.getFile(), oContextItem.getFunction(), sParams) aContextMenus += [(oContextItem.getTitle(), 'XBMC.RunPlugin(%s)' % (sTest, ), )] oListItem.addContextMenuItems(aContextMenus) #oListItem.addContextMenuItems(aContextMenus, True) return oListItem def __ContextMenuPlay(self, oGuiElement, oListItem): sPluginPath = cPluginHandler().getPluginPath() aContextMenus = [] if len(oGuiElement.getContextItems()) > 0: for oContextItem in oGuiElement.getContextItems(): oOutputParameterHandler = oContextItem.getOutputParameterHandler() sParams = oOutputParameterHandler.getParameterAsUri() sTest = '%s?site=%s&function=%s&%s' % (sPluginPath, oContextItem.getFile(), oContextItem.getFunction(), sParams) aContextMenus += [(oContextItem.getTitle(), 'XBMC.RunPlugin(%s)' % (sTest, ), )] oListItem.addContextMenuItems(aContextMenus) #oListItem.addContextMenuItems(aContextMenus, True) return oListItem def setEndOfDirectory(self, ForceViewMode = False): iHandler = cPluginHandler().getPluginHandle() # modif 22/06 if not self.listing: self.addText('cGui') xbmcplugin.addDirectoryItems(iHandler, self.listing, len(self.listing)) xbmcplugin.setPluginCategory(iHandler, '') xbmcplugin.setContent(iHandler, cGui.CONTENT) xbmcplugin.addSortMethod(iHandler, xbmcplugin.SORT_METHOD_NONE) xbmcplugin.endOfDirectory(iHandler, succeeded = True, cacheToDisc = True) # reglage vue # 50 = liste / 51 grande liste / 500 icone / 501 gallerie / 508 fanart / if ForceViewMode: xbmc.executebuiltin('Container.SetViewMode(' + str(ForceViewMode) + ')') else: if self.ADDON.getSetting('active-view') == 'true': if cGui.CONTENT == 'movies': #xbmc.executebuiltin('Container.SetViewMode(507)') xbmc.executebuiltin('Container.SetViewMode(%s)' % self.ADDON.getSetting('movie-view')) elif cGui.CONTENT == 'tvshows': xbmc.executebuiltin('Container.SetViewMode(%s)' % self.ADDON.getSetting('serie-view')) elif cGui.CONTENT == 'files': xbmc.executebuiltin('Container.SetViewMode(%s)' % self.ADDON.getSetting('default-view')) # bug affichage Kodi 18 del self.listing [:] def updateDirectory(self): xbmc.executebuiltin('Container.Refresh') def viewback(self): sPluginPath = cPluginHandler().getPluginPath() oInputParameterHandler = cInputParameterHandler() # sParams = oInputParameterHandler.getAllParameter() sId = oInputParameterHandler.getValue('sId') sTest = '%s?site=%s' % (sPluginPath, sId) xbmc.executebuiltin('XBMC.Container.Update(%s, replace)' % sTest) def viewsimil(self): sPluginPath = cPluginHandler().getPluginPath() oInputParameterHandler = cInputParameterHandler() # sFileName = oInputParameterHandler.getValue('sFileName') sTitle = oInputParameterHandler.getValue('sTitle') sCat = oInputParameterHandler.getValue('sCat') oOutputParameterHandler = cOutputParameterHandler() #oOutputParameterHandler.addParameter('searchtext', sFileName) oOutputParameterHandler.addParameter('searchtext', cUtil().CleanName(sTitle)) oOutputParameterHandler.addParameter('sCat', sCat) oOutputParameterHandler.addParameter('readdb', 'False') sParams = oOutputParameterHandler.getParameterAsUri() sTest = '%s?site=%s&function=%s&%s' % (sPluginPath, 'globalSearch', 'globalSearch', sParams) xbmc.executebuiltin('XBMC.Container.Update(%s)' % sTest) return False def selectpage(self): sPluginPath = cPluginHandler().getPluginPath() oInputParameterHandler = cInputParameterHandler() #sParams = oInputParameterHandler.getAllParameter() sId = oInputParameterHandler.getValue('sId') sFunction = oInputParameterHandler.getValue('OldFunction') siteUrl = oInputParameterHandler.getValue('siteUrl') oParser = cParser() oldNum = oParser.getNumberFromString(siteUrl) newNum = 0 if oldNum: newNum = self.showNumBoard() if newNum: try: siteUrl = siteUrl.replace(oldNum, newNum) oOutputParameterHandler = cOutputParameterHandler() oOutputParameterHandler.addParameter('siteUrl', siteUrl) sParams = oOutputParameterHandler.getParameterAsUri() sTest = '%s?site=%s&function=%s&%s' % (sPluginPath, sId, sFunction, sParams) xbmc.executebuiltin('XBMC.Container.Update(%s)' % sTest) except: return False return False def selectpage2(self): sPluginPath = cPluginHandler().getPluginPath() oInputParameterHandler = cInputParameterHandler() sId = oInputParameterHandler.getValue('sId') sFunction = oInputParameterHandler.getValue('OldFunction') siteUrl = oInputParameterHandler.getValue('siteUrl') # sParams = oInputParameterHandler.getAllParameter() selpage = self.showNumBoard() oOutputParameterHandler = cOutputParameterHandler() oOutputParameterHandler.addParameter('siteUrl', siteUrl) oOutputParameterHandler.addParameter('Selpage', selpage) sParams = oOutputParameterHandler.getParameterAsUri() sTest = '%s?site=%s&function=%s&%s' % (sPluginPath, sId, sFunction, sParams) xbmc.executebuiltin('XBMC.Container.Update(%s, replace)' % sTest) def setWatched(self): if True: #Use VStream database oInputParameterHandler = cInputParameterHandler() sSite = oInputParameterHandler.getValue('siteUrl') sTitle = oInputParameterHandler.getValue('sTitleWatched') if not sTitle: return meta = {} meta['title'] = sTitle meta['site'] = sSite db = cDb() row = db.get_watched(meta) if row: db.del_watched(meta) db.del_resume(meta) else: db.insert_watched(meta) # To test #xbmc.executebuiltin('Container.Refresh') else: # Use kodi buildin feature xbmc.executebuiltin('Action(ToggleWatched)') # Not usefull ? #xbmc.executebuiltin('Container.Refresh') def viewBA(self): oInputParameterHandler = cInputParameterHandler() sFileName = oInputParameterHandler.getValue('sFileName') from resources.lib.ba import cShowBA cBA = cShowBA() cBA.SetSearch(sFileName) cBA.SearchBA() def viewinfo(self): from resources.lib.config import WindowsBoxes # oGuiElement = cGuiElement() oInputParameterHandler = cInputParameterHandler() sTitle = oInputParameterHandler.getValue('sTitle') # sId = oInputParameterHandler.getValue('sId') sFileName = oInputParameterHandler.getValue('sFileName') sMeta = oInputParameterHandler.getValue('sMeta') sYear = oInputParameterHandler.getValue('sYear') # sMeta = 1 >> film sMeta = 2 >> serie sCleanTitle = cUtil().CleanName(sFileName) # on vire saison et episode if True: # sMeta == 2: sCleanTitle = re.sub('(?i).pisode [0-9]+', '', sCleanTitle) sCleanTitle = re.sub('(?i)saison [0-9]+', '', sCleanTitle) sCleanTitle = re.sub('(?i)S[0-9]+E[0-9]+', '', sCleanTitle) sCleanTitle = re.sub('(?i)[S|E][0-9]+', '', sCleanTitle) ui = WindowsBoxes(sTitle, sCleanTitle, sMeta, sYear) def __createItemUrl(self, oGuiElement, oOutputParameterHandler = ''): if (oOutputParameterHandler == ''): oOutputParameterHandler = cOutputParameterHandler() sParams = oOutputParameterHandler.getParameterAsUri() # cree une id unique # if oGuiElement.getSiteUrl(): # print str(hash(oGuiElement.getSiteUrl())) sPluginPath = cPluginHandler().getPluginPath() if (len(oGuiElement.getFunction()) == 0): sItemUrl = '%s?site=%s&title=%s&%s' % (sPluginPath, oGuiElement.getSiteName(), QuotePlus(oGuiElement.getCleanTitle()), sParams) else: sItemUrl = '%s?site=%s&function=%s&title=%s&%s' % (sPluginPath, oGuiElement.getSiteName(), oGuiElement.getFunction(), QuotePlus(oGuiElement.getCleanTitle()), sParams) #print sItemUrl return sItemUrl def showKeyBoard(self, sDefaultText = '', heading = ''): keyboard = xbmc.Keyboard(sDefaultText) keyboard.setHeading(heading) keyboard.doModal() if (keyboard.isConfirmed()): sSearchText = keyboard.getText() if (len(sSearchText)) > 0: return sSearchText return False def showNumBoard(self, sDefaultNum = ''): dialogs = dialog() numboard = dialogs.numeric(0, self.ADDON.VSlang(30019), sDefaultNum) #numboard.doModal() if numboard != None: return numboard return False def openSettings(self): return False def showNofication(self, sTitle, iSeconds = 0): return False def showError(self, sTitle, sDescription, iSeconds = 0): return False def showInfo(self, sTitle, sDescription, iSeconds = 0): return False
32,193
9,598
# Copyright (c) 2016-2019, Adam Karpierz # Licensed under the BSD license # http://opensource.org/licenses/BSD-3-Clause from ...jvm.lib.compat import * from ...jvm.lib import annotate from ...jvm.lib import public from ._typehandler import * # noqa @public class TypeManager(object): __slots__ = ('_state', '_handlers') def __init__(self, state=None): super(TypeManager, self).__init__() self._state = state self._handlers = {} def start(self): self._register_handler(VoidHandler) self._register_handler(BooleanHandler) self._register_handler(CharHandler) self._register_handler(ByteHandler) self._register_handler(ShortHandler) self._register_handler(IntHandler) self._register_handler(LongHandler) self._register_handler(FloatHandler) self._register_handler(DoubleHandler) self._register_handler(StringHandler) def stop(self): self._handlers = {} def _register_handler(self, hcls): thandler = hcls(self._state) self._handlers[thandler._jclass] = thandler return thandler def get_handler(self, jclass): thandler = self._handlers.get(jclass) if thandler is None: if not jclass.startswith("L"): raise ValueError("Don't know how to convert argument with " "type signature '{}'".format(jclass)) self._handlers[jclass] = thandler = ObjectHandler(self._state, jclass) return thandler
1,555
469
import os from django.conf import settings from django.core.exceptions import FieldDoesNotExist from hashids import Hashids import hashlib import string from datetime import datetime, date, timedelta from decimal import Decimal TWOPLACES = Decimal(10) ** -2 from django.db import models from django.apps import apps get_model = apps.get_model from django.http import Http404 from django.core.exceptions import ValidationError import threading from rest import helpers as rest_helpers from rest.uberdict import UberDict from rest import search from rest.privpub import PrivatePublicEncryption import importlib GRAPH_HELPERS = UberDict() GRAPH_HELPERS.restGet = None GRAPH_HELPERS.get_request = None GRAPH_HELPERS.views = None ENCRYPTER_KEY_FILE = os.path.join(settings.ROOT, "config", "encrypt_key.pem") ENCRYPTER = None if os.path.exists(ENCRYPTER_KEY_FILE): ENCRYPTER = PrivatePublicEncryption(private_key_file=ENCRYPTER_KEY_FILE) class RestError(Exception): def __init__(self, reason, code=None): self.reason = reason self.code = code def __repr__(self): return self.reason class PermisionDeniedException(RestError): def __init__(self, reason="permission denied", code=401): self.reason = reason self.code = code def requestHasPerms(request, perms, group=None): if not request.user.is_authenticated: return False, "auth required", 401 if not hasattr(request, 'member'): request.member, request.group = request.user.__class__.getMemberGroup(request, False, False) if request.member.hasPerm(perms): return True, None, None if group is None and hasattr(request, "group"): group = request.group if group and request.member.hasGroupPerm(group, perms): return True, None, None return False, "permission denied", 402 class MetaDataBase(models.Model): class Meta: abstract = True category = models.CharField(db_index=True, max_length=32, default=None, null=True, blank=True) key = models.CharField(db_index=True, max_length=80) value_format = models.CharField(max_length=16) value = models.TextField() int_value = models.IntegerField(default=None, null=True, blank=True) float_value = models.IntegerField(default=None, null=True, blank=True) def setValue(self, value): self.value = "{}".format(value) if type(value) is int and value < 2147483647: self.value_format = "I" self.int_value = value elif type(value) is float: self.value_format = "F" self.float_value = value elif isinstance(value, list): self.value_format = "L" # self.value = ",".join(value) elif isinstance(value, dict): self.value_format = "O" elif type(value) in [str, str] and len(value) < 9 and value.isdigit(): self.value_format = "I" self.int_value = value elif value in ["True", "true", "False", "false"]: self.value_format = "I" if value in ["True", "true"]: self.int_value = 1 else: self.int_value = 0 elif isinstance(value, bool): self.value_format = "I" if value: self.int_value = 1 else: self.int_value = 0 else: self.value_format = "S" def getStrictType(self, field_type): if type(self.value) is field_type: return self.value if field_type in [int, str, float, str]: return field_type(self.value) elif field_type is bool: return self.value in [True, 1, '1', 'y', 'Y', 'true'] elif field_type in [date, datetime]: return rest_helpers.parseDate(self.value) return self.value def getValue(self, field_type=None): if field_type: return self.getStrictType(field_type) elif self.value_format == 'I': return self.int_value elif self.value_format == 'F': return self.float_value elif self.value_format in ["L", "O"] and self.value: try: return eval(self.value) except: pass return self.value def __unicode__(self): if self.category: return "{}.{}={}".format(self.category, self.key, self.value) return "{}={}".format(self.key, self.value) def __str__(self): if self.category: return "{}.{}={}".format(self.category, self.key, self.value) return "{}={}".format(self.key, self.value) class MetaDataModel(object): def set_metadata(self, request, values=None): # this may get called before the model is saved if not self.id: self.save() if values is None: values = request request = None if not isinstance(values, dict): raise Exception("invalid metadata: {}".format(values)) for key, value in list(values.items()): cat = None if "." in key: cat, key = key.split('.') self.setProperty(key, value, cat, request=request) def metadata(self): return self.getProperties() def removeProperties(self, category=None): # this will remove all properties # if category is not it will remove all properties self.properties.filter(category=category).delete() def getProperties(self, category=None): ret = {} for p in self.properties.all(): if p.category: props = self.getFieldProps(p.category) if props.hidden: continue if p.category not in ret or not isinstance(ret.get(p.category, None), dict): ret[p.category] = {} props = self.getFieldProps("{}.{}".format(p.category, p.key)) if props.hidden: continue ret[p.category][p.key] = p.getValue() else: props = self.getFieldProps(p.key) if props.hidden: continue ret[p.key] = p.getValue() if category is not None: if category in ret: return ret[category] return {} return ret def __initFieldProps(self): if not hasattr(self, "__field_props"): if hasattr(self.RestMeta, "METADATA_FIELD_PROPERTIES"): # this provides extra protection for metadata fields self.__field_props = self.RestMeta.METADATA_FIELD_PROPERTIES else: self.__field_props = None def getFieldProps(self, key): self.__initFieldProps() full_key = key category = None if "." in key: category, key = key.split('.') props = UberDict() if self.__field_props: if category and self.__field_props.get(category, None): cat_props = self.__field_props.get(category, None) if cat_props: props.notify = cat_props.get("notify", None) props.on_change_name = cat_props.get("on_change", None) props.hidden = cat_props.get("hidden", False) if props.on_change_name: props.on_change = getattr(self, props.on_change_name, None) field_props = self.__field_props.get(full_key, None) if field_props: props.notify = field_props.get("notify", props.notify) props.requires = field_props.get("requires", None) props.hidden = field_props.get("hidden", False) on_change_name = field_props.get("on_change", None) if on_change_name: on_change = getattr(self, on_change_name, None) if on_change: props.on_change = on_change return props def checkFieldPerms(self, full_key, props, request=None): if not props.requires: return True if not request or not request.member: return False if request.member.hasPermission(props.requires) or request.user.is_superuser: return True # this a unauthorized attempt to change field, log and throw exception if props.notify and request.member: subject = "permission denied changing protected '{}' field".format(full_key) msg = "permission denied changing protected field '{}'\nby user: {}\nfor: {}".format( full_key, request.user.username, self ) request.member.notifyWithPermission(props.notify, subject, msg, email_only=True) raise PermisionDeniedException(subject) def setProperties(self, data, category=None, request=None, using=None): for k,v in list(data.items()): self.setProperty(k, v, category, request=request, using=using) def setProperty(self, key, value, category=None, request=None, using=None): # rest_helpers.log_print("{}:{} ({})".format(key, value, type(value))) on_change = None if not using: using = getattr(self.RestMeta, "DATABASE", using) if not request: request = RestModel.getActiveRequest() self.__initFieldProps() if isinstance(value, dict): return self.setProperties(value, key) username = "root" if request and request.member: username = request.member.username prop = None if "." in key: category, key = key.split('.') if category: # delete any keys with this category name full_key = "{}.{}".format(category, key) # this deletes anything with the key that matches the category # this works because the category is stored not in key but category field # rest_helpers.log_print("deleting key={}".format(category)) self.properties.filter(key=category).delete() else: full_key = key field_props = self.getFieldProps(full_key) if not self.checkFieldPerms(full_key, field_props, request): return False check_value = "{}".format(value) has_changed = False prop = self.properties.filter(category=category, key=key).last() old_value = None if prop: # existing property we need to make sure we delete old_value = prop.getValue() if value is None or value == "": prop.delete() has_changed = True else: has_changed = check_value != prop.value if not has_changed: return prop.setValue(value) prop.save(using=using) if field_props.on_change: field_props.on_change(key, value, old_value, category) elif value is None or value == "": # do not create none or empty property return False else: has_changed = True PropClass = self.get_fk_model("properties") prop = PropClass(parent=self, key=key, category=category) prop.setValue(value) # rest_helpers.log_print(u"saving {}.{}".format(category, key)) # rest_helpers.log_print(u"saving {} : {}".format(full_key, value)) prop.save(using=using) if hasattr(self, "_recordRestChange"): self._recordRestChange("metadata.{}".format(full_key), old_value) if field_props.notify and request and request.member: notify = field_props.get("notify") msg = "protected field '{}' changed to '{}'\nby user: {}\nfor: {}".format( full_key, value, username, self ) request.member.notifyWithPermission(notify, "protected '{}' field changed".format(full_key), msg, email_only=True) return has_changed def getProperty(self, key, default=None, category=None, field_type=None): try: if "." in key: category, key = key.split('.') return self.properties.get(category=category, key=key).getValue(field_type) except: pass return default class RestValidationError(RestError): pass class RestModel(object): class __RestMeta__: NO_SAVE_FIELDS = ["uuid", "id", "pk", "created", "modified"] NO_SHOW_FIELDS = ["password"] WHITELISTED = ["merchant", "group", "user", "member", "terminal"] class RestMeta: NO_SAVE_FIELDS = [] SAVE_FIELDS = [] GRAPHS = {} @staticmethod def generateUUID(*args, **kwargs): upper = kwargs.get("upper", True) max_length = kwargs.get("max_length", None) uuid = "" for key in args: if isinstance(key, float): key = str(float) if isinstance(key, int): uuid += Hashids().encrypt(key) if isinstance(key, str): uuid += rest_helpers.toString(hashlib.md5(rest_helpers.toBytes(key)).hexdigest()) if len(uuid) > 125: uuid = uuid[:125] if max_length != None: uuid = uuid[:max_length] if upper: return uuid.upper() return uuid @classmethod def buildGraph(cls, name): # we need to build it if hasattr(cls.RestMeta, "GRAPHS"): graphs = cls.RestMeta.GRAPHS if name in graphs: graph = graphs[name] else: graph = {} else: graph = {} if "no_uscore" not in graph: graph["no_uscore"] = False no_show_fields = RestModel.__RestMeta__.NO_SHOW_FIELDS if hasattr(cls.RestMeta, "NO_SHOW_FIELDS"): no_show_fields = cls.RestMeta.NO_SHOW_FIELDS field_names = [] for f in cls._meta.fields: if not f.name.endswith("_ptr"): if f.name not in no_show_fields: field_names.append(f.name) if "graphs" in graph: if "recurse_into" not in graph: graph["recurse_into"] = [] if "fields" in graph: graph["fields"] = graph["fields"] elif "fields" not in graph and "self" in graph["graphs"]: graph["fields"] = [] else: graph["fields"] = field_names for field in graph["graphs"]: gname = graph["graphs"][field] size = None ForeignModel = None sort = None if field.startswith("generic__"): if field not in graph["recurse_into"]: graph["recurse_into"].append((field, gname)) continue if isinstance(gname, dict): size = gname.get("size") sort = gname.get("sort") fm_name = gname.get("model") gname = gname.get("graph") if not gname: gname = "default" if fm_name: a_name, m_name = fm_name.split(".") ForeignModel = RestModel.getModel(a_name, m_name) if not field or field == "self": # this means it is referencing self foreign_graph = cls.buildGraph(gname) for part in foreign_graph: if part not in graph: graph[part] = foreign_graph[part] else: for f in foreign_graph[part]: if f not in graph[part]: graph[part].append(f) # graph[part] += foreign_graph[part] continue # print "get FK: {0}".format(field) if not ForeignModel: ForeignModel = cls.get_fk_model(field) if not ForeignModel: rest_helpers.log_print("no foreignkey: {0}".format(field)) continue # print ForeignModel # print graph["recurse_into"] # print graph["recurse_into"] if field not in graph["recurse_into"]: graph["recurse_into"].append(field) # print ForeignModel # if not hasattr(ForeignModel, "getGraph"): # foreign_graph = {} # foreign_graph["fields"] = [] # for f in ForeignModel._meta.fields: # if f.name not in RestModel.__RestMeta__.NO_SHOW_FIELDS: # foreign_graph["fields"].append(f.name) # print ForeignModel # print foreign_graph["fields"] # else: if not hasattr(ForeignModel, "getGraph"): # print "NO getGraph" continue # print "getting graph: {0} for {1}".format(gname, field) foreign_graph = ForeignModel.getGraph(gname) # print foreign_graph for part in ["fields", "recurse_into", "extra", "exclude"]: if part not in foreign_graph: continue graph_part = foreign_graph[part] if part not in graph: graph[part] = [] root_part = graph[part] for f in graph_part: if type(f) is tuple: f1, f2 = f nfname = ("{0}.{1}".format(field, f1), f2) elif graph["no_uscore"] and '_' in f: f1, f2 = f, f.replace('_', '').split('.')[-1] # print field # print f2 nfname = ("{0}.{1}".format(field, f1), f2) else: nfname = "{0}.{1}".format(field, f) if nfname not in root_part: root_part.append(nfname) del graph["graphs"] if "fields" not in graph: if graph["no_uscore"]: graph["fields"] = [] for f in field_names: if "_" in f: f1, f2 = f, f.lower().replace('_', '') # print "noscore" # print f1 # print f2 graph["fields"].append((f1, f2)) else: graph["fields"].append(f) else: graph["fields"] = field_names if "no_uscore" in graph: del graph["no_uscore"] return graph @classmethod def ro_objects(cls): using = getattr(cls.RestMeta, "RO_DATABASE", None) if using is None: using = getattr(cls.RestMeta, "DATABASE", None) # if using is None: # if settings.DATABASES.get("readonly", None) != None: # using = "readonly" if using: return cls.objects.using(using) return cls.objects @classmethod def rw_objects(cls): using = getattr(cls.RestMeta, "DATABASE", None) if using: return cls.objects.using(using) return cls.objects def safeSave(self, **kwargs): using = getattr(self.RestMeta, "DATABASE", None) if using: return self.save(using=using, **kwargs) return self.save(**kwargs) @classmethod def getGraph(cls, name): graph_key = "_graph_{0}__".format(name) if hasattr(cls, graph_key): return getattr(cls, graph_key) if not hasattr(cls, "_lock__"): cls._lock__ = threading.RLock() # cls._lock__.acquire() # try: graph = cls.buildGraph(name) # print "-" * 80 # print "SETTING GRAPH {0} FOR {1}".format(name, cls.__name__) # print graph setattr(cls, graph_key, graph) # print "." * 80 # except: # pass # cls._lock__.release() return graph def toGraph(self, request=None, graph="basic"): RestModel._setupGraphHelpers() if not request: request = GRAPH_HELPERS.get_request() return GRAPH_HELPERS.restGet(request, self, return_httpresponse=False, **self.getGraph(graph)) @classmethod def getActiveLogger(cls): return rest_helpers.getLogger(cls.getActiveRequest()) @classmethod def getActiveMember(cls): request = cls.getActiveRequest() if request: return request.member return None @classmethod def getActiveRequest(cls): if not GRAPH_HELPERS.get_request: mw = importlib.import_module("rest.middleware") GRAPH_HELPERS.get_request = mw.get_request return GRAPH_HELPERS.get_request() @classmethod def getFromRequest(cls, request): key = cls.__name__.lower() key_p = "{0}_id".format(key) lookup_fields = [key, key_p] using = getattr(cls.RestMeta, "DATABASE", None) for field in lookup_fields: value = request.DATA.get(field) if value: if not using: obj = cls.objects.filter(pk=value).first() else: obj = cls.objects.using(using).filter(pk=value).first() if obj: return obj lookup_fields = getattr(cls.RestMeta, "UNIQUE_LOOKUP", []) for field in lookup_fields: value = request.DATA.get(field) if value: q = {} q[field] = value if not using: obj = cls.objects.filter(**q).first() else: obj = cls.objects.using(using).filter(**q).first() if obj: return obj return None value = request.DATA.get(key_p) if not value: value = request.DATA.get(key) if not value: return None if using: return cls.objects.using(using).filter(pk=value).first() return cls.objects.filter(pk=value).first() @classmethod def getFromPK(cls, pk): using = getattr(cls.RestMeta, "DATABASE", None) if using: return cls.objects.using(using).filter(pk=pk).first() return cls.objects.filter(pk=pk).first() @classmethod def restEncrypt(cls, data): if ENCRYPTER: return ENCRYPTER.encrypt(data) return data @staticmethod def restGetModel(app_name, model_name): return apps.get_model(app_name, model_name) @staticmethod def getModel(app_name, model_name): return apps.get_model(app_name, model_name) def restGetGenericModel(self, field): # called by the rest module to magically parse # a component that is marked genericrelation in a graph if not hasattr(self, field): rest_helpers.log_print("model has no field: {0}".format(field)) return None name = getattr(self, field) if not name or "." not in name: return None a_name, m_name = name.split(".") model = RestModel.getModel(a_name, m_name) if not model: rest_helpers.log_print("GENERIC MODEL DOES NOT EXIST: {0}".format(name)) return model def restGetGenericRelation(self, field): # called by the rest module to magically parse # a component that is marked genericrelation in a graph GenericModel = self.restGetGenericModel(field) if not GenericModel: return None key = getattr(self, "{0}_id".format(field)) return GenericModel.rw_objects().filter(pk=key).first() @staticmethod def restGetModelDB(Model, default=None): if hasattr(Model, "RestMeta"): return getattr(Model.RestMeta, "DATABASE", default) return default @property def has_model_changed(self): if hasattr(self, "_changed__"): return len(self._changed__) > 0 return False def saveFields(self, allow_null=True, **kwargs): """ Helper method to save a list of fields """ self._changed__ = UberDict() for key, value in list(kwargs.items()): if value is None and not allow_null: continue self.restSaveField(key, value) if len(self._changed__): self.save() def restSaveField(self, fieldname, value, has_fields=False, has_no_fields=False, using=None): if not hasattr(self, "_changed__"): self._changed__ = UberDict() if fieldname.startswith("_"): return if not hasattr(self, "_field_names__"): self._field_names__ = [f.name for f in self._meta.get_fields()] # print "saving field: {0} = {1}".format(fieldname, value) if fieldname in RestModel.__RestMeta__.NO_SAVE_FIELDS: return if has_no_fields and fieldname in self.RestMeta.NO_SAVE_FIELDS: return if has_fields and fieldname not in self.RestMeta.SAVE_FIELDS: return if fieldname.endswith("_id") and not self.get_field_type(fieldname): # django will have ForeignKeys with _id, we don't want that, on_delete=models.CASCADE fieldname = fieldname[:-3] setter = "set_{0}".format(fieldname) if hasattr(self, setter): getattr(self, setter)(value) return if fieldname in self._field_names__: # TODO check if it is a function if isinstance(value, models.Model): setattr(self, fieldname, value) self._changed__[fieldname] = True return ForeignModel = self.get_fk_model(fieldname) if ForeignModel and isinstance(value, dict): obj = getattr(self, fieldname, None) if obj is None: obj = ForeignModel() if using is None: using = self.restGetModelDB(self) obj.saveFromDict(None, value, using=using) # rest_helpers.log_print("{} vs {}".format(self._state.db, obj._state.db)) # rest_helpers.log_print("saving FK to {} ({}.{}) - {}".format(fieldname, using, obj.pk, type(obj)), value) setattr(self, fieldname, obj) self._changed__[fieldname] = True return elif ForeignModel and value and (type(value) is int or value.isdigit()): # print "\tforeign model({2}) field: {0} = {1}".format(fieldname, value, ForeignModel.__class__.__name__) value = int(value) using = RestModel.restGetModelDB(ForeignModel) if using: value = ForeignModel.objects.using(using).filter(pk=value).first() else: value = ForeignModel.objects.filter(pk=value).first() elif ForeignModel and "MediaItem" in ForeignModel.__name__: if value: self.saveMediaFile(value, fieldname, None, True) return elif ForeignModel and not value: value = None # maybe we could look for to_python here to make sure we have proper conversion # thinking mainly around datetimes from epoch values if not ForeignModel: # field_model, model, direct, mm = self._meta.get_field_by_name(fieldname) field_model = self._meta.get_field(fieldname) # hack to handle save datetime fields correctly from floats try: if field_model and value != None: field_model_name = field_model.__class__.__name__ if field_model_name == "DateTimeField": value = rest_helpers.parseDateTime(value) # value = datetime.fromtimestamp(float(value)) elif field_model_name == "DateField": value = rest_helpers.parseDate(value, as_date=True) elif field_model_name == "IntegerField": value = int(value) elif field_model_name == "FloatField": value = float(value) elif field_model_name == "CurrencyField": value = Decimal(value).quantize(TWOPLACES) elif field_model_name == "BooleanField": if value in [True, 1, 'True', 'true', '1', 't', 'y', 'yes']: value = True else: value = False except Exception: return if hasattr(self, fieldname) and getattr(self, fieldname) != value: self._changed__[fieldname] = getattr(self, fieldname) setattr(self, fieldname, value) # else: # print "does not have field: {0}".format(fieldname) def saveFromRequest(self, request, **kwargs): if "files" not in kwargs: kwargs["files"] = request.FILES return self.saveFromDict(request, request.DATA, **kwargs) def _recordRestChange(self, fieldname, old_value): if not hasattr(self, "_changed__"): self._changed__ = UberDict() if "." in fieldname: fields = fieldname.split('.') root = self._changed__ for f in fields[:-1]: if f not in root: root[f] = UberDict() root = root[f] root[fields[-1]] = old_value else: self._changed__[fieldname] = old_value def saveFromDict(self, request, data, files=None, **kwargs): can_save = getattr(self.RestMeta, "CAN_SAVE", True) if not can_save: return self.restStatus(request, False, error="saving not allowed via rest for this model.") # check check for save permissions if request is None: request = RestModel.getActiveRequest() if request is None: request = UberDict(member=None, FILES=[]) if hasattr(self, "onRestCanSave"): # this should throw an error self.onRestCanSave(request) is_new = self.id is None has_fields = hasattr(self.RestMeta, "SAVE_FIELDS") and len(self.RestMeta.SAVE_FIELDS) has_no_fields = hasattr(self.RestMeta, "NO_SAVE_FIELDS") and len(self.RestMeta.NO_SAVE_FIELDS) self._field_names__ = [f.name for f in self._meta.get_fields()] # fix for multidatabase support and using readonly db for get self._state.db = kwargs.get("using", self.restGetModelDB(self, "default")) auto_save_fields = getattr(self.RestMeta, "AUTO_SAVE", None) if auto_save_fields: rest_helpers.log_print(auto_save_fields) for field in auto_save_fields: rest_helpers.log_print(field) if isinstance(field, tuple): m_field, req_field = field else: m_field = field req_field = field req_value = getattr(request, req_field, None) if request and req_value: data[m_field] = req_value rest_helpers.log_print(data) self._changed__ = UberDict() if hasattr(self.RestMeta, "POST_SAVE_FIELDS"): post_save_fields = self.RestMeta.POST_SAVE_FIELDS else: post_save_fields = [] using = kwargs.get("using", self.restGetModelDB(self)) deferred = {} group_fields = {} for fieldname in data: # we allow override via kwargs value = data.get(fieldname) if "." in fieldname: gname = fieldname[:fieldname.find('.')] fname = fieldname[fieldname.find('.')+1:] setter = "set_{0}".format(gname) if hasattr(self, setter): if gname not in group_fields: group_fields[gname] = {} group_fields[gname][fname] = value continue if fieldname in post_save_fields or fieldname.startswith("metadata"): deferred[fieldname] = value continue if fieldname not in kwargs: self.restSaveField(fieldname, value, has_fields, has_no_fields, using=using) for key, value in list(kwargs.items()): if key in post_save_fields: deferred[fieldname] = value continue self.restSaveField(key, value, has_fields, has_no_fields, using=using) self.restSaveFiles(request, files) self.on_rest_pre_save(request) self.save(using=using) for key, value in list(deferred.items()): self.restSaveField(key, value, has_fields, has_no_fields, using=using) if len(deferred): self.save(using=using) # these setters are responsible for saving themselves for gname in group_fields: setter = "set_{0}".format(gname) getattr(self, setter)(request, group_fields[gname]) if hasattr(self, "onSavedFromRequest"): self.onSavedFromRequest(request, **kwargs) elif not is_new: self.on_rest_saved(request) return self def restSaveFiles(self, request, files=None): if files is None: files = request.FILES for name in files: key = "upload__{0}".format(name) if hasattr(self, key): getattr(self, key)(files[name], name) else: ForeignModel = self.get_fk_model(name) if ForeignModel and ForeignModel.__name__ == "MediaItem": rest_helpers.log_print("saving media file: {}".format(name)) self.saveMediaFile(files[name], name) def changesFromDict(self, data): deltas = [] field_names = [f.name for f in self._meta.get_fields()] for key in data: if key not in field_names: continue # we allow override via kwargs value = data.get(key) # field_model, model, direct, mm = self._meta.get_field_by_name(key) field_model = self._meta.get_field(key) # hack to handle save datetime fields correctly from floats try: if field_model and value != None: field_model_name = field_model.__class__.__name__ if field_model_name == "DateTimeField": value = datetime.fromtimestamp(float(value)) elif field_model_name == "DateField": value = rest_helpers.parseDate(value) elif field_model_name == "IntegerField": value = int(value) elif field_model_name == "FloatField": value = float(value) elif field_model_name == "CurrencyField": value = Decimal(value).quantize(TWOPLACES) if hasattr(self, key) and getattr(self, key) != value: deltas.append(key) except: pass return deltas def copyFieldsFrom(self, obj, fields): for f in fields: if hasattr(self, f): setattr(self, f, getattr(obj, f)) def saveMediaFile(self, file, name, file_name=None, is_base64=False, group=None): """ Generic method to save a media file """ if file_name is None: file_name = name MediaItem = RestModel.getModel("medialib", "MediaItem") # make sure we set the name base64_data if is_base64: mi = MediaItem(name=file_name, base64_data=file, group=group) elif type(file) in [str, str] and (file.startswith("https:") or file.startswith("http:")): mi = MediaItem(name=name, downloadurl=file, group=group) else: mi = MediaItem(name=name, newfile=file, group=group) mi.save() setattr(self, name, mi) self.save() return mi def updateLogModel(self, request, model): if not request: request = self.getActiveRequest() if not request or not hasattr(request, "setLogModel"): rest_helpers.log_print("request does not support setLogModel") return if not self.id: self.save() request.setLogModel(model, self.id) def on_rest_pre_get(self, request): pass def on_rest_get(self, request): # check view permissions perms = getattr(self.RestMeta, "VIEW_PERMS", None) if perms: # we need to check if this user has permission group_field = getattr(self.RestMeta, "GROUP_FIELD", "group") status, error, code = requestHasPerms(request, perms, getattr(self, group_field, None)) if not status: return self.restStatus(request, False, error=error, error_code=code) graph = request.DATA.get("graph", "default") self.on_rest_pre_get(request) return self.restGet(request, graph) def on_rest_post(self, request): perms = getattr(self.RestMeta, "SAVE_PERMS", None) if perms: # we need to check if this user has permission group_field = getattr(self.RestMeta, "GROUP_FIELD", "group") status, error, code = requestHasPerms(request, perms, getattr(self, group_field, None)) if not status: return self.restStatus(request, False, error=error, error_code=code) self.saveFromRequest(request) status_only = request.DATA.get("status_only", False, field_type=bool) if status_only: return self.restStatus(request, True) graph = request.DATA.get("graph", "default") return self.restGet(request, graph) def on_rest_pre_save(self, request, **kwargs): pass def on_rest_created(self, request): self.on_rest_saved(request) def on_rest_saved(self, request): pass def on_rest_delete(self, request): can_delete = getattr(self.RestMeta, "CAN_DELETE", False) if not can_delete: return self.restStatus(request, False, error="deletion not allowed via rest for this model.") perms = getattr(self.RestMeta, "SAVE_PERMS", None) if perms: # we need to check if this user has permission group_field = getattr(self.RestMeta, "GROUP_FIELD", "group") status, error, code = requestHasPerms(request, perms, getattr(self, group_field, None)) if not status: return self.restStatus(request, False, error=error, error_code=code) self.delete() RestModel._setupGraphHelpers() return self.restStatus(request, True) @classmethod def restList(cls, request, qset, graph=None, totals=None, return_httpresponse=True): RestModel._setupGraphHelpers() sort = None if hasattr(cls.RestMeta, "DEFAULT_SORT"): sort = cls.RestMeta.DEFAULT_SORT if totals: fields = totals totals = {} for tf in fields: cls_method = "qset_totals_{}".format(tf) if hasattr(cls, cls_method): totals[tf] = getattr(cls, cls_method)(qset, request) if not graph and request != None: graph = request.DATA.get("graph", "default") return GRAPH_HELPERS.restList(request, qset, sort=sort, totals=totals, return_httpresponse=return_httpresponse, **cls.getGraph(graph)) @classmethod def toList(cls, qset, graph=None, totals=None, request=None): return cls.restList(request, qset, graph, totals, False) def restStatus(self, request, status, **kwargs): RestModel._setupGraphHelpers() return GRAPH_HELPERS.restStatus(request, status, **kwargs) def restGet(self, request, graph=None, as_dict=False): RestModel._setupGraphHelpers() if not request: request = self.getActiveRequest() if not graph and request: graph = request.DATA.get("graph", "default") elif not graph: graph = "default" return_response = not as_dict return GRAPH_HELPERS.restGet(request, self, return_httpresponse=return_response, **self.getGraph(graph)) def toDict(self, graph=None): RestModel._setupGraphHelpers() return self.restGet(None, graph=graph, as_dict=True) @classmethod def on_rest_list_filter(cls, request, qset=None): # override on do any pre filters return cls.on_rest_list_perms(request, qset) @classmethod def on_rest_list_perms(cls, request, qset=None): if request.group: group_perms = getattr(cls.RestMeta, "LIST_PERMS_GROUP", None) if group_perms is None: group_perms = getattr(cls.RestMeta, "VIEW_PERMS", None) if group_perms and request.member: has_perm = request.member.hasGroupPerm(request.group, group_perms) or request.member.hasPerm(group_perms) if not has_perm: return cls.objects.none() qset = cls.on_rest_filter_children(request, qset) else: all_perms = getattr(cls.RestMeta, "VIEW_PERMS", None) if all_perms: if not request.member.hasPerm(all_perms): return cls.objects.none() return qset @classmethod def on_rest_filter_children(cls, request, qset=None): group_field = getattr(cls.RestMeta, "GROUP_FIELD", "group") parent_kinds = getattr(cls.RestMeta, "LIST_PARENT_KINDS", ["org"]) if request.DATA.get("child_groups") or request.group.kind in parent_kinds: ids = request.group.getAllChildrenIds() ids.append(request.group.id) # to avoid future filtering issues remove group request.group = None request.DATA.remove(group_field) if group_field != "group": request.DATA.remove("group") q = {} q["{}_id__in".format(group_field)] = ids return qset.filter(**q) return qset @classmethod def on_rest_list_ready(cls, request, qset=None): # override on do any post filters return qset @classmethod def on_rest_date_filter(cls, request, qset=None): date_range_field = getattr(cls.RestMeta, "DATE_RANGE_FIELD", "created") date_range_default = getattr(cls.RestMeta, "DATE_RANGE_DEFAULT", None) if date_range_default != None: date_range_default = datetime.now() - timedelta(days=date_range_default) qset = rest_helpers.filterByDateRange(qset, request, start=date_range_default, end=datetime.now()+timedelta(days=1), field=date_range_field) else: qset = rest_helpers.filterByDateRange(qset, request, field=date_range_field) return qset @classmethod def on_rest_list(cls, request, qset=None): qset = cls.on_rest_list_query(request, qset) graph = request.DATA.get("graph", "list") format = request.DATA.get("format") if format: return cls.on_rest_list_format(request, format, qset) totals = request.DATA.getlist("totals", None) return cls.restList(request, qset, graph, totals) @classmethod def on_rest_list_query(cls, request, qset=None): cls._boundRest() request.rest_class = cls if qset is None: qset = cls.ro_objects().all() qset = cls.on_rest_list_filter(request, qset) qset = cls.filterFromRequest(request, qset) qset = cls.queryFromRequest(request, qset) qset = cls.searchFromRequest(request, qset) qset = cls.on_rest_date_filter(request, qset) qset = cls.on_rest_list_ready(request, qset) return qset @classmethod def on_rest_list_format(cls, request, format, qset): if format in ["summary", "summary_only"]: return cls.on_rest_list_summary(request, qset) if hasattr(cls.RestMeta, "FORMATS"): fields = cls.RestMeta.FORMATS.get(format) else: no_show_fields = RestModel.__RestMeta__.NO_SHOW_FIELDS if hasattr(cls.RestMeta, "NO_SHOW_FIELDS"): no_show_fields = cls.RestMeta.NO_SHOW_FIELDS fields = [] for f in cls._meta.fields: if not f.name.endswith("_ptr"): if f.name not in no_show_fields: fields.append(f.name) if fields: name = request.DATA.get("format_filename", None) format_size = request.DATA.get("format_size", 10000) if name is None: name = "{}.{}".format(cls.__name__.lower(), format) # print "csv size: {}".format(qset.count()) sort = request.DATA.get("sort", getattr(cls.RestMeta, "DEFAULT_SORT", None)) if sort: qset = qset.order_by(sort) cls._boundRest() return GRAPH_HELPERS.views.restCSV(request, qset, fields, name, format_size) @classmethod def on_rest_list_summary(cls, request, qset): if not hasattr(cls.RestMeta, "SUMMARY_FIELDS"): return cls.restList(request, qset, None) cls._boundRest() summary_info = getattr(cls.RestMeta, "SUMMARY_FIELDS") output = UberDict() output.count = qset.count() for key, value in list(summary_info.items()): if key == "sum": res = rest_helpers.getSum(qset, *value) if isinstance(res, dict): output.update(res) else: output[value[0]] = res elif key == "avg": for f in value: output["avg_{}".format(f)] = rest_helpers.getAverage(qset, f) elif key == "max": for f in value: output["max_{}".format(f)] = rest_helpers.getMax(qset, f) elif isinstance(value, dict): if "|" in key: fields = key.split("|") if len(fields) > 1: lbl = fields[0] action = fields[1] field = None if len(fields) > 2: field = fields[2] else: action = "count" lbl = key field = None act_qset = qset.filter(**value) if action == "count": output[lbl] = act_qset.count() elif action == "sum": output[lbl] = rest_helpers.getSum(act_qset, field) elif action == "avg": output[lbl] = rest_helpers.getAverage(act_qset, field) elif action == "max": output[lbl] = rest_helpers.getMax(act_qset, field) return GRAPH_HELPERS.restGet(request, output) @classmethod def on_rest_batch(cls, request, action): # this method is called when rest_batch='somme action' cls._boundRest() batch_ids = request.DATA.getlist("batch_ids", []) batch_id_field = request.DATA.get("batch_id_field", "pk") q = {} if batch_ids: q["{}__in".format(batch_id_field)] = batch_ids batch_query = request.DATA.get("batch_query", None) if batch_query: # we ignore ids when doing a query q.update(batch_query) if action == "delete": can_delete = getattr(cls.RestMeta, "CAN_DELETE", False) if not can_delete: return GRAPH_HELPERS.restStatus(request, False, error="deletion not allowed via rest for this model.") qset = cls.rw_objects().filter(**q) count = qset.delete() return GRAPH_HELPERS.restStatus(request, True, error="delete {} items".format(count)) elif action == "update": qset = cls.rw_objects().filter(**q) update_fields = request.DATA.get(["batch_data", "batch_update"]) if not isinstance(update_fields, dict): return GRAPH_HELPERS.restStatus(request, False, error="batch_update should be key/values") count = qset.update(**update_fields) return GRAPH_HELPERS.restStatus(request, True, error="updated {} items".format(count)) elif action == "create": batch_data = request.DATA.getlist("batch_data", []) items = [] exist = [] for item in batch_data: try: obj = cls.ro_objects().filter(**item).last() if not obj: obj.saveFromDict(request, item) items.append(obj) except: pass return GRAPH_HELPERS.restList(request, items) return GRAPH_HELPERS.restStatus(request, False, error="not implemented") @classmethod def on_rest_create(cls, request, pk=None): can_create = getattr(cls.RestMeta, "CAN_CREATE", True) if not can_create: return GRAPH_HELPERS.restStatus(request, False, error="creation not allowed via rest for this model.") if hasattr(cls.RestMeta, "REQUEST_DEFAULTS"): kv = {} for k, v in list(cls.RestMeta.REQUEST_DEFAULTS.items()): if hasattr(request, k): value = getattr(request, k) if value != None: kv[v] = value obj = cls.createFromRequest(request, **kv) else: obj = cls.createFromRequest(request) obj.on_rest_created(request) graph = request.DATA.get("graph", "default") return obj.restGet(request, graph) @classmethod def _boundRest(cls): RestModel._setupGraphHelpers() @staticmethod def _setupGraphHelpers(): if not GRAPH_HELPERS.views: views = importlib.import_module("rest.views") GRAPH_HELPERS.views = views GRAPH_HELPERS.restNotFound = views.restNotFound GRAPH_HELPERS.restStatus = views.restStatus GRAPH_HELPERS.restList = views.restList GRAPH_HELPERS.restGet = views.restGet if not GRAPH_HELPERS.get_request: mw = importlib.import_module("rest.middleware") GRAPH_HELPERS.get_request = mw.get_request @classmethod def get_rest_help(cls): output = UberDict() if cls.__doc__: output.doc = cls.__doc__.rstrip() else: output.doc = "" output.model_name = cls.__name__ output.fields = cls.rest_getQueryFields(True) output.graphs = {} if hasattr(cls, "RestMeta"): output.graph_names = list(getattr(cls.RestMeta, "GRAPHS", {}).keys()) for key in output.graph_names: output.graphs[key] = cls.getGraph(key) output.no_show_fields = getattr(cls.RestMeta, "NO_SHOW_FIELDS", []) output.no_save_fields = getattr(cls.RestMeta, "NO_SAVE_FIELDS", []) output.search_fields = getattr(cls.RestMeta, "SEARCH_FIELDS", []) return output @classmethod def on_rest_request(cls, request, pk=None): # check if model id is in post request.rest_class = cls cls._boundRest() if not pk: pk_fields = [] key = cls.__name__.lower() key_p = "{0}_id".format(key) pk_fields.append(key_p) # check if the cls has a field with the class name, (causes conflict) if not cls.get_field_type(key): pk_fields.append(key) pk = request.DATA.get(pk_fields, None, field_type=int) # generic rest request handler if pk: using = getattr(cls.RestMeta, "RO_DATABASE", None) if using is None: using = getattr(cls.RestMeta, "DATABASE", None) if using: obj = cls.objects.using(using).filter(pk=pk).last() else: obj = cls.objects.filter(pk=pk).last() if not obj: return GRAPH_HELPERS.views.restNotFound(request) if request.method == "GET": return obj.on_rest_get(request) elif request.method == "POST": return obj.on_rest_post(request) elif request.method == "DELETE": return obj.on_rest_delete(request) return GRAPH_HELPERS.views.restNotFound(request) if request.method == "GET": return cls.on_rest_list(request) elif request.method == "POST": if request.DATA.get("rest_batch"): return cls.on_rest_batch(request, request.DATA.get("rest_batch")) return cls.on_rest_create(request) return GRAPH_HELPERS.views.restNotFound(request) @classmethod def searchFromRequest(cls, request, qset): '''returns None if not foreignkey, otherswise the relevant model''' search_fields = getattr(cls.RestMeta, "SEARCH_FIELDS", None) search_terms = getattr(cls.RestMeta, "SEARCH_TERMS", None) q = request.DATA.get(["search", "q"]) if q: sq = search.get_query(q, search_fields, search_terms) if sq: qset = qset.filter(sq) return qset @classmethod def rest_getWHITELISTED(cls): if hasattr(cls.RestMeta, "WHITELISTED"): return cls.RestMeta.WHITELISTED return cls.__RestMeta__.WHITELISTED @classmethod def rest_getQueryFields(cls, detailed=False): field_names = [] all_fields = True if hasattr(cls.RestMeta, "QUERY_FIELDS"): field_names = cls.RestMeta.QUERY_FIELDS all_fields = "all_fields" in field_names if all_fields: for f in cls._meta.fields: if not f.name.endswith("_ptr") or f in cls.rest_getWHITELISTED(): field_names.append(f.name) if issubclass(cls, MetaDataModel): if detailed: field_names.append("metadata") else: field_names.append("properties__key") field_names.append("properties__value") if detailed: output = [] for f in field_names: if f == "metadata": t = "MetaData" fm = None else: t = cls.get_field_type(f) fm = cls.get_fk_model(f) info = {} info["name"] = f info["type"] = t if fm: info["model"] = "{}.{}".format(fm._meta.app_label, fm.__name__) try: fd = cls._meta.get_field(f) if fd.choices: info["choices"] = fd.choices if fd.help_text: info["help"] = fd.help_text() except: pass output.append(info) return output return field_names @classmethod def filterFromRequest(cls, request, qset): '''returns None if not foreignkey, otherswise the relevant model''' field_names = cls.rest_getQueryFields() q = {} # check for customer filer filter = request.DATA.get("filter") # filter must be a dictionary if filter: """ we can do customer filters but the name must be a allowed field and can only be one level deep ie no double "__" "group__member__owner" html select: name: "user_filter" field: "filter" options: [ { label: "Staff Only", value: "is_staff:1" }, { label: "Online", value: "is_online:1" }, { label: "Online", value: "is_online:1" }, ] """ if not isinstance(filter, dict): filters = filter.split(';') filter = {} for f in filters: if ":" in f: k, v = f.split(':') if v in ["true", "True"]: v = True elif v in ["false", "False"]: v = False filter[k] = v now = datetime.now() # rest_helpers.log_print(field_names) for key in filter: name = key.split('__')[0] value = filter[key] if name in field_names and value != '__': if isinstance(value, str) and ':' in value and value.startswith('__'): k, v = value.split(':') key = key + k.strip() value = v.strip() if key.endswith("__in") and ',' in value: if value.startswith("["): value = value[1:-1] value = value.split(',') elif value in ["true", "True"]: value = True elif value in ["false", "False"]: value = False if isinstance(value, str) and "(" in value and ")" in value: # this is a special function call # rest_helpers.log_print(value) if value.startswith("days("): spos = value.find("(")+1 epos = value.find(")") # rest_helpers.log_print(int(value[spos:epos])) value = now + timedelta(days=int(value[spos:epos])) # rest_helpers.log_print(now) # rest_helpers.log_print(value) elif value.startswith("hours("): spos = value.find("(")+1 epos = value.find(")") value = now + timedelta(hours=int(value[spos:epos])) elif value.startswith("minutes("): spos = value.find("(")+1 epos = value.find(")") value = now + timedelta(minutes=int(value[spos:epos])) elif value.startswith("seconds("): spos = value.find("(")+1 epos = value.find(")") value = now + timedelta(seconds=int(value[spos:epos])) else: continue if key.count('__') <= 4: q[key] = value else: rest_helpers.log_print("filterFromRequest: invalid field: {} or {}".format(name, key)) if q: rest_helpers.log_print(q) qset = qset.filter(**q) return qset @classmethod def queryFromRequest(cls, request, qset): '''returns None if not foreignkey, otherswise the relevant model''' field_names = cls.rest_getQueryFields() # group_field = getattr(self.RestMeta, "GROUP_FIELD", "group") # elif fn in ["group", "merchant"]: should change to group_field q = {} for fn in field_names: v = None if fn in cls.rest_getWHITELISTED(): if fn in ["user", "member"]: Member = RestModel.getModel("account", "Member") v = Member.getFromRequest(request) elif fn == "terminal": Terminal = ForeignModel = cls.get_fk_model(fn) # Terminal = RestModel.getModel("payauth", "Terminal") if Terminal: tid = request.DATA.get("tid") if tid: v = Terminal.ro_objects().filter(tid=tid).last() else: v = Terminal.getFromRequest(request) else: v = request.DATA.get(fn) elif fn in ["group", "merchant"]: v = request.DATA.get(fn) if not v: if request.group: v = request.group elif getattr(request, "terminal", None): v = request.terminal.merchant elif fn == "start": # this is a reserved field # TODO internally change start to _start_ continue else: v = request.DATA.get(fn) if v is None: continue if (isinstance(v, str) or isinstance(v, str)) and ',' in v: v = [a.strip() for a in v.split(',')] q["{}__in".format(fn)] = v if isinstance(v, list): q["{}__in".format(fn)] = v elif v != None: q[fn] = v if q: # rest_helpers.log_print("queryFromRequest: {}".format(q)) qset = qset.filter(**q) return qset @classmethod def createFromRequest(cls, request, **kwargs): obj = cls() return obj.saveFromRequest(request, files=request.FILES, __is_new=True, **kwargs) @classmethod def createFromDict(cls, request, data, **kwargs): obj = cls() return obj.saveFromDict(request, data, __is_new=True, **kwargs) @classmethod def get_field_type(cls, fieldname): '''returns None if not foreignkey, otherswise the relevant model''' for field in cls._meta.fields: if fieldname == field.name: return field.get_internal_type() return None @classmethod def get_fk_model(cls, fieldname): '''returns None if not foreignkey, otherswise the relevant model''' try: field = cls._meta.get_field(fieldname) return field.related_model except: return None
63,592
17,907
#!/usr/bin/env python3 __copyright__ = 'Copyright 2013-2016, http://radical.rutgers.edu' __license__ = 'MIT' import sys import radical.utils as ru import radical.pilot as rp rpu = rp.utils # ------------------------------------------------------------------------------ # if __name__ == '__main__': if len(sys.argv) <= 1: print("\n\tusage: %s <session_id>\n") sys.exit(1) sid = sys.argv[1] profiles = rpu.fetch_profiles(sid=sid, skip_existing=True) for p in profiles: print(p) profs = ru.read_profiles(profiles) for p in profs: print(type(p)) prof = ru.combine_profiles(profs) print(len(prof)) for entry in prof: print(entry) # ------------------------------------------------------------------------------
803
281
try: from django.utils.deprecation import MiddlewareMixin except ImportError: # no-op class for Django < 1.10 class MiddlewareMixin(object): pass class BrokenRequestMiddleware(MiddlewareMixin): def process_request(self, request): raise ImportError('request') class BrokenResponseMiddleware(MiddlewareMixin): def process_response(self, request, response): raise ImportError('response') class BrokenViewMiddleware(MiddlewareMixin): def process_view(self, request, func, args, kwargs): raise ImportError('view') class MetricsNameOverrideMiddleware(MiddlewareMixin): def process_response(self, request, response): request._elasticapm_transaction_name = 'foobar' return response
758
219
#!/usr/bin/env python3.6 from user import User from credential import Credential def createUser(userName,password): ''' Function to create a new user ''' newUser = User(userName,password) return newUser def saveUsers(user): ''' Function to save users ''' user.saveUser() def createCredential(firstName,lastName,accountName,password): newCredential = Credential(firstName,lastName,accountName,password) return newCredential def saveCredential(credential): ''' Function to save a new credential ''' Credential.saveCredential(credential) def delCredential(credential): ''' Function to delete a credential ''' credential.deleteCredential() def findCredential(name): ''' Function that finds a credential by name returns the credential ''' return Credential.find_by_name(name) def check_existingCredentials(name): ''' Function that checks if a credential exists with that name and return a boolean ''' return Credential.credential_exist(name) def displayCredentials(): ''' Function that returns all the saved credentials ''' return Credential.displayCredentials() def main(): print("Hello Welcome to password locker.\n Login:") userName = input("What is your name?") password = input("Enter your password :") print(f"Hello {userName}. what would you like to do??\n Create an acount First!!" ) print("-"* 15) while True: print("Us this short codes : cc - Create a new credential, dc -display credentials, fc -to search a credential, dl -to delete credential, ex -exit the credential list ") short_code = input() if short_code == 'cc': print("New Credential") print("-"*10) print ("firstName ....") firstName = input() print("lastName ...") lastName = input() print("accountName ...") accountName = input() print("password ...") password = input() saveCredential(createCredential(firstName,lastName,accountName,password)) # create and save new credential. print('\n') # print (f'New Credential {firstName} {lastName} {accountName} created') print('\n') elif short_code == 'dc': if displayCredentials(): print("Here is a list of all your credentials") print('\n') for credential in displayCredentials(): print(f"{credential.firstName} {credential.lastName} ....{credential.accountName}") print('\n') else: print('\n') print("You dont seem to have any credentials saved yet") print('\n') elif short_code =='dl': print("Are your sure you want to delete this credential\n Please insert the name of the credential:") searchName = input() deleteCredential = Credential.deleteCredential(searchName) elif short_code == 'fc': print("Enter the name you want to search for") searchName = input() searchCredential = findCredential(searchName) print(f" {searchCredential.lastName}") print('-' * 20) print(f"accountName........{searchCredential.accountName}") elif short_code == "ex": print("Bye ......") break else: print("I really didn't get that. Please use the short codes") if __name__ == '__main__': main()
3,874
989
# constants used in more than one module # Some of them might find a better home later. from lib.nofloat import PRECISION MAIN_METASERVER_URL = open("cfg/metaserver.txt").read().strip() # old value used by some features (stats, ...) METASERVER_URL = "http://jlpo.free.fr/soundrts/metaserver/" # simulation VIRTUAL_TIME_INTERVAL = 300 # milliseconds COLLISION_RADIUS = 175 # millimeters # 350 / 2 USE_RANGE_MARGIN = 175 # millimeters ORDERS_QUEUE_LIMIT = 10 MAX_NB_OF_RESOURCE_TYPES = 10 DEFAULT_MINIMAL_DAMAGE = int(.17 * PRECISION) # used for packing the orders NEWLINE_REPLACEMENT = ";" SPACE_REPLACEMENT = "," # minimal interval (in seconds) between 2 sounds ALERT_LIMIT = .5 FOOTSTEP_LIMIT = .1 # don't play events after this limit (in seconds) EVENT_LIMIT = 3 # use the profiler (warning: will slow down the game) PROFILE = False
844
332
import tensorflow as tf import numpy as np ''' TF 2.0's basic attention layers(Attention and AdditiveAttention) calculate parallelly. TO USE MONOTONIC FUNCTION, ATTENTION MUST KNOW 'n-1 ALIGNMENT'. Thus, this parallel versions do not support the monotonic function. ''' class BahdanauAttention(tf.keras.layers.Layer): ''' Refer: https://www.tensorflow.org/tutorials/text/nmt_with_attention ''' def __init__(self, size): super(BahdanauAttention, self).__init__() self.size = size def build(self, input_shapes): self.layer_Dict = { 'Query': tf.keras.layers.Dense(self.size), 'Value': tf.keras.layers.Dense(self.size), 'V': tf.keras.layers.Dense(1) } self.built = True def call(self, inputs): ''' inputs: [queries, values] queries: [Batch, Query_dim] values: [Batch, T_v, Value_dim] ''' queries, values = inputs queries = self.layer_Dict['Query'](queries) #[Batch, Att_dim] values = self.layer_Dict['Value'](values) #[Batch, T_v, Att_dim] queries = tf.expand_dims(queries, 1) #[Batch, 1, Att_dim] score = self.layer_Dict['V'](tf.nn.tanh(values + queries)) #[Batch, T_v, 1] attention_weights = tf.nn.softmax(score - tf.reduce_max(score, axis= 1, keepdims= True), axis=1) #[Batch, T_v, 1] context_vector = tf.reduce_sum(attention_weights * values, axis=1) #[Batch, T_v, Att_dim] -> [Batch, Att_dim] return context_vector, tf.squeeze(attention_weights, axis= -1) def initial_alignment_fn(self, batch_size, key_time, dtype): return tf.zeros((batch_size, key_time), dtype= dtype) class BahdanauMonotonicAttention(tf.keras.layers.Layer): ''' Refer https://www.tensorflow.org/tutorials/text/nmt_with_attention https://github.com/tensorflow/addons/blob/v0.7.1/tensorflow_addons/seq2seq/attention_wrapper.py#L1004-L1175 ''' def __init__(self, size, sigmoid_noise= 0.0, normalize= False, **kwargs): super(BahdanauMonotonicAttention, self).__init__() self.size = size self.sigmoid_noise = sigmoid_noise self.normalize = normalize def build(self, input_shapes): self.layer_Dict = { 'Query': tf.keras.layers.Dense(self.size), 'Value': tf.keras.layers.Dense(self.size), 'Key': tf.keras.layers.Dense(self.size) } self.attention_v = self.add_weight( name='attention_v', shape=[self.size,], initializer='glorot_uniform', dtype=self.dtype, trainable=True ) self.attention_score_bias = self.add_weight( name='attention_score_bias', shape=[], initializer=tf.zeros_initializer(), dtype=self.dtype, trainable=True ) if self.normalize: self.attention_g = self.add_weight( name='attention_g', shape=[], initializer= tf.initializers.constant([np.sqrt(1. / self.size),]), dtype=self.dtype, trainable=True ) self.attention_b = self.add_weight( name='attention_b', shape=[self.size,], initializer= tf.zeros_initializer(), dtype=self.dtype, trainable=True ) self.bulit = True def call(self, inputs): ''' inputs: [queries, values, previous_alignments] or [queries, values, keys, previous_alignments] query: [Batch, Query_dim] value: [Batch, T_v, Value_dim] key: [Batch, T_v, Key_dim] previous_alignment: [Batch, T_v] ''' if len(inputs) == 3: query, value, previous_alignment = inputs elif len(inputs) == 4: query, value, key, previous_alignment = inputs else: raise ValueError('Unexpected input length') query = self.layer_Dict['Query'](query) # [Batch, Att_dim] value = self.layer_Dict['Value'](value) # [Batch, T_v, Att_dim] key = self.layer_Dict['Key'](key) if len(inputs) == 4 else value # [Batch, T_v, Att_dim] query = tf.expand_dims(query, 1) # [Batch, 1, Att_dim] previous_alignment = tf.expand_dims(previous_alignment, axis= 1) # [Batch, 1, T_v] score = self._calculate_scores(query= query, key= key) context, alignment = self._apply_scores( score= score, value= value, previous_alignment= previous_alignment ) # [Batch, Att_dim], [Batch, 1, T_v] return context, alignment def _calculate_scores(self, query, key): ''' Calculates attention scores as a nonlinear sum of query and key. Args: query: Query tensor of shape `[batch_size, 1, Att_dim]`. key: Key tensor of shape `[batch_size, T_k, Att_dim]`. Returns: Tensor of shape `[batch_size, T_k]`. ''' if self.normalize: norm_v = self.attention_g * self.attention_v * tf.math.rsqrt(tf.reduce_sum(tf.square(self.attention_v))) return tf.reduce_sum(norm_v * tf.tanh(query + key + self.attention_b), axis= -1) + self.attention_score_bias #[Batch, T_k, Att_dim] -> [Batch, T_k] else: return tf.reduce_sum(self.attention_v * tf.tanh(query + key), axis= -1) + self.attention_score_bias #[Batch, T_k, Att_dim] -> [Batch, T_k] def _apply_scores(self, score, value, previous_alignment): ''' score shape: [batch_size, T_v]`. (Must T_k == T_v) value shape: [batch_size, T_v, Att_dim]`. previous_alignment shape: [batch_size, 1, T_v]`. Return: [batch_size, Att_dim], [batch_size, T_v] ''' score = tf.expand_dims(score, axis= 1) #[Batch_size, 1, T_v] alignment = self._monotonic_probability_fn(score, previous_alignment) #[Batch_size, 1, T_v] context = tf.matmul(alignment, value) #[Batch_size, 1, Att_dim] return tf.squeeze(context, axis= 1), tf.squeeze(alignment, axis= 1) def _monotonic_probability_fn(self, score, previous_alignment): if self.sigmoid_noise > 0.0: score += self.sigmoid_noise * tf.random.normal(tf.shape(score), dtype= score.dtype) p_choose_i = tf.sigmoid(score) cumprod_1mp_choose_i = self.safe_cumprod(1 - p_choose_i, axis= 2, exclusive= True) alignment = p_choose_i * cumprod_1mp_choose_i * tf.cumsum( previous_alignment / tf.clip_by_value(cumprod_1mp_choose_i, 1e-10, 1.), axis= 2 ) return alignment # https://github.com/tensorflow/addons/blob/9e9031133c8362fedf40f2d05f00334b6f7a970b/tensorflow_addons/seq2seq/attention_wrapper.py#L810 def safe_cumprod(self, x, *args, **kwargs): """Computes cumprod of x in logspace using cumsum to avoid underflow. The cumprod function and its gradient can result in numerical instabilities when its argument has very small and/or zero values. As long as the argument is all positive, we can instead compute the cumulative product as exp(cumsum(log(x))). This function can be called identically to tf.cumprod. Args: x: Tensor to take the cumulative product of. *args: Passed on to cumsum; these are identical to those in cumprod. **kwargs: Passed on to cumsum; these are identical to those in cumprod. Returns: Cumulative product of x. """ x = tf.convert_to_tensor(x, name='x') tiny = np.finfo(x.dtype.as_numpy_dtype).tiny return tf.exp(tf.cumsum(tf.math.log(tf.clip_by_value(x, tiny, 1)), *args, **kwargs)) def initial_alignment_fn(self, batch_size, key_time, dtype): return tf.one_hot( indices= tf.zeros((batch_size), dtype= tf.int32), depth= key_time, dtype= dtype ) class StepwiseMonotonicAttention(BahdanauMonotonicAttention): ''' Refer: https://gist.github.com/dy-octa/38a7638f75c21479582d7391490df37c ''' def __init__(self, size, sigmoid_noise= 2.0, normalize= False, **kwargs): super(StepwiseMonotonicAttention, self).__init__(size, sigmoid_noise, normalize, **kwargs) def _monotonic_probability_fn(self, score, previous_alignment): ''' score: [Batch_size, 1, T_v] previous_alignment: [batch_size, 1, T_v] ''' if self.sigmoid_noise > 0.0: score += self.sigmoid_noise * tf.random.normal(tf.shape(score), dtype= score.dtype) p_choose_i = tf.sigmoid(score) # [Batch_size, 1, T_v] pad = tf.zeros([tf.shape(p_choose_i)[0], 1, 1], dtype=p_choose_i.dtype) # [Batch_size, 1, 1] alignment = previous_alignment * p_choose_i + tf.concat( [pad, previous_alignment[:, :, :-1] * (1.0 - p_choose_i[:, :, :-1])], axis= -1) return alignment
9,104
3,109
import faiss import torch import logging import numpy as np from tqdm import tqdm from torch.utils.data import DataLoader from torch.utils.data.dataset import Subset def test_efficient_ram_usage(args, eval_ds, model, test_method="hard_resize"): """This function gives the same output as test(), but uses much less RAM. This can be useful when testing with large descriptors (e.g. NetVLAD) on large datasets (e.g. San Francisco). Obviously it is slower than test(), and can't be used with PCA. """ model = model.eval() if test_method == 'nearest_crop' or test_method == "maj_voting": distances = np.empty([eval_ds.queries_num * 5, eval_ds.database_num], dtype=np.float32) else: distances = np.empty([eval_ds.queries_num, eval_ds.database_num], dtype=np.float32) with torch.no_grad(): if test_method == 'nearest_crop' or test_method == 'maj_voting': queries_features = np.ones((eval_ds.queries_num * 5, args.features_dim), dtype="float32") else: queries_features = np.ones((eval_ds.queries_num, args.features_dim), dtype="float32") logging.debug("Extracting queries features for evaluation/testing") queries_infer_batch_size = 1 if test_method == "single_query" else args.infer_batch_size eval_ds.test_method = test_method queries_subset_ds = Subset(eval_ds, list(range(eval_ds.database_num, eval_ds.database_num+eval_ds.queries_num))) queries_dataloader = DataLoader(dataset=queries_subset_ds, num_workers=args.num_workers, batch_size=queries_infer_batch_size, pin_memory=(args.device=="cuda")) for inputs, indices in tqdm(queries_dataloader, ncols=100): if test_method == "five_crops" or test_method == "nearest_crop" or test_method == 'maj_voting': inputs = torch.cat(tuple(inputs)) # shape = 5*bs x 3 x 480 x 480 features = model(inputs.to(args.device)) if test_method == "five_crops": # Compute mean along the 5 crops features = torch.stack(torch.split(features, 5)).mean(1) if test_method == "nearest_crop" or test_method == 'maj_voting': start_idx = (indices[0] - eval_ds.database_num) * 5 end_idx = start_idx + indices.shape[0] * 5 indices = np.arange(start_idx, end_idx) queries_features[indices, :] = features.cpu().numpy() else: queries_features[indices.numpy()-eval_ds.database_num, :] = features.cpu().numpy() queries_features = torch.tensor(queries_features).type(torch.float32).cuda() logging.debug("Extracting database features for evaluation/testing") # For database use "hard_resize", although it usually has no effect because database images have same resolution eval_ds.test_method = "hard_resize" database_subset_ds = Subset(eval_ds, list(range(eval_ds.database_num))) database_dataloader = DataLoader(dataset=database_subset_ds, num_workers=args.num_workers, batch_size=args.infer_batch_size, pin_memory=(args.device=="cuda")) for inputs, indices in tqdm(database_dataloader, ncols=100): inputs = inputs.to(args.device) features = model(inputs) for pn, (index, pred_feature) in enumerate(zip(indices, features)): distances[:, index] = ((queries_features-pred_feature)**2).sum(1).cpu().numpy() del features, queries_features, pred_feature predictions = distances.argsort(axis=1)[:, :max(args.recall_values)] if test_method == 'nearest_crop': distances = np.array([distances[row, index] for row, index in enumerate(predictions)]) distances = np.reshape(distances, (eval_ds.queries_num, 20 * 5)) predictions = np.reshape(predictions, (eval_ds.queries_num, 20 * 5)) for q in range(eval_ds.queries_num): # sort predictions by distance sort_idx = np.argsort(distances[q]) predictions[q] = predictions[q, sort_idx] # remove duplicated predictions, i.e. keep only the closest ones _, unique_idx = np.unique(predictions[q], return_index=True) # unique_idx is sorted based on the unique values, sort it again predictions[q, :20] = predictions[q, np.sort(unique_idx)][:20] predictions = predictions[:, :20] # keep only the closer 20 predictions for each elif test_method == 'maj_voting': distances = np.array([distances[row, index] for row, index in enumerate(predictions)]) distances = np.reshape(distances, (eval_ds.queries_num, 5, 20)) predictions = np.reshape(predictions, (eval_ds.queries_num, 5, 20)) for q in range(eval_ds.queries_num): # votings, modify distances in-place top_n_voting('top1', predictions[q], distances[q], args.majority_weight) top_n_voting('top5', predictions[q], distances[q], args.majority_weight) top_n_voting('top10', predictions[q], distances[q], args.majority_weight) # flatten dist and preds from 5, 20 -> 20*5 # and then proceed as usual to keep only first 20 dists = distances[q].flatten() preds = predictions[q].flatten() # sort predictions by distance sort_idx = np.argsort(dists) preds = preds[sort_idx] # remove duplicated predictions, i.e. keep only the closest ones _, unique_idx = np.unique(preds, return_index=True) # unique_idx is sorted based on the unique values, sort it again # here the row corresponding to the first crop is used as a # 'buffer' for each query, and in the end the dimension # relative to crops is eliminated predictions[q, 0, :20] = preds[np.sort(unique_idx)][:20] predictions = predictions[:, 0, :20] # keep only the closer 20 predictions for each query del distances #### For each query, check if the predictions are correct positives_per_query = eval_ds.get_positives() # args.recall_values by default is [1, 5, 10, 20] recalls = np.zeros(len(args.recall_values)) for query_index, pred in enumerate(predictions): for i, n in enumerate(args.recall_values): if np.any(np.in1d(pred[:n], positives_per_query[query_index])): recalls[i:] += 1 break recalls = recalls / eval_ds.queries_num * 100 recalls_str = ", ".join([f"R@{val}: {rec:.1f}" for val, rec in zip(args.recall_values, recalls)]) return recalls, recalls_str def test(args, eval_ds, model, test_method="hard_resize", pca=None): """Compute features of the given dataset and compute the recalls.""" assert test_method in ["hard_resize", "single_query", "central_crop", "five_crops", "nearest_crop", "maj_voting"], f"test_method can't be {test_method}" if args.efficient_ram_testing: return test_efficient_ram_usage(args, eval_ds, model, test_method) model = model.eval() with torch.no_grad(): logging.debug("Extracting database features for evaluation/testing") # For database use "hard_resize", although it usually has no effect because database images have same resolution eval_ds.test_method = "hard_resize" database_subset_ds = Subset(eval_ds, list(range(eval_ds.database_num))) database_dataloader = DataLoader(dataset=database_subset_ds, num_workers=args.num_workers, batch_size=args.infer_batch_size, pin_memory=(args.device=="cuda")) if test_method == "nearest_crop" or test_method == 'maj_voting': all_features = np.empty((5 * eval_ds.queries_num + eval_ds.database_num, args.features_dim), dtype="float32") else: all_features = np.empty((len(eval_ds), args.features_dim), dtype="float32") for inputs, indices in tqdm(database_dataloader, ncols=100): features = model(inputs.to(args.device)) features = features.cpu().numpy() if pca != None: features = pca.transform(features) all_features[indices.numpy(), :] = features logging.debug("Extracting queries features for evaluation/testing") queries_infer_batch_size = 1 if test_method == "single_query" else args.infer_batch_size eval_ds.test_method = test_method queries_subset_ds = Subset(eval_ds, list(range(eval_ds.database_num, eval_ds.database_num+eval_ds.queries_num))) queries_dataloader = DataLoader(dataset=queries_subset_ds, num_workers=args.num_workers, batch_size=queries_infer_batch_size, pin_memory=(args.device=="cuda")) for inputs, indices in tqdm(queries_dataloader, ncols=100): if test_method == "five_crops" or test_method == "nearest_crop" or test_method == 'maj_voting': inputs = torch.cat(tuple(inputs)) # shape = 5*bs x 3 x 480 x 480 features = model(inputs.to(args.device)) if test_method == "five_crops": # Compute mean along the 5 crops features = torch.stack(torch.split(features, 5)).mean(1) features = features.cpu().numpy() if pca != None: features = pca.transform(features) if test_method == "nearest_crop" or test_method == 'maj_voting': # store the features of all 5 crops start_idx = eval_ds.database_num + (indices[0] - eval_ds.database_num) * 5 end_idx = start_idx + indices.shape[0] * 5 indices = np.arange(start_idx, end_idx) all_features[indices, :] = features else: all_features[indices.numpy(), :] = features queries_features = all_features[eval_ds.database_num:] database_features = all_features[:eval_ds.database_num] faiss_index = faiss.IndexFlatL2(args.features_dim) faiss_index.add(database_features) del database_features, all_features logging.debug("Calculating recalls") distances, predictions = faiss_index.search(queries_features, max(args.recall_values)) if test_method == 'nearest_crop': distances = np.reshape(distances, (eval_ds.queries_num, 20 * 5)) predictions = np.reshape(predictions, (eval_ds.queries_num, 20 * 5)) for q in range(eval_ds.queries_num): # sort predictions by distance sort_idx = np.argsort(distances[q]) predictions[q] = predictions[q, sort_idx] # remove duplicated predictions, i.e. keep only the closest ones _, unique_idx = np.unique(predictions[q], return_index=True) # unique_idx is sorted based on the unique values, sort it again predictions[q, :20] = predictions[q, np.sort(unique_idx)][:20] predictions = predictions[:, :20] # keep only the closer 20 predictions for each query elif test_method == 'maj_voting': distances = np.reshape(distances, (eval_ds.queries_num, 5, 20)) predictions = np.reshape(predictions, (eval_ds.queries_num, 5, 20)) for q in range(eval_ds.queries_num): # votings, modify distances in-place top_n_voting('top1', predictions[q], distances[q], args.majority_weight) top_n_voting('top5', predictions[q], distances[q], args.majority_weight) top_n_voting('top10', predictions[q], distances[q], args.majority_weight) # flatten dist and preds from 5, 20 -> 20*5 # and then proceed as usual to keep only first 20 dists = distances[q].flatten() preds = predictions[q].flatten() # sort predictions by distance sort_idx = np.argsort(dists) preds = preds[sort_idx] # remove duplicated predictions, i.e. keep only the closest ones _, unique_idx = np.unique(preds, return_index=True) # unique_idx is sorted based on the unique values, sort it again # here the row corresponding to the first crop is used as a # 'buffer' for each query, and in the end the dimension # relative to crops is eliminated predictions[q, 0, :20] = preds[np.sort(unique_idx)][:20] predictions = predictions[:, 0, :20] # keep only the closer 20 predictions for each query #### For each query, check if the predictions are correct positives_per_query = eval_ds.get_positives() # args.recall_values by default is [1, 5, 10, 20] recalls = np.zeros(len(args.recall_values)) for query_index, pred in enumerate(predictions): for i, n in enumerate(args.recall_values): if np.any(np.in1d(pred[:n], positives_per_query[query_index])): recalls[i:] += 1 break # Divide by the number of queries*100, so the recalls are in percentages recalls = recalls / eval_ds.queries_num * 100 recalls_str = ", ".join([f"R@{val}: {rec:.1f}" for val, rec in zip(args.recall_values, recalls)]) return recalls, recalls_str def top_n_voting(topn, predictions, distances, maj_weight): if topn == 'top1': n = 1 selected = 0 elif topn == 'top5': n = 5 selected = slice(0, 5) elif topn == 'top10': n = 10 selected = slice(0, 10) # find predictions that repeat in the first, first five, # or fist ten columns for each crop vals, counts = np.unique(predictions[:, selected], return_counts=True) # for each prediction that repeats more than once, # subtract from its score for val, count in zip(vals[counts > 1], counts[counts > 1]): mask = (predictions[:, selected] == val) distances[:, selected][mask] -= maj_weight * count/n
14,002
4,341
import random li=[] for i in range(10):#creating a random list using code from blackboard li=li+[random.randrange(0,50)] print(li) def insertionSort(li):#insertion sort function i=1 while(i < len(li)): j = i while (j>0) and (li[j-1] > li[j]): li[j], li[j-1] = li[j-1], li[j] j = j-1 i = i+1 return li print(insertionSort(li)) #print(li)
355
170
import nose.tools from hyperopt import fmin, rand, tpe, hp, Trials, exceptions, space_eval def test_quadratic1_rand(): trials = Trials() argmin = fmin( fn=lambda x: (x - 3) ** 2, space=hp.uniform('x', -5, 5), algo=rand.suggest, max_evals=500, trials=trials) assert len(trials) == 500 assert abs(argmin['x'] - 3.0) < .25 def test_quadratic1_tpe(): trials = Trials() argmin = fmin( fn=lambda x: (x - 3) ** 2, space=hp.uniform('x', -5, 5), algo=tpe.suggest, max_evals=50, trials=trials) assert len(trials) == 50, len(trials) assert abs(argmin['x'] - 3.0) < .25, argmin @nose.tools.raises(exceptions.DuplicateLabel) def test_duplicate_label_is_error(): trials = Trials() def fn(xy): x, y = xy return x ** 2 + y ** 2 fmin(fn=fn, space=[ hp.uniform('x', -5, 5), hp.uniform('x', -5, 5), ], algo=rand.suggest, max_evals=500, trials=trials) def test_space_eval(): space = hp.choice('a', [ ('case 1', 1 + hp.lognormal('c1', 0, 1)), ('case 2', hp.uniform('c2', -10, 10)) ]) assert space_eval(space, {'a': 0, 'c1': 1.0}) == ('case 1', 2.0) assert space_eval(space, {'a': 1, 'c2': 3.5}) == ('case 2', 3.5)
1,442
585
# -*- coding: utf-8 -*- import os from datetime import timedelta from flask import Flask, session default_timezone = 'Europe/Berlin' app = Flask(__name__, static_folder='../static', static_url_path='/static', template_folder="../templates/") app.permanent_session_lifetime = timedelta(minutes=60) app.config.update( SESSION_COOKIE_NAME = "AAsession", ERROR_LOG_FILE = "%s/app.log" % os.environ.get('OPENSHIFT_LOG_DIR', 'logs') ) @app.before_request def session_activity(): session.modified = True @app.route('/robots.txt') def serve_robots(): return 'User-agent: *\nDisallow: /' # VIEWS import aaweb.views import aaweb.forms # API import aaweb.api # additional functionalities import aaweb.error import aaweb.log
737
268
#========================================================================== # # Copyright Insight Software Consortium # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0.txt # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # #==========================================================================*/ import InsightToolkit as itk import sys reader = itk.itkImageFileReaderF2_New() reader.SetFileName( sys.argv[1] ) diffusion = itk.itkGradientAnisotropicDiffusionImageFilterF2F2_New() diffusion.SetInput(reader.GetOutput()) diffusion.SetTimeStep(0.0625) diffusion.SetConductanceParameter(9.0) diffusion.SetNumberOfIterations( 5 ); gradient = itk.itkGradientMagnitudeImageFilterF2F2_New() gradient.SetInput(diffusion.GetOutput()) watershed = itk.itkWatershedImageFilterF2_New() watershed.SetInput(gradient.GetOutput()) watershed.SetThreshold(0.01) watershed.SetLevel(0.2) writer = itk.itkImageFileWriterUL2_New() writer.SetFileName( sys.argv[2] ) writer.SetInput( watershed.GetOutput() ) writer.Update()
1,528
499
import os import sys sys.path.append("../../../monk/"); import psutil from gluon_prototype import prototype gtf = prototype(verbose=1); gtf.Prototype("sample-project-1", "sample-experiment-1"); gtf.Default(dataset_path="../../../monk/system_check_tests/datasets/dataset_cats_dogs_train", model_name="resnet18_v1", freeze_base_network=True, num_epochs=2); ######################################################## Summary ##################################################### gtf.Summary() ########################################################################################################################### ##################################################### EDA - Find Num images per class ##################################### gtf.EDA(show_img=True, save_img=True); ########################################################################################################################### ##################################################### EDA - Find Missing and corrupted images ##################################### gtf.EDA(check_missing=True, check_corrupt=True); ########################################################################################################################### ##################################################### Estimate Training Time ##################################### gtf.Estimate_Train_Time(num_epochs=50); ###########################################################################################################################
1,520
334
cache_root = '/home/ubuntu/ppa-mirror/cache/' mirror_root = '/home/ubuntu/ppa-mirror/repo' http_proxy = "188.112.194.222:8080"
126
62
import sys, getopt import sched import time import csv from pprint import pprint import urllib, urllib2 from random import randint import threading proxies_file = "" targets_file = "" proxies = [] targets = [] scheduler = sched.scheduler(time.time, time.sleep) def pick_random_proxy(): proxy_count = len(proxies) - 1; if proxy_count == 0: return None proxy_index = randint(0, proxy_count ) proxy = proxies[ proxy_index ] return proxy[ "host" ] + ":" + proxy[ "port" ] def download_file(interval, url): threading.Thread( target = lambda: download_file_impl(interval, url) ) #randomize the interval new_interval = interval + randint( -1 * interval, interval ) if new_interval < interval: new_interval = interval #repeat itself forever scheduler.enter(new_interval, 1 , download_file, (new_interval, url) ) print "==> Next download of " + url + " in " + str( new_interval ) + " seconds" def download_file_impl(interval, url): selected_proxy = pick_random_proxy(); download_was_ok = True try: request = urllib2.Request(url) if selected_proxy is None: print "NO PROXY!" else: request.set_proxy(selected_proxy, 'http') response = urllib2.urlopen(request) print "Response code: " + str( response.code ) download_was_ok = response.code == 200 except urllib2.URLError, e: download_was_ok = False pprint( e ) if( download_was_ok ): print " OK! " else: print " NOK! " def main(argv): global scheduler parse_command_line_parameters(argv) proxiesReader = csv.DictReader(open(proxies_file), dialect='excel', delimiter=',') for row in proxiesReader: proxies.append( row ) targetsReader = csv.DictReader(open(targets_file), dialect='excel', delimiter=',') for row in targetsReader: targets.append( row ) print "===============================================================================" print "Proxies file: " + proxies_file print "Targets file: " + targets_file print "-------------------------------------------------------------------------------" print "Proxies (total:" + str( len(proxies) ) + ")" pprint( proxies ) print "Targets (total:" + str( len(targets) ) + ")" pprint( targets ) print "===============================================================================" for target in targets: interval = int( target[ "interval" ] ) url = target[ "url" ] scheduler.enter(interval, 1 , download_file, (interval, url ) ) scheduler.run() def parse_command_line_parameters(argv): global proxies_file global targets_file try: opts, args = getopt.getopt(argv,"hp:t:",["proxies=","targets="]) except getopt.GetoptError: print 'downloader.py -p <proxiesfile> -t <targetsfile>' sys.exit(2) for opt, arg in opts: if opt == '-h': print 'downloader.py -p <proxiesfile> -t <targetsfile>' sys.exit() elif opt in ("-p", "--proxiesfile"): proxies_file = arg elif opt in ("-t", "--targetsfile"): targets_file = arg if __name__ == "__main__": main(sys.argv[1:])
3,031
1,118
import random from django.conf import settings from django.db.models import Prefetch from django.db.models.functions.math import Radians, Cos, ACos, Sin from ..models import Factory, ReportRecord, Image, Document def _sample(objs, k): list_of_objs = list(objs) random.shuffle(list_of_objs) return list_of_objs[:k] def _get_nearby_factories(latitude, longitude, radius): """Return nearby factories based on position and search range.""" # ref: https://stackoverflow.com/questions/574691/mysql-great-circle-distance-haversine-formula distance = 6371 * ACos( Cos(Radians(latitude)) * Cos(Radians("lat")) * Cos(Radians("lng") - Radians(longitude)) + Sin(Radians(latitude)) * Sin(Radians("lat")) ) radius_km = radius ids = Factory.objects.annotate(distance=distance).only("id").filter(distance__lt=radius_km).order_by("id") if len(ids) > settings.MAX_FACTORY_PER_GET: ids = _sample(ids, settings.MAX_FACTORY_PER_GET) return ( Factory.objects.filter(id__in=[obj.id for obj in ids]) .prefetch_related(Prefetch('report_records', queryset=ReportRecord.objects.only("created_at").all())) .prefetch_related(Prefetch('images', queryset=Image.objects.only("id").all())) .prefetch_related(Prefetch('documents', queryset=Document.objects.only('created_at', 'display_status').all())) .all() ) def _get_client_ip(request): # ref: https://stackoverflow.com/a/30558984 x_forwarded_for = request.META.get("HTTP_X_FORWARDED_FOR") if x_forwarded_for: return x_forwarded_for.split(",")[-1].strip() elif request.META.get("HTTP_X_REAL_IP"): return request.META.get("HTTP_X_REAL_IP") else: return request.META.get("REMOTE_ADDR")
1,799
616
from __future__ import absolute_import from __future__ import print_function import pytest import std_srvs.srv as std_srvs # public decorators from pyros_schemas.ros import with_service_schemas # # Testing with_service_schemas decorator # @with_service_schemas(std_srvs.Trigger) def service_callback(data, data_dict, error): # From spec http://docs.ros.org/api/std_srvs/html/srv/Trigger.html assert len(data_dict) == len(data.__slots__) == 0 return { 'success': True, 'message': 'fortytwo', } def test_decorated_service(): resp = service_callback(std_srvs.TriggerRequest()) assert isinstance(resp, std_srvs.TriggerResponse) assert resp.success == True assert resp.message == 'fortytwo' # Just in case we run this directly if __name__ == '__main__': pytest.main([ 'test_decorators.py::test_decorated_service' ])
886
294
""" @param: A: A list of integers @param: elem: An integer @return: The new length after remove """ #倒序遍历list def removeElement(self, A, elem): # write your code here for i in range(len(A)-1,-1,-1): if A[i] == elem: A.pop(i) return len(A) #遍历拷贝的list ,操作原始list def removeElement(self, A, elem): # write your code here for i in A[:]: if i == elem: A.remove(i) return len(A)
516
188
from setuptools import setup, find_packages import io from collections import OrderedDict with io.open('README.rst', 'rt', encoding='utf8') as f: readme = f.read() def get_version(): return '0.1.1' setup( name='Flask-Wired', version=get_version(), license='MIT', author='Karthikkannan Maruthamuthu', author_email='karthikkannan@gmail.com', maintainer='Karthikkannan Maruthamuthu', maintainer_email='karthikkannan@gmail.com', description='Package for Flask wiring.', long_description=readme, url='https://github.com/treebohotels/Flask-Wired', project_urls=OrderedDict(( ('Documentation', 'https://github.com/treebohotels/Flask-Wired'), ('Code', 'https://github.com/treebohotels/Flask-Wired'), ('Issue tracker', 'https://github.com/treebohotels/Flask-Wired/issues'), )), package_dir={'': '.'}, packages=find_packages(".", exclude=['tests', 'sample_app']), include_package_data=True, zip_safe=False, platforms='any', python_requires='>=3.6', install_requires=[ 'Flask==1.0.2', 'Flask-Script==2.0.6', 'Flask-Migrate==2.2.1', 'flask-marshmallow==0.9.0', 'Flask-SQLAlchemy==2.3.2', 'marshmallow-sqlalchemy==0.14.1', 'psycopg2==2.7.5', ], entry_points={ }, test_suite="tests", classifiers=[ 'Development Status :: 1 - Planning', 'Environment :: Web Environment', 'Framework :: Flask', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Internet :: WWW/HTTP :: WSGI :: Application', 'Topic :: Software Development :: Libraries :: Application Frameworks', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
2,057
696
from flask import jsonify def auth_login(http_status_code, message, token): view = jsonify({'status_code': http_status_code, 'message': message, '_token': token}) return view def auth_is_valid(http_status_code, message, token): view = jsonify({'status_code': http_status_code, 'message': message, '_token': token}) return view def auth_logout(http_status_code, message, token): view = jsonify({'status_code': http_status_code, 'message': message, '_token': token}) return view
510
163
# -*- coding: utf-8 -*- try: from setuptools import setup, find_packages except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages REQUIRES = ['six', 'pecan', 'setuptools >= 1.1.6', 'cassandra-driver', 'pymongo'] setup( name='deuce', version='0.1', description='Deuce - Block-level de-duplication as-a-service', author='Rackspace', author_email='', install_requires=REQUIRES, test_suite='deuce', zip_safe=False, include_package_data=True, packages=find_packages(exclude=['tests']) )
604
216
import telepot import time import requests from bs4 import BeautifulSoup as bs import cPickle import csv RAHUL_ID = 931906767 # You can leave this bit out if you're using a paid PythonAnywhere account # proxy_url = "http://proxy.server:3128" # telepot.api._pools = { # 'default': urllib3.ProxyManager(proxy_url=proxy_url, num_pools=3, maxsize=10, retries=False, timeout=30), # } # telepot.api._onetime_pool_spec = (urllib3.ProxyManager, dict(proxy_url=proxy_url, num_pools=1, maxsize=1, retries=False, timeout=30)) # end of the stuff that's only needed for free accounts ######################## login_url = 'https://www.placement.iitbhu.ac.in/accounts/login/' client = requests.session() login = client.get(login_url) login = bs(login.content, "html.parser") payload = { "login": "rahul.kumar.cse15@itbhu.ac.in", "password": "rahulkmr", "csrfmiddlewaretoken": login.input['value'] } result = client.post( login_url, data = payload, headers = dict(referer=login_url) ) forum = client.get("https://www.placement.iitbhu.ac.in/forum/c/notice-board/2019-20/") soup = bs(forum.content, "html.parser") #load last message delivred to users try: with open("posts", "rb") as f: posts = cPickle.load(f); except Exception as e: print e posts = soup.findAll("td", "topic-name") for i in range(len(posts)): posts[i] = posts[i].a posts.pop(0) posts.pop(0) updated = soup.findAll('td','topic-last-post') # updated.pop() # updated.pop(0) ######################### bot = telepot.Bot('940251504:AAG19YYQYtkiEOCrW0fZETvmYQSskElARcc') # chat_ids = {RAHUL_ID} with open("IDs", "rb") as f: chat_ids = cPickle.load(f) print '#################No of IDs loaded: ', len(chat_ids) ####### Commands ######## def start(msg): # with open("users.csv", "w") as f: # writer = csv.writer(f) # writer.writerow(msg['from'].values()) bot.sendMessage(msg['chat']['id'],"Hello " + msg['from']['first_name']) def add_cmd(chat_id, msg, *argv): if chat_id not in chat_ids: chat_ids.add(chat_id) with open("IDs", "wb") as f: cPickle.dump(chat_ids, f); with open("users.txt", "a") as f: writer = csv.writer(f) writer.writerow(msg['from'].values()) bot.sendMessage(chat_id, "Added your ID for notifications. Note that it may take upto 5 minutes to get update of a recent post") bot.sendMessage(RAHUL_ID, "Added:\n" + str(msg['from'].values())) else: bot.sendMessage(chat_id, "You are already added") def remove_cmd(chat_id, *argv): try: chat_ids.remove(chat_id) with open("IDs", "wb") as f: cPickle.dump(chat_ids, f); bot.sendMessage(chat_id, "Removed your ID") except KeyError: bot.sendMessage(chat_id, "You are not in the list") def allPosts(chat_id, *argv): msg = '' for i in range(len(posts)): msg += gen_msg(posts[i]) + '\n<b>Last Updated: </b>' + updated[i].string.encode() + '\n\n' bot.sendMessage(chat_id, text=msg, parse_mode="HTML") def top(chat_id, param, *argv): total = 3 if len(param) > 1 and param[1].isdigit(): total = min(15, int(param[1])) msg = '.' for i in range(total): msg += gen_msg(posts[i]) try: # print msg bot.sendMessage(chat_id, text=msg, parse_mode="HTML") except Exception as e: bot.sendMessage(chat_id, text=str(e), parse_mode="HTML") ######################### command = {'/add':add_cmd, '/remove':remove_cmd, '/all':allPosts, '/recent':top} def handle(msg): # print msg content_type, chat_type, chat_id = telepot.glance(msg) print msg['from']['first_name'], chat_id if content_type == 'text': if msg['text'][0] == '/': tokens = msg['text'].split() try: if tokens[0] == '/start': start(msg) elif tokens[0] == '/add': add_cmd(chat_id, msg) else: command[tokens[0]](chat_id, tokens) except KeyError: bot.sendMessage(chat_id, "Unknown command: {}".format(tokens[0])) else: bot.sendMessage(chat_id, "You said '{}'".format(msg["text"])) bot.message_loop(handle) print ('Listening ...') # for chat_id in chat_ids: # bot.sendMessage(chat_id, text='Server started', parse_mode="HTML") bot.sendMessage(RAHUL_ID, text='Server started', parse_mode="HTML") def gen_msg(post): string = str(post) string = string[:8] + '"https://www.placement.iitbhu.ac.in' + string[9:] + '\n-----------------\n<b>Last Updated: </b>' + updated[posts.index(post)].string.encode() + '\n' # string += ' post = client.get("https://www.placement.iitbhu.ac.in/" + post['href']) post = bs(post.content, "html.parser") post = post.find("td", "post-content") # print post.contents for x in post.contents: if type(x) is type(post.contents[0]): string += x + '\n' post = post.find("div", "attachments").a if post is not None: tmp = '<a href=' + '"https://www.placement.iitbhu.ac.in' + post['href'] + '">' tmp += post.contents[1].split()[0] tmp += '</a>' string += tmp string += '\n\n' return string def on_new(): global updated global posts posts2 = soup.findAll("td", "topic-name") for i in range(len(posts2)): posts2[i] = posts2[i].a #find how many new posts try: total = posts2.index(posts[0]) except ValueError: total = len(posts2) print total, "new posts, users = ", len(chat_ids) posts = posts2 updated = soup.findAll('td','topic-last-post') msg = '<b>Note that you need to be logged in before opening these links, else youll see 500 error in your browser</b>\n\n' for i in range(total): msg += gen_msg(posts[i]) for chat_id in chat_ids: # print "sending update to ", chat_id bot.sendMessage(chat_id, text=msg, parse_mode="HTML") #save last message delivred to users with open("posts", "wb") as f: cPickle.dump(posts, f); # Keep the program running. def main(): while 1: # bot.sendMessage(RAHUL_ID, text="Dynamic code update", parse_mode="HTML") global forum global soup try: forum = requests.get("https://www.placement.iitbhu.ac.in/forum/c/notice-board/2019-20/") soup = bs(forum.content, "html.parser") if len(posts) == 0 or soup.td.a['href'] != posts[0]['href']: on_new() except Exception as e: bot.sendMessage(RAHUL_ID, text="<b>Exception:</b>\n" + str(e), parse_mode="HTML") # else: # bot.sendMessage(RAHUL_ID, text="Error in polling TPO forum", parse_mode="HTML") try: time.sleep(1000 * 60 *1) finally: # for chat_id in chat_ids: # bot.sendMessage(chat_id, text='Server closing for maintenance, you might miss updates', parse_mode="HTML") bot.sendMessage(RAHUL_ID, text='Server closing for maintenance, you might miss updates', parse_mode="HTML") if __name__ == '__main__': main()
6,580
2,693
""" Train ===== Defines functions which train models and write model artifacts to disk. """ from __future__ import print_function import os import tempfile import tensorflow as tf from tensorflow.examples.tutorials.mnist import input_data from tensorflow_mnist import model, paths def train(path): """ Train a decision tree classifier using a floating point feature matrix and a categorical classification target. Arguments: path (str): The path indicating where to save the final model artifacts """ # Construct the model graph graph, x, y, step, initializer, accuracy, prediction = model.build() # Start a training session with tf.Session(graph=graph) as sess: # Initialize the graph sess.run(initializer) # Train the model for 1000 steps mnist = input_data.read_data_sets(tempfile.mkdtemp(), one_hot=True) for _ in range(1000): batch_xs, batch_ys = mnist.train.next_batch(100) sess.run(step, feed_dict={x: batch_xs, y: batch_ys}) # Display accuracy measurement print(sess.run(accuracy, feed_dict={x: mnist.test.images, y: mnist.test.labels})) # Save the variable data to disk os.makedirs(path) saver = tf.train.Saver() saver.save(sess, path) print('Success!') def main(): """ Load features and labels, train the neural network, and serialize model artifact. Note: This is the training entrypoint used by baklava! """ path = paths.model('mnist') train(path)
1,607
468
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import proto # type: ignore from google.protobuf import timestamp_pb2 # type: ignore __protobuf__ = proto.module( package='google.cloud.binaryauthorization.v1', manifest={ 'Policy', 'AdmissionWhitelistPattern', 'AdmissionRule', 'Attestor', 'UserOwnedGrafeasNote', 'PkixPublicKey', 'AttestorPublicKey', }, ) class Policy(proto.Message): r"""A [policy][google.cloud.binaryauthorization.v1.Policy] for container image binary authorization. Attributes: name (str): Output only. The resource name, in the format ``projects/*/policy``. There is at most one policy per project. description (str): Optional. A descriptive comment. global_policy_evaluation_mode (google.cloud.binaryauthorization_v1.types.Policy.GlobalPolicyEvaluationMode): Optional. Controls the evaluation of a Google-maintained global admission policy for common system-level images. Images not covered by the global policy will be subject to the project admission policy. This setting has no effect when specified inside a global admission policy. admission_whitelist_patterns (Sequence[google.cloud.binaryauthorization_v1.types.AdmissionWhitelistPattern]): Optional. Admission policy allowlisting. A matching admission request will always be permitted. This feature is typically used to exclude Google or third-party infrastructure images from Binary Authorization policies. cluster_admission_rules (Sequence[google.cloud.binaryauthorization_v1.types.Policy.ClusterAdmissionRulesEntry]): Optional. Per-cluster admission rules. Cluster spec format: ``location.clusterId``. There can be at most one admission rule per cluster spec. A ``location`` is either a compute zone (e.g. us-central1-a) or a region (e.g. us-central1). For ``clusterId`` syntax restrictions see https://cloud.google.com/container-engine/reference/rest/v1/projects.zones.clusters. kubernetes_namespace_admission_rules (Sequence[google.cloud.binaryauthorization_v1.types.Policy.KubernetesNamespaceAdmissionRulesEntry]): Optional. Per-kubernetes-namespace admission rules. K8s namespace spec format: [a-z.-]+, e.g. 'some-namespace' kubernetes_service_account_admission_rules (Sequence[google.cloud.binaryauthorization_v1.types.Policy.KubernetesServiceAccountAdmissionRulesEntry]): Optional. Per-kubernetes-service-account admission rules. Service account spec format: ``namespace:serviceaccount``. e.g. 'test-ns:default' istio_service_identity_admission_rules (Sequence[google.cloud.binaryauthorization_v1.types.Policy.IstioServiceIdentityAdmissionRulesEntry]): Optional. Per-istio-service-identity admission rules. Istio service identity spec format: spiffe://<domain>/ns/<namespace>/sa/<serviceaccount> or <domain>/ns/<namespace>/sa/<serviceaccount> e.g. spiffe://example.com/ns/test-ns/sa/default default_admission_rule (google.cloud.binaryauthorization_v1.types.AdmissionRule): Required. Default admission rule for a cluster without a per-cluster, per- kubernetes- service-account, or per-istio-service-identity admission rule. update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. Time when the policy was last updated. """ class GlobalPolicyEvaluationMode(proto.Enum): r"""""" GLOBAL_POLICY_EVALUATION_MODE_UNSPECIFIED = 0 ENABLE = 1 DISABLE = 2 name = proto.Field( proto.STRING, number=1, ) description = proto.Field( proto.STRING, number=6, ) global_policy_evaluation_mode = proto.Field( proto.ENUM, number=7, enum=GlobalPolicyEvaluationMode, ) admission_whitelist_patterns = proto.RepeatedField( proto.MESSAGE, number=2, message='AdmissionWhitelistPattern', ) cluster_admission_rules = proto.MapField( proto.STRING, proto.MESSAGE, number=3, message='AdmissionRule', ) kubernetes_namespace_admission_rules = proto.MapField( proto.STRING, proto.MESSAGE, number=10, message='AdmissionRule', ) kubernetes_service_account_admission_rules = proto.MapField( proto.STRING, proto.MESSAGE, number=8, message='AdmissionRule', ) istio_service_identity_admission_rules = proto.MapField( proto.STRING, proto.MESSAGE, number=9, message='AdmissionRule', ) default_admission_rule = proto.Field( proto.MESSAGE, number=4, message='AdmissionRule', ) update_time = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) class AdmissionWhitelistPattern(proto.Message): r"""An [admission allowlist pattern][google.cloud.binaryauthorization.v1.AdmissionWhitelistPattern] exempts images from checks by [admission rules][google.cloud.binaryauthorization.v1.AdmissionRule]. Attributes: name_pattern (str): An image name pattern to allowlist, in the form ``registry/path/to/image``. This supports a trailing ``*`` wildcard, but this is allowed only in text after the ``registry/`` part. This also supports a trailing ``**`` wildcard which matches subdirectories of a given entry. """ name_pattern = proto.Field( proto.STRING, number=1, ) class AdmissionRule(proto.Message): r"""An [admission rule][google.cloud.binaryauthorization.v1.AdmissionRule] specifies either that all container images used in a pod creation request must be attested to by one or more [attestors][google.cloud.binaryauthorization.v1.Attestor], that all pod creations will be allowed, or that all pod creations will be denied. Images matching an [admission allowlist pattern][google.cloud.binaryauthorization.v1.AdmissionWhitelistPattern] are exempted from admission rules and will never block a pod creation. Attributes: evaluation_mode (google.cloud.binaryauthorization_v1.types.AdmissionRule.EvaluationMode): Required. How this admission rule will be evaluated. require_attestations_by (Sequence[str]): Optional. The resource names of the attestors that must attest to a container image, in the format ``projects/*/attestors/*``. Each attestor must exist before a policy can reference it. To add an attestor to a policy the principal issuing the policy change request must be able to read the attestor resource. Note: this field must be non-empty when the evaluation_mode field specifies REQUIRE_ATTESTATION, otherwise it must be empty. enforcement_mode (google.cloud.binaryauthorization_v1.types.AdmissionRule.EnforcementMode): Required. The action when a pod creation is denied by the admission rule. """ class EvaluationMode(proto.Enum): r"""""" EVALUATION_MODE_UNSPECIFIED = 0 ALWAYS_ALLOW = 1 REQUIRE_ATTESTATION = 2 ALWAYS_DENY = 3 class EnforcementMode(proto.Enum): r"""Defines the possible actions when a pod creation is denied by an admission rule. """ ENFORCEMENT_MODE_UNSPECIFIED = 0 ENFORCED_BLOCK_AND_AUDIT_LOG = 1 DRYRUN_AUDIT_LOG_ONLY = 2 evaluation_mode = proto.Field( proto.ENUM, number=1, enum=EvaluationMode, ) require_attestations_by = proto.RepeatedField( proto.STRING, number=2, ) enforcement_mode = proto.Field( proto.ENUM, number=3, enum=EnforcementMode, ) class Attestor(proto.Message): r"""An [attestor][google.cloud.binaryauthorization.v1.Attestor] that attests to container image artifacts. An existing attestor cannot be modified except where indicated. Attributes: name (str): Required. The resource name, in the format: ``projects/*/attestors/*``. This field may not be updated. description (str): Optional. A descriptive comment. This field may be updated. The field may be displayed in chooser dialogs. user_owned_grafeas_note (google.cloud.binaryauthorization_v1.types.UserOwnedGrafeasNote): This specifies how an attestation will be read, and how it will be used during policy enforcement. update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. Time when the attestor was last updated. """ name = proto.Field( proto.STRING, number=1, ) description = proto.Field( proto.STRING, number=6, ) user_owned_grafeas_note = proto.Field( proto.MESSAGE, number=3, oneof='attestor_type', message='UserOwnedGrafeasNote', ) update_time = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) class UserOwnedGrafeasNote(proto.Message): r"""An [user owned Grafeas note][google.cloud.binaryauthorization.v1.UserOwnedGrafeasNote] references a Grafeas Attestation.Authority Note created by the user. Attributes: note_reference (str): Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: ``projects/*/notes/*``. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. public_keys (Sequence[google.cloud.binaryauthorization_v1.types.AttestorPublicKey]): Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. delegation_service_account_email (str): Output only. This field will contain the service account email address that this Attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the [note_reference][Note] in Container Analysis (``containeranalysis.notes.occurrences.viewer``). This email address is fixed for the lifetime of the Attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. """ note_reference = proto.Field( proto.STRING, number=1, ) public_keys = proto.RepeatedField( proto.MESSAGE, number=2, message='AttestorPublicKey', ) delegation_service_account_email = proto.Field( proto.STRING, number=3, ) class PkixPublicKey(proto.Message): r"""A public key in the PkixPublicKey format (see https://tools.ietf.org/html/rfc5280#section-4.1.2.7 for details). Public keys of this type are typically textually encoded using the PEM format. Attributes: public_key_pem (str): A PEM-encoded public key, as described in https://tools.ietf.org/html/rfc7468#section-13 signature_algorithm (google.cloud.binaryauthorization_v1.types.PkixPublicKey.SignatureAlgorithm): The signature algorithm used to verify a message against a signature using this key. These signature algorithm must match the structure and any object identifiers encoded in ``public_key_pem`` (i.e. this algorithm must match that of the public key). """ class SignatureAlgorithm(proto.Enum): r"""Represents a signature algorithm and other information necessary to verify signatures with a given public key. This is based primarily on the public key types supported by Tink's PemKeyType, which is in turn based on KMS's supported signing algorithms. See https://cloud.google.com/kms/docs/algorithms. In the future, BinAuthz might support additional public key types independently of Tink and/or KMS. """ _pb_options = {'allow_alias': True} SIGNATURE_ALGORITHM_UNSPECIFIED = 0 RSA_PSS_2048_SHA256 = 1 RSA_PSS_3072_SHA256 = 2 RSA_PSS_4096_SHA256 = 3 RSA_PSS_4096_SHA512 = 4 RSA_SIGN_PKCS1_2048_SHA256 = 5 RSA_SIGN_PKCS1_3072_SHA256 = 6 RSA_SIGN_PKCS1_4096_SHA256 = 7 RSA_SIGN_PKCS1_4096_SHA512 = 8 ECDSA_P256_SHA256 = 9 EC_SIGN_P256_SHA256 = 9 ECDSA_P384_SHA384 = 10 EC_SIGN_P384_SHA384 = 10 ECDSA_P521_SHA512 = 11 EC_SIGN_P521_SHA512 = 11 public_key_pem = proto.Field( proto.STRING, number=1, ) signature_algorithm = proto.Field( proto.ENUM, number=2, enum=SignatureAlgorithm, ) class AttestorPublicKey(proto.Message): r"""An [attestor public key][google.cloud.binaryauthorization.v1.AttestorPublicKey] that will be used to verify attestations signed by this attestor. Attributes: comment (str): Optional. A descriptive comment. This field may be updated. id (str): The ID of this public key. Signatures verified by BinAuthz must include the ID of the public key that can be used to verify them, and that ID must match the contents of this field exactly. Additional restrictions on this field can be imposed based on which public key type is encapsulated. See the documentation on ``public_key`` cases below for details. ascii_armored_pgp_public_key (str): ASCII-armored representation of a PGP public key, as the entire output by the command ``gpg --export --armor foo@example.com`` (either LF or CRLF line endings). When using this field, ``id`` should be left blank. The BinAuthz API handlers will calculate the ID and fill it in automatically. BinAuthz computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If ``id`` is provided by the caller, it will be overwritten by the API-calculated ID. pkix_public_key (google.cloud.binaryauthorization_v1.types.PkixPublicKey): A raw PKIX SubjectPublicKeyInfo format public key. NOTE: ``id`` may be explicitly provided by the caller when using this type of public key, but it MUST be a valid RFC3986 URI. If ``id`` is left blank, a default one will be computed based on the digest of the DER encoding of the public key. """ comment = proto.Field( proto.STRING, number=1, ) id = proto.Field( proto.STRING, number=2, ) ascii_armored_pgp_public_key = proto.Field( proto.STRING, number=3, oneof='public_key', ) pkix_public_key = proto.Field( proto.MESSAGE, number=5, oneof='public_key', message='PkixPublicKey', ) __all__ = tuple(sorted(__protobuf__.manifest))
16,942
4,826
# Copyright (c) 2006-2010, Jesse Liesch # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the author nor the # names of its contributors may be used to endorse or promote products # derived from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE IMPLIED # DISCLAIMED. IN NO EVENT SHALL JESSE LIESCH BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. from PyQt4.QtCore import * from PyQt4.QtGui import * import copy from editGrid import * from plugin import * from portfolio import * import appGlobal class SpendingModel(EditGridModel): def __init__(self, parent = None, *args): EditGridModel.__init__(self, parent, *args) self.days = 30 self.categorized = True def rebuildSpending(self): portfolio = appGlobal.getApp().portfolio table = portfolio.getSpendingTable(days = self.days, categorize = self.categorized) self.setColumns(table[0]) self.setData(table[1]) class SpendingWidget(QWidget): def __init__(self, parent): QWidget.__init__(self, parent) self.model = SpendingModel(self) portfolio = appGlobal.getApp().portfolio vbox = QVBoxLayout(self) vbox.setMargin(0) hor = QHBoxLayout() hor.setMargin(0) vbox.addLayout(hor) hor.addWidget(QLabel("Period:")) self.periods = ["One Week", "One Month", "Three Months", "One Year", "Two Years", "Three Years", "Five Years", "Ten Years", "Portfolio Inception"] value = portfolio.portPrefs.getPositionPeriod() self.period = QComboBox() self.period.addItems(self.periods) if value in self.periods: self.period.setCurrentIndex(self.periods.index(value)) else: self.period.setCurrentIndex(self.periods.index("Portfolio Inception")) hor.addWidget(self.period) self.connect(self.period, SIGNAL("currentIndexChanged(int)"), self.newPeriod) showCategorized = QCheckBox("Show Categorized") if True or portfolio.portPrefs.getPerformanceCurrent(): showCategorized.setChecked(True) self.model.categorized = True hor.addWidget(showCategorized) # Redraw when checkbox is changed self.connect(showCategorized, SIGNAL("stateChanged(int)"), self.changeCategorized) hor.addStretch(1000) self.table = EditGrid(self.model) self.newPeriod() vbox.addWidget(self.table) self.table.setSortingEnabled(True) self.table.sortByColumn(1, Qt.DescendingOrder) self.table.resizeColumnsToContents() def newPeriod(self): period = self.periods[self.period.currentIndex()] appGlobal.getApp().portfolio.portPrefs.setPositionPeriod(period) period = self.periods[self.period.currentIndex()] if period == "One Week": self.model.days = 7 elif period == "One Month": self.model.days = 30 elif period == "Three Months": self.model.days = 91 elif period == "One Year": self.model.days = 365 elif period == "Two Years": self.model.days = 365 * 2 elif period == "Three Years": self.model.days = 365 * 3 elif period == "Five Years": self.model.days = 365 * 5 + 1 elif period == "Ten Years": self.model.days = 365 * 10 + 2 else: self.model.days = 365 * 100 # 100 years self.model.rebuildSpending() self.table.resizeColumnsToContents() def changeCategorized(self, state): self.model.categorized = state != 0 #appGlobal.getApp().portfolio.portPrefs.setPerformanceCurrent(self.model.current) self.model.rebuildSpending() self.table.resizeColumnsToContents() class Plugin(PluginBase): def __init__(self): PluginBase.__init__(self) def name(self): return "Spending" def icarraVersion(self): return (0, 0, 0) def version(self): return (1, 0, 0) def forInvestment(self): return False def createWidget(self, parent): return SpendingWidget(parent)
4,841
1,805
import torch from torch.distributions import Gamma import torch.nn.functional as F import matplotlib.pyplot as plt from tqdm import tqdm from pyro.distributions import * import pyro from pyro.optim import Adam from pyro.infer import SVI, Trace_ELBO, Predictive assert pyro.__version__.startswith('1') pyro.enable_validation(True) pyro.set_rng_seed(1337) torch.set_num_threads(1) # device = torch.device("cuda:0") if torch.cuda.is_available() else torch.device("cpu") device = torch.device("cpu") data = torch.cat((MultivariateNormal(-2 * torch.ones(2), 0.1 * torch.eye(2)).sample([25]), MultivariateNormal(2 * torch.ones(2), 0.1 * torch.eye(2)).sample([25]), MultivariateNormal(torch.tensor([0., 0.]), 0.1 * torch.eye(2)).sample([25]))) data = data.to(device) N = data.shape[0] D = data.shape[1] def mix_weights(beta): beta1m_cumprod = (1 - beta).cumprod(-1) return F.pad(beta, (0, 1), value=1) * F.pad(beta1m_cumprod, (1, 0), value=1) def model(data, **kwargs): with pyro.plate("beta_plate", T - 1): beta = pyro.sample("beta", Beta(1, alpha)) zeta = 2. * torch.ones(T * D, device=device) delta = 2. * torch.ones(T * D, device=device) with pyro.plate("prec_plate", T * D): prec = pyro.sample("prec", Gamma(zeta, delta)) for t in pyro.plate("corr_chol_plate", T): corr_chol[t, ...] = pyro.sample("corr_chol_{}".format(t), LKJCorrCholesky(d=D, eta=torch.ones(1, device=device))) with pyro.plate("mu_plate", T): _std = torch.sqrt(1. / prec.view(-1, D)) sigma_chol = torch.bmm(torch.diag_embed(_std), corr_chol) mu = pyro.sample("mu", MultivariateNormal(torch.zeros(T, D, device=device), scale_tril=sigma_chol)) with pyro.plate("data", N): z = pyro.sample("z", Categorical(mix_weights(beta))) pyro.sample("obs", MultivariateNormal(mu[z], scale_tril=sigma_chol[z]), obs=data) def guide(data, **kwargs): gamma = pyro.param('gamma', alpha * torch.ones(T - 1, device=device), constraint=constraints.positive) zeta = pyro.param('zeta', lambda: Uniform(1., 2.).sample([T * D]).to(device), constraint=constraints.positive) delta = pyro.param('delta', lambda: Uniform(1., 2.).sample([T * D]).to(device), constraint=constraints.positive) psi = pyro.param('psi', lambda: Uniform(1., 2.).sample([T]).to(device), constraint=constraints.positive) tau = pyro.param('tau', lambda: MultivariateNormal(torch.zeros(D), 10 * torch.eye(2)).sample([T]).to(device)) pi = pyro.param('pi', torch.ones(N, T, device=device) / T, constraint=constraints.simplex) with pyro.plate("beta_plate", T - 1): q_beta = pyro.sample("beta", Beta(torch.ones(T - 1, device=device), gamma)) with pyro.plate("prec_plate", T * D): q_prec = pyro.sample("prec", Gamma(zeta, delta)) q_corr_chol = torch.zeros(T, D, D, device=device) for t in pyro.plate("corr_chol_plate", T): q_corr_chol[t, ...] = pyro.sample("corr_chol_{}".format(t), LKJCorrCholesky(d=D, eta=psi[t])) with pyro.plate("mu_plate", T): _q_std = torch.sqrt(1. / q_prec.view(-1, D)) q_sigma_chol = torch.bmm(torch.diag_embed(_q_std), q_corr_chol) q_mu = pyro.sample("mu", MultivariateNormal(tau, scale_tril=q_sigma_chol)) with pyro.plate("data", N): z = pyro.sample("z", Categorical(pi)) T = 5 optim = Adam({"lr": 0.01}) svi = SVI(model, guide, optim, loss=Trace_ELBO(num_particles=35)) def train(num_iterations): losses = [] pyro.clear_param_store() fig = plt.figure(figsize=(5, 5)) for j in tqdm(range(num_iterations)): loss = svi.step(data) losses.append(loss) if (j % 100) == 0: centers, covars = marginal(guide, num_samples=250) animate(fig.gca(), centers, covars) plt.draw() plt.axis('equal') plt.pause(0.001) plt.clf() return losses def truncate(alpha, centers, perc, corrs, weights): threshold = alpha**-1 / 100. true_centers = centers[weights > threshold] prec = perc.view(T, D) true_prec = prec[weights > threshold] true_corrs = corrs[weights > threshold, ...] _stds = torch.sqrt(1. / true_prec.view(-1, D)) _sigmas = torch.bmm(torch.diag_embed(_stds), true_corrs) true_sigmas = torch.zeros(len(_sigmas), D, D) for n in range(len(_sigmas)): true_sigmas[n, ...] = torch.mm(_sigmas[n, ...], _sigmas[n, ...].T) true_weights = weights[weights > threshold] / torch.sum(weights[weights > threshold]) return true_centers, true_sigmas, true_weights def marginal(guide, num_samples=25): posterior_predictive = Predictive(guide, num_samples=num_samples) posterior_samples = posterior_predictive.forward(data) mu_mean = posterior_samples['mu'].detach().mean(dim=0) prec_mean = posterior_samples['prec'].detach().mean(dim=0) corr_mean = torch.zeros(T, D, D) for t in range(T): corr_mean[t, ...] = posterior_samples['corr_chol_{}'.format(t)].detach().mean(dim=0) beta_mean = posterior_samples['beta'].detach().mean(dim=0) weights_mean = mix_weights(beta_mean) centers, sigmas, _ = truncate(alpha, mu_mean, prec_mean, corr_mean, weights_mean) return centers, sigmas def animate(axes, centers, covars): plt.scatter(data[:, 0], data[:, 1], color="blue", marker="+") from math import pi t = torch.arange(0, 2 * pi, 0.01) circle = torch.stack([torch.sin(t), torch.cos(t)], dim=0) axes.scatter(centers[:, 0], centers[:, 1], color="red") for n in range(len(covars)): ellipse = torch.mm(torch.cholesky(covars[n, ...]), circle) axes.plot(ellipse[0, :] + centers[n, 0], ellipse[1, :] + centers[n, 1], linestyle='-', linewidth=2, color='g', alpha=1.) alpha = 0.1 * torch.ones(1, device=device) elbo = train(5000) plt.figure() plt.plot(elbo)
5,913
2,334
from calplus.tests import base from calplus.v1 import utils class TestUtils(base.TestCase): """docstring for TestUtils""" def setUp(self): super(TestUtils, self).setUp() def test_get_all_driver(self): drivers = utils.get_all_driver() self.assertEqual([], drivers) def test_validate_driver(self): @utils.validate_driver def test(request, drivers): pass import inspect args = inspect.getargspec(test).args self.assertEqual(['request'], args)
539
161
# General imports import os, json, logging import click from pathlib import Path import yaml # From common from luna_core.common.custom_logger import init_logger from luna_core.common.DataStore import DataStore_v2 from luna_core.common.Node import Node from luna_core.common.config import ConfigSet from luna_core.common.sparksession import SparkConfig @click.command() @click.option('-a', '--app_config', required=True, help="application configuration yaml file. See config.yaml.template for details.") @click.option('-s', '--datastore_id', required=True, help='datastore name. usually a slide id.') @click.option('-m', '--method_param_path', required=True, help='json parameter file with path to a WSI delta table.') def cli(app_config, datastore_id, method_param_path): """Load a slide to the datastore from the whole slide image table. app_config - application configuration yaml file. See config.yaml.template for details. datastore_id - datastore name. usually a slide id. method_param_path - json parameter file with path to a WSI delta table. - job_tag: job tag to use for loading the slide - table_path: path to the whole slide image table - datastore_path: path to store data """ init_logger() with open(method_param_path, 'r') as yaml_file: method_data = yaml.safe_load(yaml_file) load_slide_with_datastore(app_config, datastore_id, method_data) def load_slide_with_datastore(app_config, datastore_id, method_data): """Load a slide to the datastore from the whole slide image table. Args: app_config (string): path to application configuration file. datastore_id (string): datastore name. usually a slide id. method_data (dict): method parameters including input, output details. Returns: None """ logger = logging.getLogger(f"[datastore={datastore_id}]") # Do some setup cfg = ConfigSet("APP_CFG", config_file=app_config) datastore = DataStore_v2(method_data["datastore_path"]) method_id = method_data["job_tag"] # fetch patient_id column patient_id_column = method_data.get("patient_id_column_name", None) if patient_id_column == "": patient_id_column = None try: spark = SparkConfig().spark_session("APP_CFG", "query_slide") slide_id = datastore_id if patient_id_column: # assumes if patient_id column, source is parquet from dremio # right now has nested row-type into dict, todo: account for map type representation of dict in dremio df = spark.read.parquet(method_data['table_path'])\ .where(f"UPPER(slide_id)='{slide_id}'")\ .select("path", "metadata", patient_id_column)\ .toPandas() if not len(df) == 1: print(df) raise ValueError(f"Resulting query record is not singular, multiple scan's exist given the container address {slide_id}") record = df.loc[0] properties = record['metadata'] properties['patient_id'] = str(record[patient_id_column]) else: df = spark.read.format("delta").load(method_data['table_path'])\ .where(f"UPPER(slide_id)='{slide_id}'")\ .select("path", "metadata")\ .toPandas() if not len(df) == 1: print(df) raise ValueError(f"Resulting query record is not singular, multiple scan's exist given the container address {slide_id}") record = df.loc[0] properties = record['metadata'] spark.stop() except Exception as e: logger.exception (f"{e}, stopping job execution...") raise e # Put results in the data store data_path = Path(record['path'].split(':')[-1]) print(data_path) datastore.put(data_path, datastore_id, method_id, "WholeSlideImage", symlink=True) with open(os.path.join(method_data["datastore_path"], datastore_id, method_id, "WholeSlideImage", "metadata.json"), "w") as fp: json.dump(properties, fp) if __name__ == "__main__": cli()
4,213
1,270
import numpy as np import cv2 from matplotlib import pyplot as plt image = cv2.imread('champaigneditedcompressed.png') kernel = np.ones((20, 20), np.float32) / 25 img = cv2.filter2D(image, -1, kernel) gray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY) corners = cv2.goodFeaturesToTrack(gray,10,0.01,10) corners = np.int0(corners) print(corners) for i in corners: x,y = i.ravel() cv2.circle(img,(x,y),3,255,-1) plt.imshow(img),plt.show()
442
206
''' Created on 12 Jun, 2015 @author: artur.mrowca ''' from enum import Enum from PyQt5.Qt import QObject from PyQt5 import QtCore from tools.ecu_logging import ECULogger import copy class AbstractInputHandler(QObject): publish_infos_sig = QtCore.pyqtSignal(list) def __init__(self): QObject.__init__(self) self.next = None self._recs = [] def set_next(self, input_handler): self.next = input_handler def subscribe(self, obj, func_name): ''' all objects that subscribe to this function publish their information here ''' self._recs.append(obj) exec('self.publish_infos_sig.connect(obj.%s)' % func_name) def publish(self, cur_time, monitor_inputs): # emit the signal to all connected receivers then call next publish try: res = [[monitor_input.time_called, str(monitor_input.mon_id), str(monitor_input.asc_id), str(monitor_input.tag), monitor_input.msg_id, str(monitor_input.message), \ monitor_input.msg_size, monitor_input.stream_id, str(monitor_input.unique_id), str(monitor_input.data)] \ for monitor_input in monitor_inputs.get() if monitor_input.tag in self._get_tags()] # if there is a result only: if res: self.publish_infos_sig.emit(copy.deepcopy(res)) except: ECULogger().log_traceback() if self.next != None: self.next.publish(cur_time, monitor_inputs) def _get_tags(self): return [] class BufferHandler(AbstractInputHandler): def __init__(self): AbstractInputHandler.__init__(self) def _get_tags(self): return [MonitorTags.BT_ECU_RECEIVE_BUFFER, MonitorTags.BT_ECU_TRANSMIT_BUFFER] class CanBusHandler(AbstractInputHandler): def __init__(self): AbstractInputHandler.__init__(self) def publish(self, cur_time, monitor_inputs): # emit the signal to all connected receivers then call next publish try: res = [[monitor_input.time_called, str(monitor_input.mon_id), str(monitor_input.asc_id), str(monitor_input.tag), monitor_input.msg_id, str(monitor_input.message), \ monitor_input.msg_size, monitor_input.stream_id, str(monitor_input.unique_id), str(monitor_input.data)] \ for monitor_input in monitor_inputs.get() if monitor_input.tag in self._get_tags()] if res: self.publish_infos_sig.emit(copy.deepcopy([cur_time, res])) except: ECULogger().log_traceback() if self.next != None: self.next.publish(cur_time, monitor_inputs) def _get_tags(self): return [MonitorTags.CB_DONE_PROCESSING_MESSAGE, MonitorTags.CB_PROCESSING_MESSAGE] class ConstellationHandler(AbstractInputHandler): def __init__(self): AbstractInputHandler.__init__(self) self.first = True def publish(self, values, monitor_inputs): ''' pushes the initial constellation exactly once ''' try: if self.first: self.publish_infos_sig.emit(values) self.first = False except: pass if self.next != None: self.next.publish(values, monitor_inputs) def _get_tags(self): return [MonitorTags.CONSELLATION_INFORMATION] class EventlineHandler(AbstractInputHandler): def __init__(self): AbstractInputHandler.__init__(self) self._tags_list =[MonitorTags.CP_SEC_INIT_AUTHENTICATION, \ MonitorTags.CP_SEC_RECEIVE_REG_MESSAGE, \ MonitorTags.CP_SEC_DECRYPTED_INNER_REG_MESSAGE, \ MonitorTags.CP_SEC_DECRYPTED_OUTER_REG_MESSAGE, \ MonitorTags.CP_SEC_VALIDATED_ECU_CERTIFICATE, \ MonitorTags.CP_SEC_CREATED_CMP_HASH_REG_MSG, \ MonitorTags.CP_SEC_COMPARED_HASH_REG_MSG, \ MonitorTags.CP_SEC_RECEIVE_REQ_MESSAGE, \ MonitorTags.CP_SEC_DECRYPTED_REQ_MESSAGE, \ MonitorTags.CP_SEC_ECNRYPTED_CONFIRMATION_MESSAGE, \ MonitorTags.CP_SEC_GENERATED_SESSION_KEY, \ MonitorTags.CP_SEC_ENCRYPTED_DENY_MESSAGE, \ MonitorTags.CP_SEC_ENCRYPTED_GRANT_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_INTENT_SEND_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_SEND_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_SEC_MOD_ADVERTISEMENT, \ MonitorTags.CP_ECU_VALIDATED_SEC_MOD_CERTIFICATE, \ MonitorTags.CP_ECU_START_CREATION_REG_MESSAGE, \ MonitorTags.CP_ECU_CREATED_ECU_KEY_REG_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_INNER_REG_MESSAGE, \ MonitorTags.CP_ECU_HASHED_INNER_REG_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_OUTER_REG_MESSAGE, \ MonitorTags.CP_ECU_SEND_REG_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_CONF_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_CONF_MESSAGE, \ MonitorTags.CP_ECU_START_CREATE_REQ_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_REQ_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_DENY_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_DENY_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_GRANT_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_GRANT_MESSAGE, \ MonitorTags.CP_SESSION_AVAILABLE_SEND_MESSAGE, \ MonitorTags.CP_SEND_CLIENT_HELLO, \ MonitorTags.CP_RECEIVE_CLIENT_HELLO, \ MonitorTags.CP_SEND_ALERT_NO_CIPHERSUITE, \ MonitorTags.CP_SEND_SERVER_HELLO, \ MonitorTags.CP_SEND_SERVER_CERTIFICATE, \ MonitorTags.CP_SEND_SERVER_KEYEXCHANGE, MonitorTags.CP_SEND_CERTIFICATE_REQUEST , \ MonitorTags.CP_SEND_SERVER_HELLO_DONE , \ MonitorTags.CP_RECEIVE_SERVER_HELLO , \ MonitorTags.CP_RECEIVE_SERVER_CERTIFICATE , \ MonitorTags.CP_RECEIVE_SERVER_KEYEXCHANGE , \ MonitorTags.CP_RECEIVE_CERTIFICATE_REQUEST , \ MonitorTags.CP_RECEIVE_SERVER_HELLO_DONE , \ MonitorTags.CP_SERVER_HELLO_DONE_VALIDATED_CERT , \ MonitorTags.CP_SEND_CLIENT_CERTIFICATE , \ MonitorTags.CP_INIT_SEND_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_ENCRYPTED_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_GENERATED_MASTERSEC_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_INIT_SEND_CERTIFICATE_VERIFY , \ MonitorTags.CP_ENCRYPTED_CERTIFICATE_VERIFY , \ MonitorTags.CP_SEND_CIPHER_SPEC , \ MonitorTags.CP_INIT_CLIENT_FINISHED , \ MonitorTags.CP_HASHED_CLIENT_FINISHED , \ MonitorTags.CP_GENERATED_HASH_FROM_PRF_CLIENT_FINISHED , \ MonitorTags.CP_RECEIVE_CLIENT_CERTIFICATE , \ MonitorTags.CP_CLIENT_CERTIFICATE_VALIDATED , \ MonitorTags.CP_RECEIVE_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_DECRYPTED_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_RECEIVE_CERTIFICATE_VERIFY , \ MonitorTags.CP_DECRYPTED_CERTIFICATE_VERIFY , \ MonitorTags.CP_GENERATED_MASTER_SECRET_CERT_VERIFY , \ MonitorTags.CP_RECEIVED_CHANGE_CIPHER_SPEC , \ MonitorTags.CP_RECEIVE_CLIENT_FINISHED , \ MonitorTags.CP_CLIENT_FINISHED_HASHED_COMPARISON_HASH , \ MonitorTags.CP_CLIENT_FINISHED_GENERATED_HASH_PRF , \ MonitorTags.CP_RECEIVE_SERVER_FINISHED , \ MonitorTags.CP_SERVER_FINISHED_HASHED_COMPARISON_HASH , \ MonitorTags.CP_SERVER_FINISHED_GENERATED_HASH_PRF , \ MonitorTags.CP_INIT_SERVER_FINISHED , \ MonitorTags.CP_HASHED_SERVER_FINISHED , \ MonitorTags.CP_GENERATED_HASH_FROM_PRF_SERVER_FINISHED , \ MonitorTags.CP_SERVER_AUTHENTICATED , \ MonitorTags.CP_CLIENT_AUTHENTICATED, \ MonitorTags.CP_RECEIVE_SIMPLE_MESSAGE, \ MonitorTags.CP_INIT_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_ENCRYPTED_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_SETUP_INIT_CREATE_KEYS, \ MonitorTags.CP_SETUP_FINISHED_CREATE_KEYS, \ MonitorTags.CP_INIT_TRANSMIT_MESSAGE, \ MonitorTags.CP_MACED_TRANSMIT_MESSAGE, \ MonitorTags.CP_RECEIVED_SIMPLE_MESSAGE, \ MonitorTags.CP_BUFFERED_SIMPLE_MESSAGE, \ MonitorTags.CP_RETURNED_AUTHENTICATED_SIMPLE_MESSAGE, \ MonitorTags.CP_RECEIVED_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_DECRYPTED_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_CHECKED_KEY_LEGID, \ MonitorTags.CP_INIT_CHECK_KEY_LEGID, \ MonitorTags.CP_INIT_VERIFYING_BUFFER_MESSAGE, \ MonitorTags.CP_FINISHED_VERIFYING_BUFFER_MESSAGE, \ MonitorTags.CP_SEND_SYNC_MESSAGE, \ MonitorTags.CP_SEND_SYNC_RESPONSE_MESSAGE, \ MonitorTags.CP_RECEIVE_SYNC_RESPONSE_MESSAGE] def publish(self, values, monitor_inputs): ''' pushes the ecu ids or the view ''' try: if values.tag == MonitorTags.ECU_ID_LIST: self.publish_infos_sig.emit([ecu.ecu_id for ecu in values.data]) else: AbstractInputHandler.publish(self, values, monitor_inputs) except: try: AbstractInputHandler.publish(self, values, monitor_inputs) except: pass if self.next != None: self.next.publish(values, monitor_inputs) def _get_tags(self): return self._tags_list def register_tag(self, tag): self._tags_list += [tag] def register_tags(self, tags_list): self._tags_list += tags_list self._tags_list = list(set(self._tags_list)) class CheckpointHandler(AbstractInputHandler): ''' reads and publishes all Checkpoint Monitor values''' def __init__(self): AbstractInputHandler.__init__(self) self._tags_list = [MonitorTags.CP_SEC_INIT_AUTHENTICATION, \ MonitorTags.CP_SEC_RECEIVE_REG_MESSAGE, \ MonitorTags.CP_SEC_DECRYPTED_INNER_REG_MESSAGE, \ MonitorTags.CP_SEC_DECRYPTED_OUTER_REG_MESSAGE, \ MonitorTags.CP_SEC_VALIDATED_ECU_CERTIFICATE, \ MonitorTags.CP_SEC_CREATED_CMP_HASH_REG_MSG, \ MonitorTags.CP_SEC_COMPARED_HASH_REG_MSG, \ MonitorTags.CP_SEC_RECEIVE_REQ_MESSAGE, \ MonitorTags.CP_SEC_DECRYPTED_REQ_MESSAGE, \ MonitorTags.CP_SEC_ECNRYPTED_CONFIRMATION_MESSAGE, \ MonitorTags.CP_SEC_GENERATED_SESSION_KEY, \ MonitorTags.CP_SEC_ENCRYPTED_DENY_MESSAGE, \ MonitorTags.CP_SEC_ENCRYPTED_GRANT_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_INTENT_SEND_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_SEND_SIMPLE_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_SEC_MOD_ADVERTISEMENT, \ MonitorTags.CP_ECU_VALIDATED_SEC_MOD_CERTIFICATE, \ MonitorTags.CP_ECU_START_CREATION_REG_MESSAGE, \ MonitorTags.CP_ECU_CREATED_ECU_KEY_REG_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_INNER_REG_MESSAGE, \ MonitorTags.CP_ECU_HASHED_INNER_REG_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_OUTER_REG_MESSAGE, \ MonitorTags.CP_ECU_SEND_REG_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_CONF_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_CONF_MESSAGE, \ MonitorTags.CP_ECU_START_CREATE_REQ_MESSAGE, \ MonitorTags.CP_ECU_ENCRYPTED_REQ_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_DENY_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_DENY_MESSAGE, \ MonitorTags.CP_ECU_RECEIVE_GRANT_MESSAGE, \ MonitorTags.CP_ECU_DECRYPTED_GRANT_MESSAGE, \ MonitorTags.CP_SESSION_AVAILABLE_SEND_MESSAGE, \ MonitorTags.CP_SEND_CLIENT_HELLO, \ MonitorTags.CP_RECEIVE_CLIENT_HELLO, \ MonitorTags.CP_SEND_ALERT_NO_CIPHERSUITE, \ MonitorTags.CP_SEND_SERVER_HELLO, \ MonitorTags.CP_SEND_SERVER_CERTIFICATE, \ MonitorTags.CP_SEND_SERVER_KEYEXCHANGE, MonitorTags.CP_SEND_CERTIFICATE_REQUEST , \ MonitorTags.CP_SEND_SERVER_HELLO_DONE , \ MonitorTags.CP_RECEIVE_SERVER_HELLO , \ MonitorTags.CP_RECEIVE_SERVER_CERTIFICATE , \ MonitorTags.CP_RECEIVE_SERVER_KEYEXCHANGE , \ MonitorTags.CP_RECEIVE_CERTIFICATE_REQUEST , \ MonitorTags.CP_RECEIVE_SERVER_HELLO_DONE , \ MonitorTags.CP_SERVER_HELLO_DONE_VALIDATED_CERT , \ MonitorTags.CP_SEND_CLIENT_CERTIFICATE , \ MonitorTags.CP_INIT_SEND_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_ENCRYPTED_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_GENERATED_MASTERSEC_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_INIT_SEND_CERTIFICATE_VERIFY , \ MonitorTags.CP_ENCRYPTED_CERTIFICATE_VERIFY , \ MonitorTags.CP_SEND_CIPHER_SPEC , \ MonitorTags.CP_INIT_CLIENT_FINISHED , \ MonitorTags.CP_HASHED_CLIENT_FINISHED , \ MonitorTags.CP_GENERATED_HASH_FROM_PRF_CLIENT_FINISHED , \ MonitorTags.CP_RECEIVE_CLIENT_CERTIFICATE , \ MonitorTags.CP_CLIENT_CERTIFICATE_VALIDATED , \ MonitorTags.CP_RECEIVE_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_DECRYPTED_CLIENT_KEYEXCHANGE , \ MonitorTags.CP_RECEIVE_CERTIFICATE_VERIFY , \ MonitorTags.CP_DECRYPTED_CERTIFICATE_VERIFY , \ MonitorTags.CP_GENERATED_MASTER_SECRET_CERT_VERIFY , \ MonitorTags.CP_RECEIVED_CHANGE_CIPHER_SPEC , \ MonitorTags.CP_RECEIVE_CLIENT_FINISHED , \ MonitorTags.CP_CLIENT_FINISHED_HASHED_COMPARISON_HASH , \ MonitorTags.CP_CLIENT_FINISHED_GENERATED_HASH_PRF , \ MonitorTags.CP_RECEIVE_SERVER_FINISHED , \ MonitorTags.CP_SERVER_FINISHED_HASHED_COMPARISON_HASH , \ MonitorTags.CP_SERVER_FINISHED_GENERATED_HASH_PRF , \ MonitorTags.CP_INIT_SERVER_FINISHED , \ MonitorTags.CP_HASHED_SERVER_FINISHED , \ MonitorTags.CP_GENERATED_HASH_FROM_PRF_SERVER_FINISHED , \ MonitorTags.CP_SERVER_AUTHENTICATED , \ MonitorTags.CP_CLIENT_AUTHENTICATED, \ MonitorTags.CP_RECEIVE_SIMPLE_MESSAGE, \ MonitorTags.CP_INIT_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_ENCRYPTED_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_SETUP_INIT_CREATE_KEYS, \ MonitorTags.CP_SETUP_FINISHED_CREATE_KEYS, \ MonitorTags.CP_INIT_TRANSMIT_MESSAGE, \ MonitorTags.CP_MACED_TRANSMIT_MESSAGE, \ MonitorTags.CP_RECEIVED_SIMPLE_MESSAGE, \ MonitorTags.CP_BUFFERED_SIMPLE_MESSAGE, \ MonitorTags.CP_RETURNED_AUTHENTICATED_SIMPLE_MESSAGE, \ MonitorTags.CP_RECEIVED_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_DECRYPTED_EXCHANGE_FIRST_KEY_KN, \ MonitorTags.CP_INIT_CHECK_KEY_LEGID, \ MonitorTags.CP_CHECKED_KEY_LEGID, \ MonitorTags.CP_INIT_VERIFYING_BUFFER_MESSAGE, \ MonitorTags.CP_FINISHED_VERIFYING_BUFFER_MESSAGE, \ MonitorTags.CP_SEND_SYNC_MESSAGE, \ MonitorTags.CP_SEND_SYNC_RESPONSE_MESSAGE, \ MonitorTags.CP_RECEIVE_SYNC_RESPONSE_MESSAGE] # override def publish(self, cur_time, monitor_inputs): # emit the signal to all connected receivers then call next publish try: res = [[monitor_input.time_called, str(monitor_input.mon_id), str(monitor_input.asc_id), str(monitor_input.tag), monitor_input.msg_id, str(monitor_input.message), \ monitor_input.msg_size, monitor_input.stream_id, str(monitor_input.unique_id), str(monitor_input.data)] \ for monitor_input in monitor_inputs.get() if monitor_input.tag in self._get_tags()] self.publish_infos_sig.emit([None, None]) except: ECULogger().log_traceback() if self.next != None: self.next.publish(cur_time, monitor_inputs) def _get_tags(self): return self._tags_list def register_tag(self, tag): self._tags_list += [tag] def register_tags(self, tags_list): self._tags_list += tags_list class InputHandlerChain(object): def add_handler(self, handler): try: self._next_handler.set_next(handler) except: self._handler = handler self._next_handler = handler return handler def handler(self): return self._handler class MonitorInput(object): ''' Keeps the input data of a monitor. Tag defines the type of data arriving Possible Tags are MonitorTags ''' def __init__(self, data, monitor_tag, mon_id=False, time_called=False, asc_id=None, \ msg_id=-1, message=None, msg_size=-1, stream_id=-1, unique_id=None): self.data = data self.tag = monitor_tag self.mon_id = mon_id self.time_called = time_called self.calling_object = None self.asc_id = asc_id self.msg_id = msg_id self.message = message self.msg_size = msg_size self.stream_id = stream_id self.unique_id = unique_id class MonitorTags(Enum): # Buffer Tags BT_ECU_TRANSMIT_BUFFER = 1 BT_ECU_RECEIVE_BUFFER = 2 # Receiving/Sending Times, Message Contents SPAETER MIT CHECKPOINTS ZUSAMMENFUEHREN # Checkpoints - SEC MOD SIDE CP_SEC_INIT_AUTHENTICATION = 7 # Sec Mod. initialized authentication CP_SEC_RECEIVE_REG_MESSAGE = 8 # Sec Mod receive the registration message CP_SEC_DECRYPTED_INNER_REG_MESSAGE = 10 # Sec Mod decrypted inner reg Msg CP_SEC_DECRYPTED_OUTER_REG_MESSAGE = 11 # Sec Mod decrypted outer reg Msg CP_SEC_VALIDATED_ECU_CERTIFICATE = 12 # Sec Mod. validated the ECU Certificate CP_SEC_CREATED_CMP_HASH_REG_MSG = 13 # Create the comparision hash for validation of reg msg hash CP_SEC_COMPARED_HASH_REG_MSG = 14 # Finished comparing the reg msg hash to the created hash and sends the message CP_SEC_RECEIVE_REQ_MESSAGE = 15 # Sec Mod. received the request message CP_SEC_DECRYPTED_REQ_MESSAGE = 16 # Sec Mod decrypted the request message CP_SEC_ECNRYPTED_CONFIRMATION_MESSAGE = 9 # Sec Mod created confirmation message and sends it CP_SEC_GENERATED_SESSION_KEY = 17 # Sec Mod. generated the session key CP_SEC_ENCRYPTED_DENY_MESSAGE = 18 # Sec. Mod encrypted the deny message CP_SEC_ENCRYPTED_GRANT_MESSAGE = 19 # Sec. Mod encrypted the grant message # Checkpoints - ECU SIDE CP_ECU_RECEIVE_SIMPLE_MESSAGE = 20 # ECU receives a encrypted simple message CP_ECU_DECRYPTED_SIMPLE_MESSAGE = 21 # ECU decrypted the received simple message CP_ECU_INTENT_SEND_SIMPLE_MESSAGE = 22 # ECU decides on comm. module that it wants to send a simple message CP_ECU_ENCRYPTED_SEND_SIMPLE_MESSAGE = 23 # ECU encrypted message and sends it CP_ECU_RECEIVE_SEC_MOD_ADVERTISEMENT = 24 # ECU receives the SEC Module advertisement CP_ECU_VALIDATED_SEC_MOD_CERTIFICATE = 25 # ECU validated the sec. mod. certificate CP_ECU_START_CREATION_REG_MESSAGE = 26 # ECU starts to create the registration message CP_ECU_CREATED_ECU_KEY_REG_MESSAGE = 27 # ECU created the sym. ECU key CP_ECU_ENCRYPTED_INNER_REG_MESSAGE = 28 # ECU encrypted inner reg. msg CP_ECU_HASHED_INNER_REG_MESSAGE = 29 # ECU hashed inner reg. msg CP_ECU_ENCRYPTED_OUTER_REG_MESSAGE = 30 # ECU encrypted the outer reg. msg CP_ECU_SEND_REG_MESSAGE = 31 # ECU sends the reg. message CP_ECU_RECEIVE_CONF_MESSAGE = 32 # ECU receives a confirmation message CP_ECU_DECRYPTED_CONF_MESSAGE = 33 # ECU decrypted the confirmation message CP_ECU_START_CREATE_REQ_MESSAGE = 34 # ECU Starts creation of request message CP_ECU_ENCRYPTED_REQ_MESSAGE = 35 # ECU encrypted the request message and sends it CP_ECU_RECEIVE_DENY_MESSAGE = 36 # ECU receives a deny message CP_ECU_DECRYPTED_DENY_MESSAGE = 37 # ECU decrypted the deny message CP_ECU_RECEIVE_GRANT_MESSAGE = 38 # ECU receives a grant message CP_ECU_DECRYPTED_GRANT_MESSAGE = 39 # ECU decrypted the grant message CP_ECU_ALREADY_AUTHENTICATED = 40 # The ECU is already authenticated and the message is discareded # Checkpoints - TLS CP_SESSION_AVAILABLE_SEND_MESSAGE = 43 # There is a session available for this stream and the message is transmitted CP_SEND_CLIENT_HELLO = 44 # No session is available for that stream. Send the client hello message CP_RECEIVE_CLIENT_HELLO = 45 # Receive the client hello and answer CP_SEND_ALERT_NO_CIPHERSUITE = 46 # alert message if the wrong ciphersuite was chosen CP_SEND_SERVER_HELLO = 47 # send the server Hello message CP_SEND_SERVER_CERTIFICATE = 48 # send the server Certificate message CP_SEND_SERVER_KEYEXCHANGE = 49 # send the server Keyexchange message CP_SEND_CERTIFICATE_REQUEST = 50 # send the certificate request message CP_SEND_SERVER_HELLO_DONE = 51 # send the server Hello done message CP_RECEIVE_SERVER_HELLO = 52 CP_RECEIVE_SERVER_CERTIFICATE = 53 CP_RECEIVE_SERVER_KEYEXCHANGE = 54 CP_RECEIVE_CERTIFICATE_REQUEST = 55 CP_RECEIVE_SERVER_HELLO_DONE = 56 CP_SERVER_HELLO_DONE_VALIDATED_CERT = 57 CP_SEND_CLIENT_CERTIFICATE = 58 CP_INIT_SEND_CLIENT_KEYEXCHANGE = 59 CP_ENCRYPTED_CLIENT_KEYEXCHANGE = 60 CP_GENERATED_MASTERSEC_CLIENT_KEYEXCHANGE = 61 CP_INIT_SEND_CERTIFICATE_VERIFY = 62 CP_ENCRYPTED_CERTIFICATE_VERIFY = 63 CP_SEND_CIPHER_SPEC = 64 # Send the cipher spec message CP_INIT_CLIENT_FINISHED = 65 # start to create client finished message CP_HASHED_CLIENT_FINISHED = 66 # finished to hash the client finished message data CP_GENERATED_HASH_FROM_PRF_CLIENT_FINISHED = 67 # Hash generated and sending message started CP_RECEIVE_CLIENT_CERTIFICATE = 68 # receive the client certificate CP_CLIENT_CERTIFICATE_VALIDATED = 69 # Finished validation of client certificate CP_RECEIVE_CLIENT_KEYEXCHANGE = 70 CP_DECRYPTED_CLIENT_KEYEXCHANGE = 71 CP_RECEIVE_CERTIFICATE_VERIFY = 72 CP_DECRYPTED_CERTIFICATE_VERIFY = 73 CP_GENERATED_MASTER_SECRET_CERT_VERIFY = 74 CP_RECEIVED_CHANGE_CIPHER_SPEC = 75 # CP_RECEIVED_CLIENT_FINISHED = 76 CP_RECEIVE_CLIENT_FINISHED = 83 CP_CLIENT_FINISHED_HASHED_COMPARISON_HASH = 84 CP_CLIENT_FINISHED_GENERATED_HASH_PRF = 85 CP_RECEIVE_SERVER_FINISHED = 80 CP_SERVER_FINISHED_HASHED_COMPARISON_HASH = 81 CP_SERVER_FINISHED_GENERATED_HASH_PRF = 82 CP_INIT_SERVER_FINISHED = 77 # start to create SERVER finished message CP_HASHED_SERVER_FINISHED = 78 # finished to hash the SERVER finished message data CP_GENERATED_HASH_FROM_PRF_SERVER_FINISHED = 79 # Hash generated and sending message started CP_SERVER_AUTHENTICATED = 86 CP_CLIENT_AUTHENTICATED = 87 CP_RECEIVE_SIMPLE_MESSAGE = 88 # Checkpoints - TESLA CP_INIT_EXCHANGE_FIRST_KEY_KN = 89 # Intention to send the Key K N to receiver xy CP_ENCRYPTED_EXCHANGE_FIRST_KEY_KN = 90 # Encryption finished for Key K_N to receiver xy CP_SETUP_INIT_CREATE_KEYS = 91 # Start the creation of keys CP_SETUP_FINISHED_CREATE_KEYS = 92 # Finished creating keys CP_INIT_TRANSMIT_MESSAGE = 93 # Intention to send a simple message CP_MACED_TRANSMIT_MESSAGE = 94 # Finished MAC Creation of message now send it CP_RECEIVED_SIMPLE_MESSAGE = 95 # Received a simple message CP_BUFFERED_SIMPLE_MESSAGE = 96 # Added simple message to buffer CP_RETURNED_AUTHENTICATED_SIMPLE_MESSAGE = 97 # Authenticated messages are returned CP_RECEIVED_EXCHANGE_FIRST_KEY_KN = 98 # received first key message CP_DECRYPTED_EXCHANGE_FIRST_KEY_KN = 99 # decrypted first key message CP_INIT_CHECK_KEY_LEGID = 99.5 # start to check if key legid CP_CHECKED_KEY_LEGID = 100 # checked the key legidity CP_INIT_VERIFYING_BUFFER_MESSAGE = 101 # Start validation of message in buffer CP_FINISHED_VERIFYING_BUFFER_MESSAGE = 102 # Done validation of message in buffer CP_SEND_SYNC_MESSAGE = 103 # send the time sync message from the ECU CP_SEND_SYNC_RESPONSE_MESSAGE = 104 CP_RECEIVE_SYNC_RESPONSE_MESSAGE = 105 # End message CP_RECEIVE_SYNC_MESSAGE = 106 # sync message was received # CAN BUS TAGS CB_DONE_PROCESSING_MESSAGE = 41 CB_PROCESSING_MESSAGE = 42 # Constellation Handler CONSELLATION_INFORMATION = 107 ECU_ID_LIST = 108
25,581
9,666
################################################################################ # Copyright: Tobias Weber 2019 # # Apache 2.0 License # # This file contains all code related to pid check objects # ################################################################################ import re import requests from breadp.checks import Check from breadp.checks.result import BooleanResult class IsValidDoiCheck(Check): """ Checks whether an RDP has a valid DOI as PID Methods ------- _do_check(self, rdp) returns a BooleanResult """ def __init__(self): Check.__init__(self) self.id = 0 self.version = "0.0.1" def _do_check(self, rdp): if not rdp.pid: msg = "RDP has no PID" return BooleanResult(False, msg, False) if re.match(r"^10\.\d{4}\d*/.*", rdp.pid): return BooleanResult(True, "", True) msg = "{} is not a valid DOI".format(rdp.pid) return BooleanResult(False, msg, True) class DoiResolvesCheck(Check): """ Checks whether the DOI of an RDP resolves Methods ------- _do_check(self, rdp) returns a BooleanResult """ def __init__(self): Check.__init__(self) self.id = 1 self.version = "0.0.1" def _do_check(self, rdp): if not rdp.pid: msg = "RDP has no PID" return BooleanResult(False, msg, False) try: response = requests.head('https://doi.org/' + rdp.pid) except Exception as e: msg = "{}: {}".format(type(e).__name__, e) return BooleanResult(False, msg, False) if response.status_code != 302: msg = "Could not resolve {}, status code: {}".format( rdp.pid, response.status_code) return BooleanResult(False, msg, True) msg = "Location of resolved doi: {}".format( response.headers.get('Location')) return BooleanResult(True, msg, True)
1,994
594
from unittest import TestCase from src.domain.holiday import Holiday import src.infrastructure.persistence.holiday_dynamo_repository as repository HOLIDAY = Holiday( date='2021-12-25', name='Natal', category='NATIONAL' ) class TestHolidayDynamoRepository(TestCase): def test_holiday_must_be_dict_when_given_as_object(self): response = repository.translate_holiday_to_dynamo(HOLIDAY) self.assertIsInstance(response, dict, 'Holiday was not returned as a dict')
495
170
from config.defs import * import peewee db = peewee.SqliteDatabase(DATABASE_NAME) class BaseModel(peewee.Model): class Meta: database = db class Partida(BaseModel): id_time_casa = peewee.CharField() id_time_visitante = peewee.CharField() time_casa = peewee.CharField() time_visitante = peewee.CharField() data = peewee.DateField() time_da_casa_venceu = peewee.IntegerField() HG = peewee.FloatField() AG = peewee.FloatField() PH = peewee.FloatField() PD = peewee.FloatField() PA = peewee.FloatField() MAX_H = peewee.FloatField() MAX_D = peewee.FloatField() MAX_A = peewee.FloatField() AVG_H = peewee.FloatField() AVG_D = peewee.FloatField() AVG_A = peewee.FloatField() class Meta: db_table = 'partidas' class Substituicao(BaseModel): # Pode ser INTERVALO tempo = peewee.CharField() tipo_tatico = peewee.CharField(null=True) efetividade = peewee.IntegerField() class Meta: db_table = 'substituicoes' class Penalti(BaseModel): tempo = peewee.CharField() class Meta: db_table = 'penaltis' class CartaoAmarelo(BaseModel): tempo = peewee.CharField() id_jogador = peewee.CharField() class Meta: db_table = 'cartoes_amarelos' class CartaoVermelho(BaseModel): tempo = peewee.CharField() id_jogador = peewee.CharField() class Meta: db_table = 'cartoes_vermelhos' class GolContra(BaseModel): tempo = peewee.CharField() id_jogador = peewee.CharField() class Meta: db_table = 'gols_contra' class Gol(BaseModel): tempo = peewee.CharField() id_jogador = peewee.CharField() class Meta: db_table = 'gols' class Time(BaseModel): api_id = peewee.IntegerField() nome = peewee.CharField() class Meta: db_table = "times" ### Relacionamentos ### class PartidasSubstituicoes(BaseModel): partida = peewee.ForeignKeyField(Partida) substituicao = peewee.ForeignKeyField(Substituicao) class Meta: db_table = 'partidas_substituicoes' class PartidasPenaltis(BaseModel): partida = peewee.ForeignKeyField(Partida) penalti = peewee.ForeignKeyField(Penalti) class Meta: db_table = 'partidas_penaltis' class PartidasCartoesAmarelos(BaseModel): partida = peewee.ForeignKeyField(Partida) cartoes_amarelos = peewee.ForeignKeyField(CartaoAmarelo) class Meta: db_table = 'partidas_cartoes_amarelos' class PartidasCartoesVermelhos(BaseModel): partida = peewee.ForeignKeyField(Partida) cartoes_vermelhos = peewee.ForeignKeyField(CartaoVermelho) class Meta: db_table = 'partidas_cartoes_vermelhos' class PartidasGolsContra(BaseModel): partida = peewee.ForeignKeyField(Partida) gols_contra = peewee.ForeignKeyField(GolContra) class Meta: db_table = 'partidas_gols_contra' class PartidasGols(BaseModel): partida = peewee.ForeignKeyField(Partida) gols = peewee.ForeignKeyField(Gol) class Meta: db_table = 'partidas_gols' db.create_tables([Partida, Substituicao, Penalti, CartaoAmarelo, CartaoVermelho, GolContra, Gol, Time]) db.create_tables([PartidasSubstituicoes, PartidasPenaltis, PartidasCartoesAmarelos, PartidasCartoesVermelhos, PartidasGolsContra, PartidasGols])
3,136
1,267
class AnimationException(SystemException,ISerializable,_Exception): """ The exception that is thrown when an error occurs while animating a property. """ def add_SerializeObjectState(self,*args): """ add_SerializeObjectState(self: Exception,value: EventHandler[SafeSerializationEventArgs]) """ pass def remove_SerializeObjectState(self,*args): """ remove_SerializeObjectState(self: Exception,value: EventHandler[SafeSerializationEventArgs]) """ pass def __init__(self,*args): """ x.__init__(...) initializes x; see x.__class__.__doc__ for signaturex.__init__(...) initializes x; see x.__class__.__doc__ for signaturex.__init__(...) initializes x; see x.__class__.__doc__ for signature """ pass def __reduce_ex__(self,*args): pass def __str__(self,*args): pass Clock=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets the clock that generates the animated values. Get: Clock(self: AnimationException) -> AnimationClock """ Property=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets the animated dependency property. Get: Property(self: AnimationException) -> DependencyProperty """ Target=property(lambda self: object(),lambda self,v: None,lambda self: None) """Gets the animated object. Get: Target(self: AnimationException) -> IAnimatable """
1,386
426
r""" ======= quantnn ======= The quantnn package provides functionality for probabilistic modeling and prediction using deep neural networks. The two main features of the quantnn package are implemented by the :py:class:`~quantnn.qrnn.QRNN` and :py:class:`~quantnn.qrnn.DRNN` classes, which implement quantile regression neural networks (QRNNs) and density regression neural networks (DRNNs), respectively. The modules :py:mod:`quantnn.quantiles` and :py:mod:`quantnn.density` provide generic (backend agnostic) functions to manipulate probabilistic predictions. """ import logging as _logging import os from rich.logging import RichHandler from quantnn.neural_network_model import set_default_backend, get_default_backend from quantnn.qrnn import QRNN from quantnn.drnn import DRNN from quantnn.quantiles import ( cdf, pdf, posterior_mean, probability_less_than, probability_larger_than, sample_posterior, sample_posterior_gaussian, quantile_loss, ) _LOG_LEVEL = os.environ.get("QUANTNN_LOG_LEVEL", "WARNING").upper() _logging.basicConfig( level=_LOG_LEVEL, format="%(message)s", datefmt="[%X]", handlers=[RichHandler()] )
1,165
367
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import attr import magnum as mn import numpy as np @attr.s(auto_attribs=True, slots=True) class RobotTarget: """ Data class to define the target needed as input for the motion planner. """ # End-effector in world coordinate frame. ee_target_pos: np.ndarray = None obj_id_target: int = None joints_target: np.ndarray = None is_guess: bool = False @attr.s(auto_attribs=True, slots=True) class ObjectGraspTarget: """ Data class to define the target needed as input for the grasp planner. """ # Bounding Box bb: mn.Range3D translation: mn.Matrix4
805
262
#!/usr/bin/env python #-*- coding: utf-8 -*- class ListNode(object): def __init__(self, x): self.val = x self.next = None class TwoNumbers(object): @staticmethod def builderListNode(nums): if nums is not None: head = ListNode(str(nums)[0]) a = head for i in str(nums)[1:]: b = ListNode(i) a.next = b a = a.next return head @staticmethod def addTwoNumbers(n1, n2): s1 = "" s2 = "" while n1 is not None: s1 += str(n1.val) n1 = n1.next while n2 is not None: s2 += str(n2.val) n2 = n2.next summation = str(int(s1) + int(s2)) head = ListNode(summation[0]) temp = head for val in summation[1:]: temp.next = ListNode(val) temp = temp.next return head @staticmethod def printLS(node): if not node: return None res = '' while node: res += str(node.val) + ' -> ' node = node.next print res if __name__ == "__main__": tn = TwoNumbers() l1 = tn.builderListNode(1234) l2 = tn.builderListNode(34) tn.printLS(tn.addTwoNumbers(l1, l2))
1,314
442
import logging import tensorflow as tf import recsys.util.proto.config_pb2 as config def int64_feature(val): return tf.train.Feature(int64_list = tf.train.Int64List(value=[val])) def int64_list_feature(val_list): return tf.train.Feature(int64_list = tf.train.Int64List(value=val_list)) def bytes_feature(val): return tf.train.Feature(bytes_list = tf.train.BytesList(value=[val])) def bytes_list_feature(val_list): return tf.train.Feature(bytes_list = tf.train.BytesList(value=val_list)) def float_feature(val): return tf.train.Feature(float_list = tf.train.FloatList(value=[val])) def float_list_feature(val_list): return tf.train.Feature(float_list = tf.train.FloatList(value=val_list)) def string_feature(str): return tf.train.Feature(bytes_list = tf.train.BytesList(value=[str.encode('utf-8')])) def string_list_feature(str_list): str_bytes_list = [k.encode('utf-8') for k in str_list] return bytes_list_feature(str_bytes_list) # Constructs a tf.Example with feature dictionary where key is feature name and # value is tf.train.Feature def example_from_feature_dict(feature_dict): return tf.train.Example(features = tf.train.Features(feature = feature_dict)) def get_int64_feature(example, feature_name): return example.features.feature[feature_name].int64_list.value[0] def get_int64_list_feature(example, feature_name): return list(example.features.feature[feature_name].int64_list.value) def get_float_feature(example, feature_name): return example.features.feature[feature_name].float_list.value[0] def get_float_list_feature(example, feature_name): return list(example.features.feature[feature_name].float_list.value) def get_bytes_feature(example, feature_name): return example.features.feature[feature_name].bytes_list.value[0] def get_bytes_list_feature(example, feature_name): return example.features.feature[feature_name].bytes_list.value def get_string_feature(example, feature_name): return example.features.feature[feature_name].bytes_list.value[0].decode('utf-8') def get_string_list_feature(example, feature_name): return [s.decode('utf-8') for s in example.features.feature[feature_name].bytes_list.value] # Reads batched features and labels from given files, and consumes them through # callback function "consum_batch_fn". # feature_spec: dictionary specifying the type of each feature. # input_config: settings for generating batched features and labels. # consume_batch_fn: callback function that defines the consumption of the # batched features and labels. def fetch_and_process_features(filenames, feature_spec, input_config, consume_batch_fn): # Reads examples from the filenames and parse them into features. def _read_and_decode(filename_queue, feature_spec, batch_size = 2, capacity = 30, num_threads = 2, min_after_dequeue = 10): reader = tf.TFRecordReader() _, serialized_example = reader.read(filename_queue) features = tf.parse_single_example( serialized_example, features = feature_spec) batched_features = tf.train.shuffle_batch(features, batch_size = batch_size, capacity = capacity, num_threads = num_threads, min_after_dequeue = min_after_dequeue) return batched_features filename_queue = tf.train.string_input_producer( filenames, num_epochs = input_config.num_epochs) features = _read_and_decode( filename_queue, feature_spec, batch_size = input_config.batch_size, capacity = input_config.capacity, num_threads = input_config.num_threads, min_after_dequeue = input_config.min_after_dequeue ) init_op = tf.group(tf.global_variables_initializer(), tf.local_variables_initializer()) with tf.Session() as sess: sess.run(init_op) coord = tf.train.Coordinator() threads = tf.train.start_queue_runners(coord=coord) for i in range(input_config.num_batches): logging.info('current batch:{}'.format(i)) consume_batch_fn(sess, features) coord.request_stop() coord.join(threads)
3,961
1,328
''' This module handles starring of modules. ''' import web from app import RENDER from components import model, session class StarModule(object): ''' Class handles starring and unstarring of modules. ''' def GET(self): ''' This function is called when /starModule is accessed. ''' web.header('X-Frame-Options', 'SAMEORIGIN') web.header('X-Content-Type-Options', 'nosniff') web.header('X-XSS-Protection', '1') if not session.validate_session(): raise web.seeother('/login') else: module_code = web.input().code action = web.input().action return_path = web.input().return_path # modify return path if individual module info to include aySem if return_path == '/individualModuleInfo': target_ay = web.input().aysem return_path = return_path + '?code=' + module_code + '&aysem=' + target_ay if action == "star": model.star_module(module_code, web.cookies().get('user')) else: model.unstar_module(module_code, web.cookies().get('user')) raise web.seeother(return_path) class StarredModulesList(object): ''' Class handles showing of starredModules ''' URL_THIS_PAGE = '/starredModules' def GET(self): ''' This function is called when /starredModules is accessed. ''' web.header('X-Frame-Options', 'SAMEORIGIN') web.header('X-Content-Type-Options', 'nosniff') web.header('X-XSS-Protection', '1') if not session.validate_session(): raise web.seeother('/login') else: starred_module_infos = model.get_starred_modules(web.cookies().get('user')) return RENDER.starredModulesListing(starred_module_infos)
1,894
538
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Feb 23 16:21:35 2021 @author: Paolo Cozzi <paolo.cozzi@ibba.cnr.it> """ import os import logging import pathlib import pycountry import mongoengine from enum import Enum from typing import Union from pymongo import database, ReturnDocument from dotenv import find_dotenv, load_dotenv from .utils import get_project_dir SPECIES2CODE = { "Sheep": "OA", "Goat": "CH" } SMARTERDB = "smarter" DB_ALIAS = "smarterdb" # Get an instance of a logger logger = logging.getLogger(__name__) class SmarterDBException(Exception): pass def global_connection(database_name: str = SMARTERDB): # find .env automagically by walking up directories until it's found, then # load up the .env entries as environment variables load_dotenv(find_dotenv()) # TODO: track connection somewhere return mongoengine.connect( database_name, username=os.getenv("MONGODB_SMARTER_USER"), password=os.getenv("MONGODB_SMARTER_PASS"), authentication_source='admin', alias=DB_ALIAS) def complement(genotype: str): bases = { "A": "T", "T": "A", "C": "G", "G": "C", "/": "/" } result = "" for base in genotype: result += bases[base] return result class SmarterInfo(mongoengine.Document): """A class to track database status informations""" id = mongoengine.StringField(primary_key=True) version = mongoengine.StringField(required=True) working_assemblies = mongoengine.DictField() plink_specie_opt = mongoengine.DictField() last_updated = mongoengine.DateTimeField() meta = { 'db_alias': DB_ALIAS, 'collection': 'smarterInfo' } def __str__(self): return f"{self.id}: {self.version}" class Counter(mongoengine.Document): """A class to deal with counter collection (created when initializing smarter database) """ id = mongoengine.StringField(primary_key=True) sequence_value = mongoengine.IntField(required=True, default=0) meta = { 'db_alias': DB_ALIAS, 'collection': 'counters' } def __str__(self): return f"{self.id}: {self.sequence_value}" class SupportedChip(mongoengine.Document): name = mongoengine.StringField(required=True, unique=True) species = mongoengine.StringField(required=True) manifacturer = mongoengine.StringField() n_of_snps = mongoengine.IntField(default=0) meta = { 'db_alias': DB_ALIAS, 'collection': 'supportedChips' } def __str__(self): return f"'{self.name}' ({self.species})" class BreedAlias(mongoengine.EmbeddedDocument): fid = mongoengine.StringField(required=True) dataset = mongoengine.ReferenceField( 'Dataset', db_field="dataset_id") country = mongoengine.StringField() def __str__(self): return f"{self.fid}: {self.dataset}" class Breed(mongoengine.Document): species = mongoengine.StringField(required=True) name = mongoengine.StringField(required=True) code = mongoengine.StringField(required=True) aliases = mongoengine.ListField( mongoengine.EmbeddedDocumentField(BreedAlias)) n_individuals = mongoengine.IntField() meta = { 'db_alias': DB_ALIAS, 'collection': 'breeds', 'indexes': [ { 'fields': [ "species", "code" ], 'unique': True, 'collation': {'locale': 'en', 'strength': 1} }, { 'fields': [ "species", "name" ], 'unique': True, 'collation': {'locale': 'en', 'strength': 1} } ] } def __str__(self): return f"{self.name} ({self.code}) {self.species}" def get_or_create_breed( species: str, name: str, code: str, aliases: list = []): logger.debug(f"Checking: '{species}':'{name}':'{code}'") # get a breed object relying on parameters qs = Breed.objects(species=species, name=name, code=code) modified = False if qs.count() == 1: breed = qs.get() logger.debug(f"Got {breed}") for alias in aliases: if alias not in breed.aliases: # track for update modified = True logger.info(f"Adding '{alias}' to '{breed}' aliases") breed.aliases.append(alias) elif qs.count() == 0: logger.debug("Create a new breed object") modified = True breed = Breed( species=species, name=name, code=code, aliases=aliases, n_individuals=0 ) else: # should never see this relying on collection unique keys raise SmarterDBException( f"Got {qs.count()} results for '{species}':'{name}':'{code}'") if modified: logger.debug(f"Save '{breed}' to database") breed.save() return breed, modified class Dataset(mongoengine.Document): """Describe a dataset instace with fields owned by data types""" file = mongoengine.StringField(required=True, unique=True) uploader = mongoengine.StringField() size_ = mongoengine.StringField(db_field="size") partner = mongoengine.StringField() # HINT: should country, species and breeds be a list of items? country = mongoengine.StringField() species = mongoengine.StringField() breed = mongoengine.StringField() n_of_individuals = mongoengine.IntField() n_of_records = mongoengine.IntField() trait = mongoengine.StringField() gene_array = mongoengine.StringField() # add type tag type_ = mongoengine.ListField(mongoengine.StringField(), db_field="type") # file contents contents = mongoengine.ListField(mongoengine.StringField()) # track the original chip_name with dataset chip_name = mongoengine.StringField() meta = { 'db_alias': DB_ALIAS, 'collection': 'dataset' } def __str__(self): return f"file={self.file}, uploader={self.uploader}" @property def working_dir(self) -> pathlib.PosixPath: """returns the locations of dataset working directory. Could exists or not Returns: pathlib.PosixPath: a subdirectory in /data/interim/ """ if not self.id: raise SmarterDBException( "Can't define working dir. Object need to be stored in " "database") return get_project_dir() / f"data/interim/{self.id}" @property def result_dir(self) -> pathlib.PosixPath: """returns the locations of dataset processed directory. Could exists or not Returns: pathlib.PosixPath: a subdirectory in /data/processed/ """ if not self.id: raise SmarterDBException( "Can't define result dir. Object need to be stored in " "database") return get_project_dir() / f"data/processed/{self.id}" def getNextSequenceValue( sequence_name: str, mongodb: database.Database): # this method is something similar to findAndModify, # update a document and after get the UPDATED document # https://docs.mongodb.com/manual/reference/method/db.collection.findAndModify/index.html#db.collection.findAndModify sequenceDocument = mongodb.counters.find_one_and_update( {"_id": sequence_name}, {"$inc": {"sequence_value": 1}}, return_document=ReturnDocument.AFTER ) return sequenceDocument['sequence_value'] def getSmarterId( species: str, country: str, breed: str, mongodb: database.Database): # species, country and breed shold be defined in order to call this func if not species or not country or not breed: raise SmarterDBException( "species, country and breed should be defined when calling " "getSmarterId" ) # get species code if species not in SPECIES2CODE: raise SmarterDBException( "Species %s not managed by smarter" % (species)) species_code = SPECIES2CODE[species] # get country code (two letters) country = pycountry.countries.get(name=country) country_code = country.alpha_2 # get breed code from database breed_code = mongodb.breeds.find_one( {"species": species, "name": breed})["code"] # derive sequence_name from species sequence_name = f"sample{species}" # get the sequence number and define smarter id sequence_id = getNextSequenceValue(sequence_name, mongodb) # padding numbers sequence_id = str(sequence_id).zfill(9) smarter_id = f"{country_code}{species_code}-{breed_code}-{sequence_id}" return smarter_id class SEX(bytes, Enum): UNKNOWN = (0, "Unknown") MALE = (1, "Male") FEMALE = (2, "Female") def __new__(cls, value, label): obj = bytes.__new__(cls, [value]) obj._value_ = value obj.label = label return obj def __str__(self): return self.label @classmethod def from_string(cls, value: str): """Get proper type relying on input string Args: value (str): required sex as string Returns: SEX: A sex instance (MALE, FEMALE, UNKNOWN) """ if type(value) != str: raise SmarterDBException("Provided value should be a 'str' type") value = value.upper() if value in ['M', 'MALE', "1"]: return cls.MALE elif value in ['F', 'FEMALE', "2"]: return cls.FEMALE else: logger.debug( f"Unmanaged sex '{value}': return '{cls.UNKNOWN}'") return cls.UNKNOWN class Phenotype(mongoengine.DynamicEmbeddedDocument): """A class to deal with Phenotype. A dynamic document and not a generic DictField since that there can be attributes which could be enforced to have certain values. All other attributes could be set without any assumptions """ purpose = mongoengine.StringField() chest_girth = mongoengine.FloatField() height = mongoengine.FloatField() length = mongoengine.FloatField() def __str__(self): return f"{self.to_json()}" class SAMPLETYPE(Enum): FOREGROUND = 'foreground' BACKGROUND = 'background' class SampleSpecies(mongoengine.Document): original_id = mongoengine.StringField(required=True) smarter_id = mongoengine.StringField(required=True, unique=True) country = mongoengine.StringField(required=True) species = mongoengine.StringField(required=True) breed = mongoengine.StringField(required=True) breed_code = mongoengine.StringField(min_length=3) # this will be a original_id alias (a different sample name in original # data file) alias = mongoengine.StringField() # required to search a sample relying only on original ID dataset = mongoengine.ReferenceField( Dataset, db_field="dataset_id", reverse_delete_rule=mongoengine.DENY ) # add type tag type_ = mongoengine.EnumField(SAMPLETYPE, db_field="type", required=True) # track the original chip_name with sample chip_name = mongoengine.StringField() # define enum types for sex sex = mongoengine.EnumField(SEX) # GPS location # NOTE: X, Y where X is longitude, Y latitude locations = mongoengine.ListField(mongoengine.PointField(), default=None) # additional (not modelled) metadata metadata = mongoengine.DictField(default=None) # for phenotypes phenotype = mongoengine.EmbeddedDocumentField(Phenotype, default=None) meta = { 'abstract': True, } def save(self, *args, **kwargs): """Custom save method. Deal with smarter_id before save""" if not self.smarter_id: logger.debug(f"Determining smarter id for {self.original_id}") # get the pymongo connection object conn = mongoengine.connection.get_db(alias=DB_ALIAS) # even is species, country and breed are required fields for # SampleSpecies document, their value will not be evaluated until # super().save() is called. I can't call it before determining # a smarter_id self.smarter_id = getSmarterId( self.species, self.country, self.breed, conn) # default save method super(SampleSpecies, self).save(*args, **kwargs) def __str__(self): return f"{self.smarter_id} ({self.breed})" class SampleSheep(SampleSpecies): # try to model relationship between samples father_id = mongoengine.LazyReferenceField( 'SampleSheep', passthrough=True, reverse_delete_rule=mongoengine.NULLIFY ) mother_id = mongoengine.LazyReferenceField( 'SampleSheep', passthrough=True, reverse_delete_rule=mongoengine.NULLIFY ) meta = { 'db_alias': DB_ALIAS, 'collection': 'sampleSheep' } class SampleGoat(SampleSpecies): # try to model relationship between samples father_id = mongoengine.LazyReferenceField( 'SampleGoat', passthrough=True, reverse_delete_rule=mongoengine.NULLIFY ) mother_id = mongoengine.LazyReferenceField( 'SampleGoat', passthrough=True, reverse_delete_rule=mongoengine.NULLIFY ) meta = { 'db_alias': DB_ALIAS, 'collection': 'sampleGoat' } def get_or_create_sample( SampleSpecies: Union[SampleGoat, SampleSheep], original_id: str, dataset: Dataset, type_: str, breed: Breed, country: str, chip_name: str = None, sex: SEX = None, alias: str = None) -> Union[SampleGoat, SampleSheep]: """Get or create a sample providing attributes (search for original_id in provided dataset Args: SampleSpecies: (Union[SampleGoat, SampleSheep]): the class required for insert/update original_id (str): The original_id in the dataset dataset (Dataset): the dataset instance used to register sample type_ (str): "background" or "foreground" breed (Breed): A breed instance country (str): Country as a string chip_name (str): the chip name sex (SEX): A SEX instance alias (str): an original_id alias Returns: Union[SampleGoat, SampleSheep]: a SampleSpecies instance """ created = False # search for sample in database qs = SampleSpecies.objects( original_id=original_id, dataset=dataset) if qs.count() == 1: logger.debug(f"Sample '{original_id}' found in database") sample = qs.get() elif qs.count() == 0: # insert sample into database logger.info(f"Registering sample '{original_id}' in database") sample = SampleSpecies( original_id=original_id, country=country, species=dataset.species, breed=breed.name, breed_code=breed.code, dataset=dataset, type_=type_, chip_name=chip_name, sex=sex, alias=alias ) sample.save() # incrementing breed n_individuals counter breed.n_individuals += 1 breed.save() created = True else: raise SmarterDBException( f"Got {qs.count()} results for '{original_id}'") return sample, created def get_sample_type(dataset: Dataset): """ test if foreground or background dataset Args: dataset (Dataset): the dataset instance used to register sample Returns: str: sample type ("background" or "foreground") """ type_ = None for sampletype in SAMPLETYPE: if sampletype.value in dataset.type_: logger.debug( f"Found {sampletype.value} in {dataset.type_}") type_ = sampletype.value break return type_ class Consequence(mongoengine.EmbeddedDocument): pass class Location(mongoengine.EmbeddedDocument): ss_id = mongoengine.StringField() version = mongoengine.StringField(required=True) chrom = mongoengine.StringField(required=True) position = mongoengine.IntField(required=True) alleles = mongoengine.StringField() illumina = mongoengine.StringField(required=True) illumina_forward = mongoengine.StringField() illumina_strand = mongoengine.StringField() affymetrix_ab = mongoengine.StringField() strand = mongoengine.StringField() imported_from = mongoengine.StringField(required=True) # this could be the manifactured date or the last updated date = mongoengine.DateTimeField() consequences = mongoengine.ListField( mongoengine.EmbeddedDocumentField(Consequence), default=None) def __init__(self, *args, **kwargs): illumina_top = None # remove illumina top from arguments if 'illumina_top' in kwargs: illumina_top = kwargs.pop('illumina_top') # initialize base object super(Location, self).__init__(*args, **kwargs) # fix illumina top if necessary if illumina_top: self.illumina_top = illumina_top @property def illumina_top(self): """Return genotype in illumina top format""" if self.illumina_strand in ['BOT', 'bottom']: return complement(self.illumina) elif (not self.illumina_strand or self.illumina_strand in ['TOP', 'top']): return self.illumina else: raise SmarterDBException( f"{self.illumina_strand} not managed") @illumina_top.setter def illumina_top(self, genotype: str): if (not self.illumina_strand or self.illumina_strand in ['TOP', 'top']): self.illumina = genotype elif self.illumina_strand in ['BOT', 'bottom']: self.illumina = complement(genotype) else: raise SmarterDBException( f"{self.illumina_strand} not managed") def __str__(self): return ( f"({self.imported_from}:{self.version}) " f"{self.chrom}:{self.position} [{self.illumina_top}]" ) def __eq__(self, other): if super().__eq__(other): return True else: # check by positions for attribute in ["chrom", "position"]: if getattr(self, attribute) != getattr(other, attribute): return False # check genotype equality if self.illumina_top != other.illumina_top: return False return True def __check_coding(self, genotype: list, coding: str, missing: str): """Internal method to check genotype coding""" # get illumina data as an array data = getattr(self, coding).split("/") for allele in genotype: # mind to missing values. If missing can't be equal to illumina_top if allele in missing: continue if allele not in data: return False return True def is_top(self, genotype: list, missing: list = ["0", "-"]) -> bool: """Return True if genotype is compatible with illumina TOP coding Args: genotype (list): a list of two alleles (ex ['A','C']) missing (str): missing allele string (def "0") Returns: bool: True if in top coordinates """ return self.__check_coding(genotype, "illumina_top", missing) def is_forward(self, genotype: list, missing: list = ["0", "-"]) -> bool: """Return True if genotype is compatible with illumina FORWARD coding Args: genotype (list): a list of two alleles (ex ['A','C']) missing (str): missing allele string (def "0") Returns: bool: True if in top coordinates """ return self.__check_coding(genotype, "illumina_forward", missing) def is_ab(self, genotype: list, missing: list = ["0", "-"]) -> bool: """Return True if genotype is compatible with illumina AB coding Args: genotype (list): a list of two alleles (ex ['A','B']) missing (str): missing allele string (def "-") Returns: bool: True if in top coordinates """ for allele in genotype: # mind to missing valies if allele not in ["A", "B"] + missing: return False return True def is_affymetrix( self, genotype: list, missing: list = ["0", "-"]) -> bool: """Return True if genotype is compatible with affymetrix coding Args: genotype (list): a list of two alleles (ex ['A','C']) missing (str): missing allele string (def "0") Returns: bool: True if in top coordinates """ return self.__check_coding(genotype, "affymetrix_ab", missing) def forward2top(self, genotype: list, missing: list = ["0", "-"]) -> list: """Convert an illumina forward SNP in a illumina top snp Args: genotype (list): a list of two alleles (ex ['A','C']) missing (str): missing allele string (def "0") Returns: list: The genotype in top format """ # get illumina data as an array forward = self.illumina_forward.split("/") top = self.illumina_top.split("/") result = [] for allele in genotype: # mind to missing values if allele in missing: result.append("0") elif allele not in forward: raise SmarterDBException( f"{genotype} is not in forward coding") else: result.append(top[forward.index(allele)]) return result def ab2top(self, genotype: list, missing: list = ["0", "-"]) -> list: """Convert an illumina ab SNP in a illumina top snp Args: genotype (list): a list of two alleles (ex ['A','B']) missing (str): missing allele string (def "-") Returns: list: The genotype in top format """ # get illumina data as a dict top = self.illumina_top.split("/") top = {"A": top[0], "B": top[1]} result = [] for allele in genotype: # mind to missing values if allele in missing: result.append("0") elif allele not in ["A", "B"]: raise SmarterDBException( f"{genotype} is not in ab coding") else: result.append(top[allele]) return result def affy2top(self, genotype: list, missing: list = ["0", "-"]) -> list: """Convert an affymetrix SNP in a illumina top snp Args: genotype (list): a list of two alleles (ex ['A','C']) missing (str): missing allele string (def "0") Returns: list: The genotype in top format """ # get illumina data as an array affymetrix = self.affymetrix_ab.split("/") top = self.illumina_top.split("/") result = [] for allele in genotype: # mind to missing values if allele in missing: result.append("0") elif allele not in affymetrix: raise SmarterDBException( f"{genotype} is not in affymetrix coding") else: result.append(top[affymetrix.index(allele)]) return result class VariantSpecies(mongoengine.Document): rs_id = mongoengine.StringField() chip_name = mongoengine.ListField(mongoengine.StringField()) name = mongoengine.StringField(unique=True) # sequence should model both illumina or affymetrix sequences sequence = mongoengine.DictField() locations = mongoengine.ListField( mongoengine.EmbeddedDocumentField(Location)) # HINT: should sender be a Location attribute? sender = mongoengine.StringField() # Affymetryx specific fields # more probe could be assigned to the same SNP probeset_id = mongoengine.ListField(mongoengine.StringField()) affy_snp_id = mongoengine.StringField() cust_id = mongoengine.StringField() # abstract class with custom indexes # TODO: need a index for position (chrom, position, version) meta = { 'abstract': True, 'indexes': [ { 'fields': [ "locations.chrom", "locations.position" ], }, 'probeset_id', 'rs_id' ] } def __str__(self): return (f"name='{self.name}', rs_id='{self.rs_id}'") def save(self, *args, **kwargs): """Custom save method. Deal with variant name before save""" if not self.name and self.affy_snp_id: logger.debug(f"Set variant name to {self.affy_snp_id}") self.name = self.affy_snp_id # default save method super(VariantSpecies, self).save(*args, **kwargs) def get_location_index(self, version: str, imported_from='SNPchiMp v.3'): """Returns location index for assembly version and imported source Args: version (str): assembly version (ex: 'Oar_v3.1') imported_from (str): coordinates source (ex: 'SNPchiMp v.3') Returns: int: the index of the location requested """ for index, location in enumerate(self.locations): if (location.version == version and location.imported_from == imported_from): return index raise SmarterDBException( f"Location '{version}' '{imported_from}' is not in locations" ) def get_location(self, version: str, imported_from='SNPchiMp v.3'): """Returns location for assembly version and imported source Args: version (str): assembly version (ex: 'Oar_v3.1') imported_from (str): coordinates source (ex: 'SNPchiMp v.3') Returns: Location: the genomic coordinates """ def custom_filter(location: Location): if (location.version == version and location.imported_from == imported_from): return True return False locations = list(filter(custom_filter, self.locations)) if len(locations) != 1: raise SmarterDBException( "Couldn't determine a unique location for " f"'{self.name}' '{version}' '{imported_from}'") return locations[0] class VariantSheep(VariantSpecies): meta = { 'db_alias': DB_ALIAS, 'collection': 'variantSheep' } class VariantGoat(VariantSpecies): meta = { 'db_alias': DB_ALIAS, 'collection': 'variantGoat' }
27,362
8,100
#! /usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from __future__ import annotations import torch from botorch.exceptions.errors import BotorchError, BotorchTensorDimensionError from botorch.utils.multi_objective.box_decomposition import NondominatedPartitioning from botorch.utils.testing import BotorchTestCase class TestNonDominatedPartitioning(BotorchTestCase): def test_non_dominated_partitioning(self): tkwargs = {"device": self.device} for dtype in (torch.float, torch.double): tkwargs["dtype"] = dtype partitioning = NondominatedPartitioning(num_outcomes=2) # assert error is raised if pareto_Y has not been computed with self.assertRaises(BotorchError): partitioning.pareto_Y # test eps # no pareto_Y self.assertEqual(partitioning.eps, 1e-6) partitioning = NondominatedPartitioning(num_outcomes=2, eps=1.0) # eps set self.assertEqual(partitioning.eps, 1.0) # set pareto_Y partitioning = NondominatedPartitioning(num_outcomes=2) Y = torch.zeros(1, 2, **tkwargs) partitioning.update(Y=Y) self.assertEqual(partitioning.eps, 1e-6 if dtype == torch.float else 1e-8) # test _update_pareto_Y partitioning.Y = -Y self.assertFalse(partitioning._update_pareto_Y()) # test m=2 arange = torch.arange(3, 9, **tkwargs) pareto_Y = torch.stack([arange, 11 - arange], dim=-1) Y = torch.cat( [ pareto_Y, torch.tensor( [[8.0, 2.0], [7.0, 1.0]], **tkwargs ), # add some non-pareto elements ], dim=0, ) partitioning = NondominatedPartitioning(num_outcomes=2, Y=Y) sorting = torch.argsort(pareto_Y[:, 0], descending=True) self.assertTrue(torch.equal(pareto_Y[sorting], partitioning.pareto_Y)) ref_point = torch.zeros(2, **tkwargs) inf = float("inf") expected_cell_bounds = torch.tensor( [ [ [8.0, 0.0], [7.0, 3.0], [6.0, 4.0], [5.0, 5.0], [4.0, 6.0], [3.0, 7.0], [0.0, 8.0], ], [ [inf, inf], [8.0, inf], [7.0, inf], [6.0, inf], [5.0, inf], [4.0, inf], [3.0, inf], ], ], **tkwargs ) cell_bounds = partitioning.get_hypercell_bounds(ref_point) self.assertTrue(torch.equal(cell_bounds, expected_cell_bounds)) # test compute hypervolume hv = partitioning.compute_hypervolume(ref_point) self.assertEqual(hv, 49.0) # test error when reference is not worse than all pareto_Y with self.assertRaises(ValueError): partitioning.compute_hypervolume(pareto_Y.max(dim=0).values) # test error with partition_non_dominated_space_2d for m=3 partitioning = NondominatedPartitioning( num_outcomes=3, Y=torch.zeros(1, 3, **tkwargs) ) with self.assertRaises(BotorchTensorDimensionError): partitioning.partition_non_dominated_space_2d() # test m=3 pareto_Y = torch.tensor( [[1.0, 6.0, 8.0], [2.0, 4.0, 10.0], [3.0, 5.0, 7.0]], **tkwargs ) partitioning = NondominatedPartitioning(num_outcomes=3, Y=pareto_Y) sorting = torch.argsort(pareto_Y[:, 0], descending=True) self.assertTrue(torch.equal(pareto_Y[sorting], partitioning.pareto_Y)) ref_point = torch.tensor([-1.0, -2.0, -3.0], **tkwargs) expected_cell_bounds = torch.tensor( [ [ [1.0, 4.0, 7.0], [-1.0, -2.0, 10.0], [-1.0, 4.0, 8.0], [1.0, -2.0, 10.0], [1.0, 4.0, 8.0], [-1.0, 6.0, -3.0], [1.0, 5.0, -3.0], [-1.0, 5.0, 8.0], [2.0, -2.0, 7.0], [2.0, 4.0, 7.0], [3.0, -2.0, -3.0], [2.0, -2.0, 8.0], [2.0, 5.0, -3.0], ], [ [2.0, 5.0, 8.0], [1.0, 4.0, inf], [1.0, 5.0, inf], [2.0, 4.0, inf], [2.0, 5.0, inf], [1.0, inf, 8.0], [2.0, inf, 8.0], [2.0, inf, inf], [3.0, 4.0, 8.0], [3.0, 5.0, 8.0], [inf, 5.0, 8.0], [inf, 5.0, inf], [inf, inf, inf], ], ], **tkwargs ) cell_bounds = partitioning.get_hypercell_bounds(ref_point) # cell bounds can have different order num_matches = ( (cell_bounds.unsqueeze(0) == expected_cell_bounds.unsqueeze(1)) .all(dim=-1) .any(dim=0) .sum() ) self.assertTrue(num_matches, 9) # test compute hypervolume hv = partitioning.compute_hypervolume(ref_point) self.assertEqual(hv, 358.0) # TODO: test approximate decomposition
6,143
1,969
# -*- coding: utf-8 -*- from __future__ import absolute_import from __future__ import unicode_literals import ast import sys import pytest from pyupgrade import _fix_dict_set from pyupgrade import _fix_escape_sequences from pyupgrade import _fix_format_literals from pyupgrade import _fix_fstrings from pyupgrade import _fix_new_style_classes from pyupgrade import _fix_percent_format from pyupgrade import _fix_six from pyupgrade import _fix_super from pyupgrade import _fix_tokens from pyupgrade import _fix_unicode_literals from pyupgrade import _imports_unicode_literals from pyupgrade import _is_bytestring from pyupgrade import _percent_to_format from pyupgrade import _simplify_conversion_flag from pyupgrade import main from pyupgrade import parse_format from pyupgrade import parse_percent_format from pyupgrade import unparse_parsed_string @pytest.mark.parametrize( 's', ( '', 'foo', '{}', '{0}', '{named}', '{!r}', '{:>5}', '{{', '}}', '{0!s:15}' ), ) def test_roundtrip_text(s): assert unparse_parsed_string(parse_format(s)) == s @pytest.mark.parametrize( ('s', 'expected'), ( ('{:}', '{}'), ('{0:}', '{0}'), ('{0!r:}', '{0!r}'), ), ) def test_intentionally_not_round_trip(s, expected): # Our unparse simplifies empty parts, whereas stdlib allows them ret = unparse_parsed_string(parse_format(s)) assert ret == expected @pytest.mark.parametrize( ('s', 'expected'), ( # Don't touch empty set literals ('set()', 'set()'), # Don't touch set(empty literal) with newlines in them (may create # syntax errors) ('set((\n))', 'set((\n))'), # Don't touch weird looking function calls -- use autopep8 or such # first ('set (())', 'set (())'), ('set ((1, 2))', 'set ((1, 2))'), # Take a set literal with an empty tuple / list and remove the arg ('set(())', 'set()'), ('set([])', 'set()'), # Remove spaces in empty set literals ('set(( ))', 'set()'), # Some "normal" test cases ('set((1, 2))', '{1, 2}'), ('set([1, 2])', '{1, 2}'), ('set(x for x in y)', '{x for x in y}'), ('set([x for x in y])', '{x for x in y}'), # These are strange cases -- the ast doesn't tell us about the parens # here so we have to parse ourselves ('set((x for x in y))', '{x for x in y}'), ('set(((1, 2)))', '{1, 2}'), # The ast also doesn't tell us about the start of the tuple in this # generator expression ('set((a, b) for a, b in y)', '{(a, b) for a, b in y}'), # The ast also doesn't tell us about the start of the tuple for # tuple of tuples ('set(((1, 2), (3, 4)))', '{(1, 2), (3, 4)}'), # Lists where the first element is a tuple also gives the ast trouble # The first element lies about the offset of the element ('set([(1, 2), (3, 4)])', '{(1, 2), (3, 4)}'), ( 'set(\n' ' [(1, 2)]\n' ')', '{\n' ' (1, 2)\n' '}', ), ('set([((1, 2)), (3, 4)])', '{((1, 2)), (3, 4)}'), # And it gets worse ('set((((1, 2),),))', '{((1, 2),)}'), # Some multiline cases ('set(\n(1, 2))', '{\n1, 2}'), ('set((\n1,\n2,\n))\n', '{\n1,\n2,\n}\n'), # Nested sets ( 'set((frozenset(set((1, 2))), frozenset(set((3, 4)))))', '{frozenset({1, 2}), frozenset({3, 4})}', ), # Remove trailing commas on inline things ('set((1,))', '{1}'), ('set((1, ))', '{1}'), # Remove trailing commas after things ('set([1, 2, 3,],)', '{1, 2, 3}'), ('set((x for x in y),)', '{x for x in y}'), ( 'set(\n' ' (x for x in y),\n' ')', '{\n' ' x for x in y\n' '}', ), ), ) def test_sets(s, expected): ret = _fix_dict_set(s) assert ret == expected @pytest.mark.xfail(sys.version_info >= (3, 7), reason='genexp trailing comma') @pytest.mark.parametrize( ('s', 'expected'), ( ('set(x for x in y,)', '{x for x in y}'), ( 'set(\n' ' x for x in y,\n' ')', '{\n' ' x for x in y\n' '}', ), ), ) def test_sets_generators_trailing_comas(s, expected): ret = _fix_dict_set(s) assert ret == expected @pytest.mark.parametrize( ('s', 'expected'), ( # Don't touch irrelevant code ('x = 5', 'x = 5'), ('dict()', 'dict()'), # Don't touch syntax errors ('(', '('), # Don't touch strange looking calls ('dict ((a, b) for a, b in y)', 'dict ((a, b) for a, b in y)'), # dict of generator expression ('dict((a, b) for a, b in y)', '{a: b for a, b in y}'), ('dict((a, b,) for a, b in y)', '{a: b for a, b in y}'), ('dict((a, b, ) for a, b in y)', '{a: b for a, b in y}'), ('dict([a, b] for a, b in y)', '{a: b for a, b in y}'), # Parenthesized target ('dict(((a, b)) for a, b in y)', '{a: b for a, b in y}'), # dict of list comprehension ('dict([(a, b) for a, b in y])', '{a: b for a, b in y}'), # ast doesn't tell us about the tuple in the list ('dict([(a, b), c] for a, b, c in y)', '{(a, b): c for a, b, c in y}'), # ast doesn't tell us about parenthesized keys ('dict(((a), b) for a, b in y)', '{(a): b for a, b in y}'), # Nested dictcomps ( 'dict((k, dict((k2, v2) for k2, v2 in y2)) for k, y2 in y)', '{k: {k2: v2 for k2, v2 in y2} for k, y2 in y}', ), # This doesn't get fixed by autopep8 and can cause a syntax error ('dict((a, b)for a, b in y)', '{a: b for a, b in y}'), # Need to remove trailing commas on the element ( 'dict(\n' ' (\n' ' a,\n' ' b,\n' ' )\n' ' for a, b in y\n' ')', # Ideally, this'll go through some other formatting tool before # being committed. Shrugs! '{\n' ' a:\n' ' b\n' ' for a, b in y\n' '}', ), # Don't rewrite kwargd dicts ( 'dict(((a, b) for a, b in y), x=1)', 'dict(((a, b) for a, b in y), x=1)', ), ( 'dict(((a, b) for a, b in y), **kwargs)', 'dict(((a, b) for a, b in y), **kwargs)', ), # Don't gobble the last paren in a dictcomp ( 'x(\n' ' dict(\n' ' (a, b) for a, b in y\n' ' )\n' ')', 'x(\n' ' {\n' ' a: b for a, b in y\n' ' }\n' ')', ) ), ) def test_dictcomps(s, expected): ret = _fix_dict_set(s) assert ret == expected @pytest.mark.parametrize( ('s', 'expected'), ( # Don't touch syntax errors ('"{0}"format(1)', '"{0}"format(1)'), # Don't touch py27 format strings ("'{}'.format(1)", "'{}'.format(1)"), # Don't touch invalid format strings ("'{'.format(1)", "'{'.format(1)"), ("'}'.format(1)", "'}'.format(1)"), # Don't touch non-format strings ("x = ('{0} {1}',)\n", "x = ('{0} {1}',)\n"), # Don't touch non-incrementing integers ("'{0} {0}'.format(1)", "'{0} {0}'.format(1)"), # Simplest case ("'{0}'.format(1)", "'{}'.format(1)"), ("'{0:x}'.format(30)", "'{:x}'.format(30)"), ("x = '{0}'.format(1)", "x = '{}'.format(1)"), # Multiline strings ("'''{0}\n{1}\n'''.format(1, 2)", "'''{}\n{}\n'''.format(1, 2)"), # Multiple implicitly-joined strings ("'{0}' '{1}'.format(1, 2)", "'{}' '{}'.format(1, 2)"), # Multiple implicitly-joined strings over lines ( 'print(\n' " 'foo{0}'\n" " 'bar{1}'.format(1, 2)\n" ')', 'print(\n' " 'foo{}'\n" " 'bar{}'.format(1, 2)\n" ')', ), # Multiple implicitly-joind strings over lines with comments ( 'print(\n' " 'foo{0}' # ohai\n" " 'bar{1}'.format(1, 2)\n" ')', 'print(\n' " 'foo{}' # ohai\n" " 'bar{}'.format(1, 2)\n" ')', ), # Formats can be embedded in formats, leave these alone? ("'{0:<{1}}'.format(1, 4)", "'{0:<{1}}'.format(1, 4)"), # joined by backslash ( 'x = "foo {0}" \\\n' ' "bar {1}".format(1, 2)', 'x = "foo {}" \\\n' ' "bar {}".format(1, 2)', ) ), ) def test_format_literals(s, expected): ret = _fix_format_literals(s) assert ret == expected @pytest.mark.parametrize( ('s', 'expected'), ( ('import x', False), ('from foo import bar', False), ('x = 5', False), ('from __future__ import unicode_literals', True), ( '"""docstring"""\n' 'from __future__ import unicode_literals', True, ), ( 'from __future__ import absolute_import\n' 'from __future__ import unicode_literals\n', True, ), ), ) def test_imports_unicode_literals(s, expected): assert _imports_unicode_literals(s) is expected @pytest.mark.parametrize( ('s', 'py3_plus', 'expected'), ( # Syntax errors are unchanged ('(', False, '('), # Without py3-plus, no replacements ("u''", False, "u''"), # With py3-plus, it removes u prefix ("u''", True, "''"), # Importing unicode_literals also cause it to remove it ( 'from __future__ import unicode_literals\n' 'u""\n', False, 'from __future__ import unicode_literals\n' '""\n', ), # Regression: string containing newline ('"""with newline\n"""', True, '"""with newline\n"""'), ), ) def test_unicode_literals(s, py3_plus, expected): ret = _fix_unicode_literals(s, py3_plus=py3_plus) assert ret == expected @pytest.mark.parametrize( 's', ( '""', r'r"\d"', r"r'\d'", r'r"""\d"""', r"r'''\d'''", # python2 has a bug where `rb'foo'` is tokenized as NAME + STRING r'rb"\d"', # make sure we don't replace an already valid string r'"\\d"', # in python2 `'\u2603'` is literally \\u2603, but transforming based # on that would be incorrect in python3. # intentionally timid here to avoid breaking working python3 code '"\\u2603"', # don't touch already valid escapes r'"\r\n"', # don't touch escaped newlines '"""\\\n"""', '"""\\\r\n"""', '"""\\\r"""', ), ) def test_fix_escape_sequences_noop(s): assert _fix_escape_sequences(s) == s @pytest.mark.parametrize( ('s', 'expected'), ( # no valid escape sequences, make a raw literal (r'"\d"', r'r"\d"'), # when there are valid escape sequences, need to use backslashes (r'"\n\d"', r'"\n\\d"'), # `ur` is not a valid string prefix in python3.x (r'u"\d"', r'u"\\d"'), # `rb` is not a valid string prefix in python2.x (r'b"\d"', r'br"\d"'), # 8 and 9 aren't valid octal digits (r'"\8"', r'r"\8"'), (r'"\9"', r'r"\9"'), # explicit byte strings should not honor string-specific escapes ('b"\\u2603"', 'br"\\u2603"'), ), ) def test_fix_escape_sequences(s, expected): assert _fix_escape_sequences(s) == expected @pytest.mark.xfail(sys.version_info >= (3,), reason='python2 "feature"') @pytest.mark.parametrize( ('s', 'expected'), ( ('5L', '5'), ('5l', '5'), ('123456789123456789123456789L', '123456789123456789123456789'), ), ) def test_long_literals(s, expected): assert _fix_tokens(s) == expected @pytest.mark.parametrize( 's', ( # Any number of zeros is considered a legal token '0', '00', # Don't modify non octal literals '1', '12345', '1.2345', ), ) def test_noop_octal_literals(s): assert _fix_tokens(s) == s @pytest.mark.xfail(sys.version_info >= (3,), reason='python2 "feature"') @pytest.mark.parametrize( ('s', 'expected'), ( ('0755', '0o755'), ('05', '5'), ), ) def test_fix_octal_literal(s, expected): assert _fix_tokens(s) == expected @pytest.mark.parametrize('s', ("b''", 'b""', 'B""', "B''", "rb''", "rb''")) def test_is_bytestring_true(s): assert _is_bytestring(s) is True @pytest.mark.parametrize('s', ('', '""', "''", 'u""', '"b"')) def test_is_bytestring_false(s): assert _is_bytestring(s) is False @pytest.mark.parametrize( ('s', 'expected'), ( ( '""', ( ('""', None), ), ), ( '"%%"', ( ('"', (None, None, None, None, '%')), ('"', None), ), ), ( '"%s"', ( ('"', (None, None, None, None, 's')), ('"', None), ), ), ( '"%s two! %s"', ( ('"', (None, None, None, None, 's')), (' two! ', (None, None, None, None, 's')), ('"', None), ), ), ( '"%(hi)s"', ( ('"', ('hi', None, None, None, 's')), ('"', None), ), ), ( '"%()s"', ( ('"', ('', None, None, None, 's')), ('"', None), ), ), ( '"%#o"', ( ('"', (None, '#', None, None, 'o')), ('"', None), ), ), ( '"% #0-+d"', ( ('"', (None, ' #0-+', None, None, 'd')), ('"', None), ), ), ( '"%5d"', ( ('"', (None, None, '5', None, 'd')), ('"', None), ), ), ( '"%*d"', ( ('"', (None, None, '*', None, 'd')), ('"', None), ) ), ( '"%.f"', ( ('"', (None, None, None, '.', 'f')), ('"', None), ), ), ( '"%.5f"', ( ('"', (None, None, None, '.5', 'f')), ('"', None), ), ), ( '"%.*f"', ( ('"', (None, None, None, '.*', 'f')), ('"', None), ), ), ( '"%ld"', ( ('"', (None, None, None, None, 'd')), ('"', None), ), ), ( '"%(complete)#4.4f"', ( ('"', ('complete', '#', '4', '.4', 'f')), ('"', None), ), ), ), ) def test_parse_percent_format(s, expected): assert parse_percent_format(s) == expected @pytest.mark.parametrize( ('s', 'expected'), ( ('%s', '{}'), ('%%%s', '%{}'), ('%(foo)s', '{foo}'), ('%2f', '{:2f}'), ('%r', '{!r}'), ('%a', '{!a}'), ), ) def test_percent_to_format(s, expected): assert _percent_to_format(s) == expected @pytest.mark.parametrize( ('s', 'expected'), ( ('', ''), (' ', ' '), (' ', ' '), ('#0- +', '#<+'), ('-', '<'), ), ) def test_simplify_conversion_flag(s, expected): assert _simplify_conversion_flag(s) == expected @pytest.mark.parametrize( 's', ( # cannot determine whether `unknown_type` is tuple or not '"%s" % unknown_type', # format of bytestring cannot be changed to `.format(...)` 'b"%s" % (b"bytestring",)', # out-of-order parameter consumption '"%*s" % (5, "hi")', '"%.*s" % (5, "hi")', # potential conversion to int required '"%d" % (flt,)', '"%i" % (flt,)', '"%u" % (flt,)', # potential conversion to character required '"%c" % (some_string,)', # different output vs .format() in python 2 '"%#o" % (123,)', # no format equivalent '"%()s" % {"": "empty"}', # different output in python2 / python 3 '"%4%" % ()', # no equivalent in format specifier '"%.2r" % (1.25)', '"%.2a" % (1.25)', # non-string mod 'i % 3', # dict format but not keyed arguments '"%s" % {"k": "v"}', # dict format must have valid identifiers '"%()s" % {"": "bar"}', '"%(1)s" % {"1": "bar"}', # don't trigger `SyntaxError: keyword argument repeated` '"%(a)s" % {"a": 1, "a": 2}', # don't rewrite string-joins in dict literal '"%(ab)s" % {"a" "b": 1}', # don't rewrite strangely styled things '"%(a)s" % {"a" : 1}', # don't rewrite non-str keys '"%(1)s" % {1: 2, "1": 2}', # don't rewrite keyword keys '"%(and)s" % {"and": 2}', ), ) def test_percent_format_noop(s): assert _fix_percent_format(s) == s def _has_16806_bug(): # See https://bugs.python.org/issue16806 return ast.parse('"""\n"""').body[0].value.col_offset == -1 @pytest.mark.xfail(not _has_16806_bug(), reason='multiline string parse bug') def test_percent_format_noop_if_bug_16806(): s = '"""%s\n""" % ("issue16806",)' assert _fix_percent_format(s) == s @pytest.mark.parametrize( ('s', 'expected'), ( # tuple ('"trivial" % ()', '"trivial".format()'), ('"%s" % ("simple",)', '"{}".format("simple")'), ('"%s" % ("%s" % ("nested",),)', '"{}".format("{}".format("nested"))'), ('"%s%% percent" % (15,)', '"{}% percent".format(15)'), ('"%3f" % (15,)', '"{:3f}".format(15)'), ('"%-5s" % ("hi",)', '"{:<5}".format("hi")'), ('"%9s" % (5,)', '"{:>9}".format(5)'), ('"brace {} %s" % (1,)', '"brace {{}} {}".format(1)'), ( '"%s" % (\n' ' "trailing comma",\n' ')\n', '"{}".format(\n' ' "trailing comma",\n' ')\n', ), # dict ('"%(k)s" % {"k": "v"}', '"{k}".format(k="v")'), ('"%(to_list)s" % {"to_list": []}', '"{to_list}".format(to_list=[])'), ), ) def test_percent_format(s, expected): assert _fix_percent_format(s) == expected @pytest.mark.xfail @pytest.mark.parametrize( ('s', 'expected'), ( # currently the approach does not attempt to consider joined strings ( 'paren_continue = (\n' ' "foo %s "\n' ' "bar %s" % (x, y)\n' ')\n', 'paren_continue = (\n' ' "foo {} "\n' ' "bar {}".format(x, y)\n' ')\n', ), ( 'paren_string = (\n' ' "foo %s "\n' ' "bar %s"\n' ') % (x, y)\n', 'paren_string = (\n' ' "foo {} "\n' ' "bar {}"\n' ').format(x, y)\n', ), ( 'paren_continue = (\n' ' "foo %(foo)s "\n' ' "bar %(bar)s" % {"foo": x, "bar": y}\n' ')\n', 'paren_continue = (\n' ' "foo {foo} "\n' ' "bar {bar}".format(foo=x, bar=y)\n' ')\n', ), ( 'paren_string = (\n' ' "foo %(foo)s "\n' ' "bar %(bar)s"\n' ') % {"foo": x, "bar": y}\n', 'paren_string = (\n' ' "foo {foo} "\n' ' "bar {bar}"\n' ').format(foo=x, bar=y)\n', ), ), ) def test_percent_format_todo(s, expected): assert _fix_percent_format(s) == expected @pytest.mark.parametrize( 's', ( # syntax error 'x(', 'class C(Base):\n' ' def f(self):\n' ' super().f()\n', # super class doesn't match class name 'class C(Base):\n' ' def f(self):\n' ' super(Base, self).f()\n', # super outside of a class (technically legal!) 'def f(self):\n' ' super(C, self).f()\n', # super used in a comprehension 'class C(Base):\n' ' def f(self):\n' ' return [super(C, self).f() for _ in ()]\n', 'class C(Base):\n' ' def f(self):\n' ' return {super(C, self).f() for _ in ()}\n', 'class C(Base):\n' ' def f(self):\n' ' return (super(C, self).f() for _ in ())\n', 'class C(Base):\n' ' def f(self):\n' ' return {True: super(C, self).f() for _ in ()}\n', # nested comprehension 'class C(Base):\n' ' def f(self):\n' ' return [\n' ' (\n' ' [_ for _ in ()],\n' ' super(C, self).f(),\n' ' )\n' ' for _ in ()' ' ]\n', # super in a closure 'class C(Base):\n' ' def f(self):\n' ' def g():\n' ' super(C, self).f()\n' ' g()\n', 'class C(Base):\n' ' def f(self):\n' ' g = lambda: super(C, self).f()\n' ' g()\n', ), ) def test_fix_super_noop(s): assert _fix_super(s) == s @pytest.mark.parametrize( ('s', 'expected'), ( ( 'class C(Base):\n' ' def f(self):\n' ' super(C, self).f()\n', 'class C(Base):\n' ' def f(self):\n' ' super().f()\n', ), ( 'class C(Base):\n' ' def f(self):\n' ' super (C, self).f()\n', 'class C(Base):\n' ' def f(self):\n' ' super ().f()\n', ), ( 'class Outer(object):\n' ' class C(Base):\n' ' def f(self):\n' ' super (C, self).f()\n', 'class Outer(object):\n' ' class C(Base):\n' ' def f(self):\n' ' super ().f()\n', ), ( 'class C(Base):\n' ' f = lambda self: super(C, self).f()\n', 'class C(Base):\n' ' f = lambda self: super().f()\n' ), ( 'class C(Base):\n' ' @classmethod\n' ' def f(cls):\n' ' super(C, cls).f()\n', 'class C(Base):\n' ' @classmethod\n' ' def f(cls):\n' ' super().f()\n', ), ), ) def test_fix_super(s, expected): assert _fix_super(s) == expected @pytest.mark.parametrize( 's', ( # syntax error 'x = (', # does not inherit from `object` 'class C(B): pass', ), ) def test_fix_new_style_classes_noop(s): assert _fix_new_style_classes(s) == s @pytest.mark.parametrize( ('s', 'expected'), ( ( 'class C(object): pass', 'class C: pass', ), ( 'class C(\n' ' object,\n' '): pass', 'class C: pass', ), ( 'class C(B, object): pass', 'class C(B): pass', ), ( 'class C(B, (object)): pass', 'class C(B): pass', ), ( 'class C(B, ( object )): pass', 'class C(B): pass', ), ( 'class C((object)): pass', 'class C: pass', ), ( 'class C(\n' ' B,\n' ' object,\n' '): pass\n', 'class C(\n' ' B,\n' '): pass\n', ), ( 'class C(\n' ' B,\n' ' object\n' '): pass\n', 'class C(\n' ' B\n' '): pass\n', ), # only legal in python2 ( 'class C(object, B): pass', 'class C(B): pass', ), ( 'class C((object), B): pass', 'class C(B): pass', ), ( 'class C(( object ), B): pass', 'class C(B): pass', ), ( 'class C(\n' ' object,\n' ' B,\n' '): pass', 'class C(\n' ' B,\n' '): pass', ), ), ) def test_fix_new_style_classes(s, expected): assert _fix_new_style_classes(s) == expected @pytest.mark.parametrize( 's', ( # syntax error 'x = (', # weird attributes 'isinstance(s, six . string_types)', # weird space at beginning of decorator '@ six.python_2_unicode_compatible\n' 'class C: pass', # unrelated 'from os import path', 'from six import moves', # unrelated decorator '@mydec\n' 'class C: pass', # renaming things for weird reasons 'from six import StringIO as text_type\n' 'isinstance(s, text_type)\n', # weird spaces at begining of calls 'six.u ("bar")', 'from six import u\nu ("bar")', 'six.raise_from (exc, exc_from)', 'from six import raise_from\nraise_from (exc, exc_from)', # don't rewrite things that would become `raise` in non-statements 'print(six.raise_from(exc, exc_from))', ) ) def test_fix_six_noop(s): assert _fix_six(s) == s @pytest.mark.parametrize( ('s', 'expected'), ( ( 'isinstance(s, six.text_type)', 'isinstance(s, str)', ), ( 'isinstance(s, six.string_types)', 'isinstance(s, str)', ), ( 'issubclass(tp, six.string_types)', 'issubclass(tp, str)', ), ( 'STRING_TYPES = six.string_types', 'STRING_TYPES = (str,)', ), ( 'from six import string_types\n' 'isinstance(s, string_types)\n', 'from six import string_types\n' 'isinstance(s, str)\n', ), ( 'from six import string_types\n' 'STRING_TYPES = string_types\n', 'from six import string_types\n' 'STRING_TYPES = (str,)\n', ), ( '@six.python_2_unicode_compatible\n' 'class C: pass', 'class C: pass', ), ( '@six.python_2_unicode_compatible\n' '@other_decorator\n' 'class C: pass', '@other_decorator\n' 'class C: pass', ), ( 'from six import python_2_unicode_compatible\n' '@python_2_unicode_compatible\n' 'class C: pass', 'from six import python_2_unicode_compatible\n' 'class C: pass', ), ( 'six.get_unbound_method(meth)\n', 'meth\n', ), ( 'from six import get_unbound_method\n' 'get_unbound_method(meth)\n', 'from six import get_unbound_method\n' 'meth\n', ), ( 'six.indexbytes(bs, i)\n', 'bs[i]\n', ), ( 'six.assertCountEqual(\n' ' self,\n' ' arg1,\n' ' arg2,\n' ')', 'self.assertCountEqual(\n' ' arg1,\n' ' arg2,\n' ')', ), ( 'six.assertCountEqual(\n' ' self,\\\n' ' arg1,\n' ' arg2,\n' ')', 'self.assertCountEqual(\\\n' ' arg1,\n' ' arg2,\n' ')', ), ( 'six.assertCountEqual(\n' ' self, # hello\n' ' arg1,\n' ' arg2,\n' ')', 'self.assertCountEqual(\n' ' arg1,\n' ' arg2,\n' ')', ), ( 'six.assertCountEqual(\n' ' self,\n' ' arg1,\n' ' (1, 2, 3),\n' ')', 'self.assertCountEqual(\n' ' arg1,\n' ' (1, 2, 3),\n' ')', ), ( 'six.raise_from(exc, exc_from)\n', 'raise exc from exc_from\n', ), ( 'six.reraise(tp, exc, tb)\n', 'raise exc.with_traceback(tb)\n', ), ( 'from six import raise_from\n' 'raise_from(exc, exc_from)\n', 'from six import raise_from\n' 'raise exc from exc_from\n', ), ( 'six.reraise(\n' ' tp,\n' ' exc,\n' ' tb,\n' ')\n', 'raise exc.with_traceback(tb)\n', ), ), ) def test_fix_six(s, expected): assert _fix_six(s) == expected @pytest.mark.xfail(sys.version_info < (3,), reason='py3+ metaclass') @pytest.mark.parametrize( ('s', 'expected'), ( ( 'class C(object, metaclass=ABCMeta): pass', 'class C(metaclass=ABCMeta): pass', ), ), ) def test_fix_new_style_classes_py3only(s, expected): assert _fix_new_style_classes(s) == expected @pytest.mark.parametrize( 's', ( # syntax error '(', # weird syntax '"{}" . format(x)', # spans multiple lines '"{}".format(\n a,\n)', # starargs '"{} {}".format(*a)', '"{foo} {bar}".format(**b)"', # likely makes the format longer '"{0} {0}".format(arg)', '"{x} {x}".format(arg)', '"{x.y} {x.z}".format(arg)', # bytestrings don't participate in `.format()` or `f''` # but are legal in python 2 'b"{} {}".format(a, b)', # for now, too difficult to rewrite correctly '"{:{}}".format(x, y)', ), ) def test_fix_fstrings_noop(s): assert _fix_fstrings(s) == s @pytest.mark.parametrize( ('s', 'expected'), ( ('"{} {}".format(a, b)', 'f"{a} {b}"'), ('"{1} {0}".format(a, b)', 'f"{b} {a}"'), ('"{x.y}".format(x=z)', 'f"{z.y}"'), ('"{.x} {.y}".format(a, b)', 'f"{a.x} {b.y}"'), ('"{} {}".format(a.b, c.d)', 'f"{a.b} {c.d}"'), ('"hello {}!".format(name)', 'f"hello {name}!"'), ('"{}{{}}{}".format(escaped, y)', 'f"{escaped}{{}}{y}"'), # TODO: poor man's f-strings? # '"{foo}".format(**locals())' ), ) def test_fix_fstrings(s, expected): assert _fix_fstrings(s) == expected def test_main_trivial(): assert main(()) == 0 def test_main_noop(tmpdir): f = tmpdir.join('f.py') f.write('x = 5\n') assert main((f.strpath,)) == 0 assert f.read() == 'x = 5\n' def test_main_changes_a_file(tmpdir, capsys): f = tmpdir.join('f.py') f.write('x = set((1, 2, 3))\n') assert main((f.strpath,)) == 1 out, _ = capsys.readouterr() assert out == 'Rewriting {}\n'.format(f.strpath) assert f.read() == 'x = {1, 2, 3}\n' def test_main_keeps_line_endings(tmpdir, capsys): f = tmpdir.join('f.py') f.write_binary(b'x = set((1, 2, 3))\r\n') assert main((f.strpath,)) == 1 assert f.read_binary() == b'x = {1, 2, 3}\r\n' def test_main_syntax_error(tmpdir): f = tmpdir.join('f.py') f.write('from __future__ import print_function\nprint 1\n') assert main((f.strpath,)) == 0 def test_main_non_utf8_bytes(tmpdir, capsys): f = tmpdir.join('f.py') f.write_binary('# -*- coding: cp1252 -*-\nx = €\n'.encode('cp1252')) assert main((f.strpath,)) == 1 out, _ = capsys.readouterr() assert out == '{} is non-utf-8 (not supported)\n'.format(f.strpath) def test_keep_percent_format(tmpdir): f = tmpdir.join('f.py') f.write('"%s" % (1,)') assert main((f.strpath, '--keep-percent-format')) == 0 assert f.read() == '"%s" % (1,)' assert main((f.strpath,)) == 1 assert f.read() == '"{}".format(1)' def test_py3_plus_argument_unicode_literals(tmpdir): f = tmpdir.join('f.py') f.write('u""') assert main((f.strpath,)) == 0 assert f.read() == 'u""' assert main((f.strpath, '--py3-plus')) == 1 assert f.read() == '""' def test_py3_plus_super(tmpdir): f = tmpdir.join('f.py') f.write( 'class C(Base):\n' ' def f(self):\n' ' super(C, self).f()\n', ) assert main((f.strpath,)) == 0 assert f.read() == ( 'class C(Base):\n' ' def f(self):\n' ' super(C, self).f()\n' ) assert main((f.strpath, '--py3-plus')) == 1 assert f.read() == ( 'class C(Base):\n' ' def f(self):\n' ' super().f()\n' ) def test_py3_plus_new_style_classes(tmpdir): f = tmpdir.join('f.py') f.write('class C(object): pass\n') assert main((f.strpath,)) == 0 assert f.read() == 'class C(object): pass\n' assert main((f.strpath, '--py3-plus')) == 1 assert f.read() == 'class C: pass\n' def test_py36_plus_fstrings(tmpdir): f = tmpdir.join('f.py') f.write('"{} {}".format(hello, world)') assert main((f.strpath,)) == 0 assert f.read() == '"{} {}".format(hello, world)' assert main((f.strpath, '--py36-plus')) == 1 assert f.read() == 'f"{hello} {world}"'
34,139
12,279
file_berita = open("berita.txt", "r") berita = file_berita.read() berita = berita.split() berita = [x.lower() for x in berita] berita = list(set(berita)) berita = sorted(berita) print (berita)
194
79
# Copyright 2016-present CERN – European Organization for Nuclear Research # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import matplotlib.pyplot as plt import pandas as pd from qf_lib.common.enums.orientation import Orientation from qf_lib.plotting.charts.bar_chart import BarChart from qf_lib.plotting.decorators.data_element_decorator import DataElementDecorator from qf_lib.plotting.helpers.index_translator import IndexTranslator index = ['constant', 'b', 'c', 'd'] # index = [0, 4, 5, 6] labels_to_locations_dict = { 'constant': 0, 'b': 4, 'c': 5, 'd': 6 } colors = ['orange'] + ['forestgreen'] * 3 def main(): # using automatic mapping between labels and locations bar_chart2 = BarChart(orientation=Orientation.Horizontal, index_translator=IndexTranslator(), thickness=1.0, color=colors, align='center') bar_chart2.add_decorator(DataElementDecorator(pd.Series(data=[1, 2, 3, 4], index=index))) bar_chart2.add_decorator(DataElementDecorator(pd.Series(data=[3, 1, 2, 4], index=index))) bar_chart2.plot() # using custom mapping between labels and locations bar_chart = BarChart(orientation=Orientation.Horizontal, index_translator=IndexTranslator(labels_to_locations_dict), thickness=1.0, color=colors, align='center') bar_chart.add_decorator(DataElementDecorator(pd.Series(data=[1, 2, 3, 4], index=index))) bar_chart.add_decorator(DataElementDecorator(pd.Series(data=[3, 1, 2, 4], index=index))) bar_chart.plot() plt.show(block=True) if __name__ == '__main__': main()
2,139
703
import pytest from responses import RequestsMock from tests import loader def test_parameter_cannot_be_parsed(responses: RequestsMock, tmpdir): responses.add( responses.GET, url="http://test/", json={ "swagger": "2.0", "paths": { "/test": { "get": { "parameters": [ { "in": "query", "name": "param", "schema": {}, } ], "responses": { "200": { "description": "return value", "schema": {"type": "string"}, } }, } } }, }, match_querystring=True, ) with pytest.raises(Exception) as exception_info: loader.load( tmpdir, { "invalid": { "open_api": {"definition": "http://test/"}, "formulas": {"dynamic_array": {"lock_excel": True}}, } }, ) assert ( str(exception_info.value) == "Unable to extract parameters from {'in': 'query', 'name': 'param', 'schema': {}, 'server_param_name': 'param'}" ) def test_parameter_with_more_than_one_field_type(responses: RequestsMock, tmpdir): responses.add( responses.GET, url="http://test/", json={ "swagger": "2.0", "paths": { "/test": { "get": { "parameters": [ { "in": "query", "name": "param", "type": ["string", "integer"], } ], "responses": { "200": { "description": "return value", } }, } } }, }, match_querystring=True, ) with pytest.raises(Exception) as exception_info: loader.load( tmpdir, { "invalid": { "open_api": {"definition": "http://test/"}, "formulas": {"dynamic_array": {"lock_excel": True}}, } }, ) assert ( str(exception_info.value) == "Unable to guess field type amongst ['string', 'integer']" )
2,771
660
"""Package for YooKassa Payout API Python Client Library."""
61
17
import toml from delfino.constants import ENTRY_POINT, PYPROJECT_TOML_FILENAME from delfino.models.pyproject_toml import PyprojectToml from tests.constants import PROJECT_ROOT class TestEntrypointConstant: @staticmethod def should_match_entrypoint_in_pyproject_toml(): file_path = PROJECT_ROOT / PYPROJECT_TOML_FILENAME pyproject_toml = PyprojectToml(**toml.load(file_path)) assert ENTRY_POINT in pyproject_toml.tool.poetry.scripts
467
162
from random import randint from typing import Callable, List, Optional class Coin: """Simulates a coin.""" def __init__(self) -> None: self.__head = False self.__toss_count = 0 self.__head_count = 0 def toss(self) -> None: """Toss a coin.""" r = randint(1, 2) self.__head = True if r == 1 else False self.__toss_count += 1 if self.__head: self.__head_count += 1 def get_head_percentage(self) -> float: """Returns the percentages of heads relative to the total umber of coin tosses.""" return self.__head_count * 100 / self.__toss_count def is_head(self) -> bool: """Check if the coins shows heads.""" return self.__head def get_head_count(self) -> int: """Return the number of tossed heads.""" return self.__head_count def get_toss_count(self) -> int: """Return the number of tosses.""" return self.__toss_count class Player: def __init__(self, name: str, bet_amount: float) -> None: self.__name = name self.__amounts = [bet_amount] @property def name(self) -> str: """Name of the player.""" return self.__name @property def is_winner(self) -> bool: """ Check if the player is a winner. If the current amount of a player is greater or equal to the initial amount it is a winner. """ return self.__amounts[-1] >= self.__amounts[0] @property def is_total_loss(self) -> bool: """ Check if player lost everything. It is assumed that a total lost occurs if amount drops below 1% of the initial bet. """ return self.amount < self.amounts[0] / 100 @property def amount(self) -> float: """ Current amount of the player. """ return self.__amounts[-1] @property def amounts(self) -> List[float]: """ The amounts for all rounds of the player. The initial amount (bet) is stored at index 0. """ return self.__amounts def add_new_amount(self, amount: float) -> None: self.__amounts.append(amount) class RoundResults: def __init__(self, players: List[Player]) -> None: self.__total_amounts: List[float] = [] self.__number_of_winners: List[int] = [] self.__number_of_losers: List[int] = [] self.__number_of_total_losses: List[int] = [] self.__winner_percentages: List[float] = [] self.__min_amounts: List[float] = [] self.__max_amounts: List[float] = [] self.__avg_amounts: List[float] = [] self.add_round(players) def add_round(self, players: List[Player]) -> None: total_amount = 0 number_of_winners = 0 number_of_total_losses = 0 min_amount = players[0].amount max_amount = 0 for player in players: total_amount += player.amount min_amount = min(player.amount, min_amount) max_amount = max(player.amount, max_amount) if player.is_winner: number_of_winners += 1 if player.is_total_loss: number_of_total_losses += 1 winner_percentage = number_of_winners * 100 / len(players) self.__total_amounts.append(total_amount) self.__number_of_winners.append(number_of_winners) self.__number_of_losers.append(len(players) - number_of_winners) self.__number_of_total_losses.append(number_of_total_losses) self.__winner_percentages.append(winner_percentage) self.__min_amounts.append(min_amount) self.__max_amounts.append(max_amount) self.__avg_amounts.append(total_amount / len(players)) @property def number_of_rounds(self) -> int: return len(self.__total_amounts) - 1 @property def total_amounts(self) -> List[float]: return self.__total_amounts @property def avg_amounts(self) -> List[float]: return self.__avg_amounts @property def number_of_winners(self) -> List[int]: return self.__number_of_winners @property def number_of_losers(self) -> List[int]: return self.__number_of_losers @property def number_of_total_losses(self) -> List[int]: return self.__number_of_total_losses @property def winner_percentages(self) -> List[float]: return self.__winner_percentages @property def min_amounts(self) -> List[float]: return self.__min_amounts @property def max_amounts(self) -> List[float]: return self.__max_amounts class Gamble: def __init__( self, name: str, number_of_players: int, number_of_rounds: int, bet_amount: float, gain_percentage: int, loss_percentage: int, ) -> None: assert number_of_players > 0 assert number_of_rounds > 0 assert bet_amount > 0 assert gain_percentage >= 0 assert loss_percentage >= 0 and loss_percentage <= 100 self.__coin = Coin() self.__name: str = name self.__gain_factor: float = 1.0 + gain_percentage / 100.0 self.__loss_factor: float = 1.0 - loss_percentage / 100.0 self.__number_of_rounds: int = number_of_rounds self.__progress_callback: Optional[Callable[[str, int, int], None]] = None self.__players = [] for i in range(1, number_of_players + 1): self.__players.append(Player(name="p" + str(i), bet_amount=bet_amount)) self.__round_results = RoundResults(self.__players) def set_progress_callback(self, callback: Callable[[str, int, int], None]) -> None: self.__progress_callback = callback @property def name(self) -> str: return self.__name def _apply_rule(self, amount: float) -> float: self.__coin.toss() amount = ( amount * self.__gain_factor if self.__coin.is_head() else amount * self.__loss_factor ) return round(amount, 2) def _play_round(self, round_index: int) -> None: for player in self.__players: player.add_new_amount(self._apply_rule(player.amount)) return def play(self) -> None: for index in range(1, self.__number_of_rounds + 1): self._play_round(index) self.__round_results.add_round(self.__players) if self.__progress_callback: self.__progress_callback(self.name, index, self.__number_of_rounds) return @property def results(self) -> RoundResults: return self.__round_results @property def players(self) -> List[Player]: return self.__players @property def max_amount(self) -> float: return max(self.results.max_amounts)
6,888
2,142
#!/bin/python
16
8
""" @Fire https://github.com/fire717 """ cfg = { ##### Global Setting 'GPU_ID': '0', "num_workers":8, "random_seed":42, "cfg_verbose":True, "save_dir": "output/", "num_classes": 17, "width_mult":1.0, "img_size": 192, ##### Train Setting 'img_path':"./data/croped/imgs", 'train_label_path':'./data/croped/train2017.json', 'val_label_path':'./data/croped/val2017.json', 'balance_data':False, 'log_interval':10, 'save_best_only': True, 'pin_memory': True, ##### Train Hyperparameters 'learning_rate':0.001,#1.25e-4 'batch_size':64, 'epochs':120, 'optimizer':'Adam', #Adam SGD 'scheduler':'MultiStepLR-70,100-0.1', #default SGDR-5-2 CVPR step-4-0.8 MultiStepLR 'weight_decay' : 5.e-4,#0.0001, 'class_weight': None,#[1., 1., 1., 1., 1., 1., 1., ] 'clip_gradient': 5,#1, ##### Test 'test_img_path':"./data/croped/imgs", #"../data/eval/imgs", #"../data/eval/imgs", #"../data/all/imgs" #"../data/true/mypc/crop_upper1" #../data/coco/small_dataset/imgs #"../data/testimg" 'exam_label_path':'../data/all/data_all_new.json', 'eval_img_path':'../data/eval/imgs', 'eval_label_path':'../data/eval/mypc.json', }
1,285
593
from nmigen import * from nmigen.back.pysim import * # Function to calculate a clock divider which creates the # desired output frequency from a given input frequency. # Verifies that the divider is a positive integer, and that # the resulting signal doesn't deviate more than expected. def clock_divider( inf, outf, max_err ): # Calculate the divisor. div = inf // outf # Check that the rounded clock divider is a positive integer. if div < 1: print( "Error: Invalid input / output frequencies to " + "clock divider (%d / %d)" %( inf, outf ) ) raise ValueError # Check that the error value does not exceed the given threshold. err = ( ( inf / div ) - outf ) / outf if err > max_err: print( "Error: Clock divider error rate is too high " + "(%1.2f > %1.2f)" %( err, max_err ) ) raise ValueError # Return the divisor. return div # Basic work-in-progress UART modules. # - TX / RX only, no flow control or USART. # - Samples during the middle of the clock period, no oversampling. # - 8-bit words, 1 stop bit, no parity bit. # - Receives bits LSB-first only. # - Configurable baud rate. # UART receiver. class UART_RX( Elaboratable ): def __init__( self, clk_freq, baud_rate ): # Calculate the clock divider for the given frequency / baud rate. self.clk_div = clock_divider( clk_freq, baud_rate, 0.05 ) # Signals with external connections. self.rx = Signal() # Internal signals. self.rx_count = Signal( range( self.clk_div ) ) self.rx_buf = Signal( 8 ) self.rx_bit = Signal( 3 ) self.rx_ack = Signal() self.rx_fix = Signal() self.rx_strobe = Signal() def elaborate( self, platform ): m = Module() # Increment the RX counter each clock edge. with m.If( self.rx_count == self.clk_div - 1 ): m.d.sync += self.rx_count.eq( 0 ) with m.Else(): m.d.sync += self.rx_count.eq( self.rx_count + 1 ) # Toggle the "rx_strobe" value whenever "rx_count" ticks over. m.d.comb += self.rx_strobe.eq( self.rx_count == 0 ) # RX state machine. with m.FSM() as fsm: # RX state machine: "Idle" state. # - Move to the "Start" state when the "RX" line goes low. # - Set the "rx_count" value to half of the clock divider, so # that bits are sampled in the middle of each clock period. # - Reset the other peripheral signals. with m.State( "RX_IDLE" ): with m.If( ~self.rx ): m.next = "RX_START" m.d.sync += [ self.rx_ack.eq( 0 ), self.rx_fix.eq( 0 ), self.rx_bit.eq( 0 ), self.rx_buf.eq( 0x00 ), self.rx_count.eq( self.clk_div // 2 ) ] # RX state machine: "Start" state. # - Wait half a cycle until the clock divider ticks over, then # move to the "Data" state. with m.State( "RX_START" ): with m.If( self.rx_strobe ): m.next = "RX_DATA" # RX state machine: "Data" state. # - Wait one cycle, then prepend the RX bit and move to the # "Stop" state iff 8 bits have been received. with m.State( "RX_DATA" ): with m.If( self.rx_strobe ): m.d.sync += [ self.rx_buf.eq( Cat( self.rx_buf[ 1 : 8 ], self.rx ) ), self.rx_bit.eq( self.rx_bit + 1 ) ] with m.If( self.rx_bit == 7 ): m.next = "RX_STOP" # RX state machine: "Stop" state. # - Wait one cycle, then move to the "Full" state if the # "RX" line is asserted and to the "Error" state if not. with m.State( "RX_STOP" ): with m.If( self.rx_strobe ): with m.If( self.rx ): m.next = "RX_FULL" with m.Else(): m.next = "RX_ERROR" # RX state machine: "Full" state. # - Wait for an external "rx_ack" signal to indicate that the # value has been read, then move to the "Idle" state. # - Move to the "Error" state if a start bit is detected before # an "rx_ack" signal is received. with m.State( "RX_FULL" ): with m.If( self.rx_ack ): m.next = "RX_IDLE" with m.Elif( ~self.rx ): m.next = "RX_ERROR" # RX state machine: "Error" state. # - Wait for an external "rx_fix" signal, then return to "Idle". with m.State( "RX_ERROR" ): with m.If( self.rx_fix ): m.next = "RX_IDLE" return m # UART transmitter class UART_TX( Elaboratable ): def __init__( self, clk_freq, baud_rate ): # Calculate the clock divider for the given frequency / baud rate. self.clk_div = clock_divider( clk_freq, baud_rate, 0.05 ) # Signals with external connections. self.tx = Signal() # TX signals. self.tx_count = Signal( range( self.clk_div ) ) self.tx_buf = Signal( 8 ) self.tx_bit = Signal( 3 ) self.tx_new = Signal() self.tx_fix = Signal() self.tx_strobe = Signal() def elaborate( self, platform ): m = Module() # Increment the TX counter each clock edge. with m.If( self.tx_count == self.clk_div - 1 ): m.d.sync += self.tx_count.eq( 0 ) with m.Else(): m.d.sync += self.tx_count.eq( self.tx_count + 1 ) # Toggle the "tx_strobe" value whenever "tx_count" ticks over. m.d.comb += self.tx_strobe.eq( self.tx_count == 0 ) # TX state machine. with m.FSM() as fsm: # TX state machine: "Idle" state. # - Move to "Start" state when the "tx_new" signal goes high. # - Set "tx_count" divider to 1, and reset the other signals. # - Send a start bit by pulling "tx" low. with m.State( "TX_IDLE" ): with m.If( self.tx_new ): m.next = "TX_START" m.d.sync += [ self.tx_new.eq( 0 ), self.tx_fix.eq( 0 ), self.tx_bit.eq( 0 ), self.tx_count.eq( 1 ), self.tx.eq( 0 ) ] # TX state machine: "Start" state. # - Wait a cycle until the clock divider ticks over, then # move to the "Data" state. # - Move to the "Error" state if an unexpected "tx_new" occurs. with m.State( "TX_START" ): with m.If( self.tx_strobe ): m.next = "TX_DATA" with m.Elif( self.tx_new ): m.next = "TX_ERROR" # TX state machine: "Data" state. # - Set the "tx" value based on the current bit. # - Wait a cycle, then increment the bit number. # - Move to the "Stop" state iff 8 bits have been sent. # - Move to the "Error" state if an unexpected "tx_new" occurs. with m.State( "TX_DATA" ): m.d.sync += self.tx.eq( self.tx_buf.bit_select( self.tx_bit, 1 ) ) with m.If( self.tx_strobe ): m.d.sync += self.tx_bit.eq( self.tx_bit + 1 ) with m.If( self.tx_bit == 7 ): m.next = "TX_STOP" with m.Elif( self.tx_new ): m.next = "TX_ERROR" # TX state machine: "Stop" state. # - Pull the "tx" line high, clear the "tx_buf" buffer, # and then wait one cycle before moving. # to the "Idle" state. with m.State( "TX_STOP" ): m.d.sync += self.tx.eq( 1 ) m.d.sync += self.tx_buf.eq( 0x00 ) with m.If( self.tx_strobe ): m.next = "TX_IDLE" # TX state machine: "Error" state. with m.State( "TX_ERROR" ): with m.If( self.tx_fix ): m.next = "TX_IDLE" m.d.sync += self.tx.eq( 1 ) return m # Combined UART interface with both TX and RX modules. class UART( Elaboratable ): def __init__( self, uart_rx, uart_tx ): self.uart_rx = uart_rx self.uart_tx = uart_tx def elaborate( self, platform ): m = Module() m.submodules.rx = self.uart_rx m.submodules.tx = self.uart_tx return m # # Simple UART testbench. # # Helper UART test method to simulate receiving a byte. def uart_rx_byte( uart, val ): # Simulate a "start bit". yield uart.rx.eq( 0 ) # Wait one cycle. for i in range( uart.clk_div ): yield Tick() # Simulate the byte with one cycle between each bit. for i in range( 8 ): if val & ( 1 << i ): yield uart.rx.eq( 1 ) else: yield uart.rx.eq( 0 ) for j in range( uart.clk_div ): yield Tick() # Simulate the "stop bit", and wait one cycle. yield uart.rx.eq( 1 ) for i in range( uart.clk_div ): yield Tick() # Helper UART test method to simulate transmitting a buffered byte. def uart_tx_byte( uart ): # Send a "start bit". yield uart.tx.eq( 0 ) # Wait one cycle. for i in range( uart.clk_div ): yield Tick() # Send the byte with one cycle between each bit. for i in range( 8 ): yield uart.tx.eq( uart.tx_buf.bit_select( i, 1 ) ) for j in range( uart.clk_div ): yield Tick() # Send the "stop bit", and wait one cycle. yield uart.tx.eq( 1 ) for i in range( uart.clk_div ): yield Tick() # UART 'receive' testbench. def uart_rx_test( uart_rx ): # Simulate receiving "0xAF". yield from uart_rx_byte( uart_rx, 0xAF ) # Wait a couple of cycles, then send "rx_ack". for i in range( uart_rx.clk_div * 2 ): yield Tick() yield uart_rx.rx_ack.eq( 1 ) # Simulate receiving "0x42". yield from uart_rx_byte( uart_rx, 0x42 ) # Simulate receiving "0x24" (should cause an error) yield from uart_rx_byte( uart_rx, 0x24 ) # Wait a cycle, then send the "fix" signal and re-send 0x24. for i in range( uart_rx.clk_div ): yield Tick() yield uart_rx.rx_fix.eq( 1 ) yield from uart_rx_byte( uart_rx, 0x24 ) # Send "rx_ack" and wait a cycle to see the end state. yield uart_rx.rx_ack.eq( 1 ) for i in range( uart_rx.clk_div ): yield Tick() # UART 'transmit' testbench. def uart_tx_test( uart_tx ): # Set the "tx_buf" value to 0xAF, and set the "tx_new" signal. yield uart_tx.tx_buf.eq( 0xAF ) yield uart_tx.tx_new.eq( 1 ) # Simulate transmitting the byte. yield from uart_tx_byte( uart_tx ) # Wait a couple of cycles. for i in range( uart_tx.clk_div * 2 ): yield Tick() # Setup the TX module to send 0x42. yield uart_tx.tx_buf.eq( 0x42 ) yield uart_tx.tx_new.eq( 1 ) # Simulate sending the byte. yield from uart_tx_byte( uart_tx ) # Wait a cycle to see the end state. for i in range( uart_tx.clk_div ): yield Tick() # Create a UART module and run tests on it. # (The baud rate is set to a high value to speed up the simulation.) if __name__ == "__main__": #uart_rx = UART_RX( 24000000, 9600 ) #uart_tx = UART_TX( 24000000, 9600 ) uart_rx = UART_RX( 24000000, 1000000 ) uart_tx = UART_TX( 24000000, 1000000 ) uart = UART( uart_rx, uart_tx ) # Run the UART tests. with Simulator( uart, vcd_file = open( 'test.vcd', 'w' ) ) as sim: def proc_rx(): yield from uart_rx_test( uart.uart_rx ) def proc_tx(): yield from uart_tx_test( uart.uart_tx ) # Run the UART test with a 24MHz clock. sim.add_clock( 24e-6 ) sim.add_sync_process( proc_rx ) sim.add_sync_process( proc_tx ) sim.run()
11,038
4,224
## @file # This file is used to define the FMMT dependent external tool management class. # # Copyright (c) 2021-, Intel Corporation. All rights reserved.<BR> # SPDX-License-Identifier: BSD-2-Clause-Patent ## import glob import logging import os import shutil import sys import tempfile import uuid from edk2basetools.FMMT.PI.Common import * from edk2basetools.FMMT.utils.FmmtLogger import FmmtLogger as logger import subprocess def ExecuteCommand(cmd: list) -> None: subprocess.run(cmd,stdout=subprocess.DEVNULL) class GUIDTool: def __init__(self, guid: str, short_name: str, command: str) -> None: self.guid: str = guid self.short_name: str = short_name self.command: str = command def pack(self, buffer: bytes) -> bytes: """ compress file. """ tool = self.command if tool: tmp = tempfile.mkdtemp(dir=os.environ.get('tmp')) ToolInputFile = os.path.join(tmp, "pack_uncompress_sec_file") ToolOuputFile = os.path.join(tmp, "pack_sec_file") try: file = open(ToolInputFile, "wb") file.write(buffer) file.close() command = [tool, '-e', '-o', ToolOuputFile, ToolInputFile] ExecuteCommand(command) buf = open(ToolOuputFile, "rb") res_buffer = buf.read() except Exception as msg: logger.error(msg) return "" else: buf.close() if os.path.exists(tmp): shutil.rmtree(tmp) return res_buffer else: logger.error( "Error parsing section: EFI_SECTION_GUID_DEFINED cannot be parsed at this time.") logger.info("Its GUID is: %s" % self.guid) return "" def unpack(self, buffer: bytes) -> bytes: """ buffer: remove common header uncompress file """ tool = self.command if tool: tmp = tempfile.mkdtemp(dir=os.environ.get('tmp')) ToolInputFile = os.path.join(tmp, "unpack_sec_file") ToolOuputFile = os.path.join(tmp, "unpack_uncompress_sec_file") try: file = open(ToolInputFile, "wb") file.write(buffer) file.close() command = [tool, '-d', '-o', ToolOuputFile, ToolInputFile] ExecuteCommand(command) buf = open(ToolOuputFile, "rb") res_buffer = buf.read() except Exception as msg: logger.error(msg) return "" else: buf.close() if os.path.exists(tmp): shutil.rmtree(tmp) return res_buffer else: logger.error("Error parsing section: EFI_SECTION_GUID_DEFINED cannot be parsed at this time.") logger.info("Its GUID is: %s" % self.guid) return "" class GUIDTools: ''' GUIDTools is responsible for reading FMMTConfig.ini, verify the tools and provide interfaces to access those tools. ''' default_tools = { struct2stream(ModifyGuidFormat("a31280ad-481e-41b6-95e8-127f4c984779")): GUIDTool("a31280ad-481e-41b6-95e8-127f4c984779", "TIANO", "TianoCompress"), struct2stream(ModifyGuidFormat("ee4e5898-3914-4259-9d6e-dc7bd79403cf")): GUIDTool("ee4e5898-3914-4259-9d6e-dc7bd79403cf", "LZMA", "LzmaCompress"), struct2stream(ModifyGuidFormat("fc1bcdb0-7d31-49aa-936a-a4600d9dd083")): GUIDTool("fc1bcdb0-7d31-49aa-936a-a4600d9dd083", "CRC32", "GenCrc32"), struct2stream(ModifyGuidFormat("d42ae6bd-1352-4bfb-909a-ca72a6eae889")): GUIDTool("d42ae6bd-1352-4bfb-909a-ca72a6eae889", "LZMAF86", "LzmaF86Compress"), struct2stream(ModifyGuidFormat("3d532050-5cda-4fd0-879e-0f7f630d5afb")): GUIDTool("3d532050-5cda-4fd0-879e-0f7f630d5afb", "BROTLI", "BrotliCompress"), } def __init__(self, tooldef_file: str=None) -> None: self.dir = os.path.dirname(__file__) self.tooldef_file = tooldef_file if tooldef_file else os.path.join( self.dir, "FMMTConfig.ini") self.tooldef = dict() self.load() def VerifyTools(self) -> None: """ Verify Tools and Update Tools path. """ path_env = os.environ.get("PATH") path_env_list = path_env.split(os.pathsep) path_env_list.append(os.path.dirname(__file__)) path_env_list = list(set(path_env_list)) for tool in self.tooldef.values(): cmd = tool.command if os.path.isabs(cmd): if not os.path.exists(cmd): print("Tool Not found %s" % cmd) else: for syspath in path_env_list: if glob.glob(os.path.join(syspath, cmd+"*")): break else: print("Tool Not found %s" % cmd) def load(self) -> None: if os.path.exists(self.tooldef_file): with open(self.tooldef_file, "r") as fd: config_data = fd.readlines() for line in config_data: try: guid, short_name, command = line.split() new_format_guid = struct2stream(ModifyGuidFormat(guid.strip())) self.tooldef[new_format_guid] = GUIDTool( guid.strip(), short_name.strip(), command.strip()) except: print("GuidTool load error!") continue else: self.tooldef.update(self.default_tools) self.VerifyTools() def __getitem__(self, guid) -> None: return self.tooldef.get(guid) guidtools = GUIDTools()
5,853
1,957
from challenges.tree_intersection.tree_intersection import find_intersection from challenges.tree.tree import BinarySearchTree def test_find_intersection(): tree1 = BinarySearchTree() tree1.add(1) tree1.add(2) tree1.add(3) tree1.add(4) tree1.add(5) tree1.add(6) tree1.add(7) tree1.add(8) tree2 = BinarySearchTree() tree2.add(12) tree2.add(12) tree2.add(13) tree2.add(14) tree2.add(15) tree2.add(16) tree2.add(7) tree2.add(8) actual = find_intersection(tree1, tree2) expected = [7,8] assert actual == expected def test_empty_binary_tree(): tree1 = BinarySearchTree() tree1.add(1) tree1.add(2) tree1.add(3) tree1.add(4) tree1.add(5) tree1.add(6) tree1.add(7) tree1.add(8) tree2 = BinarySearchTree() actual = find_intersection(tree1, tree2) expected = [] assert actual == expected def test_first_empty_binary_tree(): tree2 = BinarySearchTree() tree2.add(1) tree2.add(2) tree2.add(3) tree2.add(4) tree2.add(5) tree2.add(6) tree2.add(7) tree2.add(8) tree1 = BinarySearchTree() actual = find_intersection(tree1, tree2) expected = [] assert actual == expected def test_same_tree(): tree1 = BinarySearchTree() tree1.add(1) tree1.add(2) tree1.add(3) tree1.add(4) tree1.add(5) tree1.add(6) tree1.add(7) tree1.add(8) tree2 = BinarySearchTree() tree2.add(1) tree2.add(2) tree2.add(3) tree2.add(4) tree2.add(5) tree2.add(6) tree2.add(7) tree2.add(8) actual = find_intersection(tree1, tree2) expected = [1,2,3,4,5,6,7,8] assert actual == expected
1,709
709
from setuptools import setup, find_packages setup( name = "imgdup", version = "1.3", packages = find_packages(), scripts = ['imgdup.py'], install_requires = ['pillow>=2.8.1'], # metadata for upload to PyPI author = "Radu Ioan Fericean", author_email = "radu@fericean.ro", description = "Visual similarity image finder and cleaner (image deduplication tool)", license = "MIT", keywords = "deduplication duplicate images image visual finder", url = "https://github.com/rif/imgdup", # project home page, if any )
562
176
k = float(input("Digite uma distância em quilometros: ")) m = k / 1.61 print("A distância digitada é de {} quilometros, essa distância convertida é {:.2f} milhas" .format(k,m))
178
73
from django.conf.urls.defaults import * urlpatterns = patterns('rcs.wiki.views', url(r'^((?:[A-Z]+[a-z]+){2,})/$', 'page', {}, name="wiki_page"), url(r'^((?:[A-Z]+[a-z]+){2,})/edit/$', 'edit', {}, name="wiki_edit"), url(r'^((?:[A-Z]+[a-z]+){2,})/attachments/$', 'attachments', {}, name="wiki_attachments"), url(r'^((?:[A-Z]+[a-z]+){2,})/rev/([a-f0-9]+)/$', 'revision', {}, name="wiki_revision"), url(r'^((?:[A-Z]+[a-z]+){2,})/diff/([\w]+)/([\w]+)/$', 'diff', {}, name="wiki_diff"), url(r'^list/$', 'list', {}, name="wiki_list"), url(r'^recent/$', 'recent', {}, name="wiki_recent"), url(r'^$', 'index', {}, name="wiki_index"), )
660
300
import cv2 as cv import numpy as np import pandas as pd import heapq def sort_contours_horizontally(contours): dic = {} for contour in contours: x_points = contour[:, :, 0] dic[min(x_points)[0]] = contour dic = dict(sorted(dic.items())) return dic.values() def sort_contours(contours): dic = {} error = 5 for contour in contours: x_points = contour[:, :, 0] y_points = contour[:, :, 1] key = min(x_points)[0] min_key = list(filter(lambda x: -error < x - key < error, dic.keys())) if min_key: key = min_key[0] dic[key].append((min(y_points)[0], contour)) else: dic[key] = [(min(y_points)[0], contour)] dic = dict(sorted(dic.items())) for key in dic.keys(): dic[key] = list(sorted(dic[key])) return [tup[1] for sublist in dic.values() for tup in sublist] def is_valid_test(tests): # multi answers => WRONG answers = list(filter(lambda x: x > 0.7, tests)) if len(answers) != 1: return False return True if __name__ == "__main__": img = cv.imread('sample1_2.jpg') cop = img.copy() img = cv.cvtColor(img, cv.COLOR_BGR2GRAY) img = cv.adaptiveThreshold(img, 255, cv.ADAPTIVE_THRESH_MEAN_C, cv.THRESH_BINARY_INV, 15, 10) kernel = cv.getStructuringElement(cv.MORPH_RECT, (4, 1)) img = cv.morphologyEx(img, cv.MORPH_CLOSE, kernel) contours, hierarchy = cv.findContours(img, cv.RETR_LIST, cv.CHAIN_APPROX_NONE) contours = list(filter(lambda x: 300 < cv.contourArea(x) < 450, contours)) contours = sort_contours(contours) answers = [] for i, contour in enumerate(contours): x, y, w, h = cv.boundingRect(contour) roi = cv.cvtColor(cop[y: y+h, x: x + w], cv.COLOR_BGR2GRAY) roi_cop = roi.copy() roi = cv.adaptiveThreshold(roi, 255, cv.ADAPTIVE_THRESH_MEAN_C, cv.THRESH_BINARY_INV, 15, 10) kernel = cv.getStructuringElement(cv.MORPH_ELLIPSE, (1, 3)) roi = cv.morphologyEx(roi, cv.MORPH_CLOSE, kernel) cnts, hierarchy = cv.findContours(roi, cv.RETR_EXTERNAL, cv.CHAIN_APPROX_NONE) cnts = list(filter(cv.contourArea, cnts)) cnts = sort_contours_horizontally(cnts) tests = list(map(cv.boundingRect, cnts)) coord = [(x, y)] for j, test in enumerate(tests): # each test is a contour coord.append(test) x, y, w, h = test area = w * h filled_area = np.count_nonzero(roi[y: y+h, x: x+w]) tests[j] = filled_area / area if is_valid_test(tests): choice = tests.index(max(tests)) + 1 answers.append(choice) X, Y = coord[0] x, y, w, h = coord[choice] pt1 = (X + x, Y + y) pt2 = (X + x + w, Y + y + h) cv.rectangle(cop, pt1, pt2, (0, 255, 0), 2) else: answers.append(-1) for i in range(len(answers)): print(i + 1, ":", answers[i]) data = { "Q": [i for i in range(1, len(answers) + 1)], "A": answers } data = pd.DataFrame(data) data.to_excel('./sample1.xlsx', 'Answer Sheet 1') cv.imwrite('output.jpg', cop) cv.imshow('Detected Choices', cop) cv.waitKey()
3,359
1,306
# 7. Write a program that estimates the average number of drawings it takes before the user’s # numbers are picked in a lottery that consists of correctly picking six different numbers that # are between 1 and 10. To do this, run a loop 1000 times that randomly generates a set of # user numbers and simulates drawings until the user’s numbers are drawn. Find the average # number of drawings needed over the 1000 times the loop runs. import random lottery_numbers = [i for i in range(1, 11)] avg = 0 for i in range(1000): user = random.randint(1, 10) lott = random.choice(lottery_numbers) if lott == user: avg = avg + 1 print('Average number of drawings:', round(1000 / avg, 4))
704
228
from django.conf import settings from django.utils.importlib import import_module from humfrey.update.transform.base import Transform def get_transforms(): try: return get_transforms._cache except AttributeError: pass transforms = {'__builtins__': {}} for class_path in settings.UPDATE_TRANSFORMS: module_path, class_name = class_path.rsplit('.', 1) transform = getattr(import_module(module_path), class_name) assert issubclass(transform, Transform) transforms[transform.__name__] = transform get_transforms._cache = transforms return transforms def evaluate_pipeline(pipeline): return eval('(%s)' % pipeline, get_transforms())
709
203
from .db_base import db, env class Device(db.Model): __tablename__ = 'device' id = db.Column(db.Integer, primary_key=True) if env == 'development': model = db.Column(db.String(16, 'utf8mb4_unicode_ci')) serial = db.Column(db.String(20, 'utf8mb4_unicode_ci')) company_id = db.Column(db.ForeignKey('company.id', onupdate='CASCADE'), index=True) owner_id = db.Column(db.ForeignKey('user.id', onupdate='CASCADE'), index=True) ip = db.Column(db.String(15, 'utf8mb4_unicode_ci'), server_default=db.FetchedValue()) else: model = db.Column(db.String(16, 'utf8mb4_unicode_ci'), nullable=False) serial = db.Column(db.String(20, 'utf8mb4_unicode_ci'), nullable=False, unique=True) company_id = db.Column(db.ForeignKey('company.id', onupdate='CASCADE'), nullable=False, index=True) owner_id = db.Column(db.ForeignKey('user.id', onupdate='CASCADE'), nullable=False, index=True) ip = db.Column(db.String(15, 'utf8mb4_unicode_ci'), nullable=False, unique=True, server_default=db.FetchedValue()) created = db.Column(db.DateTime) created_by_id = db.Column(db.ForeignKey('user.id', onupdate='CASCADE'), index=True) last_edited = db.Column(db.DateTime) edited_by_id = db.Column(db.ForeignKey('user.id', onupdate='CASCADE'), index=True) is_deleted = db.Column(db.Integer) company = db.relationship('Company', primaryjoin='Device.company_id == Company.id', backref='devices') created_by = db.relationship('User', primaryjoin='Device.created_by_id == User.id', backref='created_devices') edited_by = db.relationship('User', primaryjoin='Device.edited_by_id == User.id', backref='edited_devices') owner = db.relationship('User', primaryjoin='Device.owner_id == User.id', backref='owned_devices') def __repr__(self): return f'<Device {self.model} | {self.serial}>' def to_dict(self): return dict( id=self.id, model=self.model, serial=self.serial, company=self.company, owner=self.owner, ip=self.ip, created=self.created, created_by=self.created_by, last_edited=self.last_edited, edited_by=self.edited_by, is_deleted=self.is_deleted )
2,448
834
from lib.remote.remote_util import RemoteMachineShellConnection from pytests.tuqquery.tuq import QueryTests class TokenTests(QueryTests): def setUp(self): if not self._testMethodName == 'suite_setUp': self.skip_buckets_handle = True super(TokenTests, self).setUp() self.n1ql_port = self.input.param("n1ql_port", 8093) self.scan_consistency = self.input.param("scan_consistency", 'REQUEST_PLUS') def tearDown(self): server = self.master shell = RemoteMachineShellConnection(server) # shell.execute_command("""curl -X DELETE -u Administrator:password http://{0}:8091/pools/default/buckets/beer-sample""".format(server.ip)) self.sleep(20) super(TokenTests, self).tearDown() def test_tokens_secondary_indexes(self): self.rest.load_sample("beer-sample") self.sleep(20) created_indexes = [] self.query = 'create primary index on `beer-sample`' self.run_cbq_query() self.query = 'create index idx1 on `beer-sample`(description,name )' self.run_cbq_query() self.query = 'create index idx2 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description) END ,description,name )' self.run_cbq_query() self.query = 'create index idx3 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"lower","names":true,"specials":false}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx4 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"upper","names":false,"specials":true}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx5 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"upper","names":false}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx6 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"upper"}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx7 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx8 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"":""}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx9 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"random"}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx10 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"names":"random"}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx11 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"specials":"random"}) END ,description,name )' self.run_cbq_query() self.query = 'create index idx12 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description) END )' self.run_cbq_query() self.query = 'create index idx13 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"lower"}) END )' self.run_cbq_query() self.query = 'create index idx14 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"upper"}) END )' self.run_cbq_query() self.query = 'create index idx15 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"lower","names":true,"specials":false}) END )' self.run_cbq_query() self.query = 'create index idx16 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"upper","names":false,"specials":true}) END )' self.run_cbq_query() self.query = 'create index idx17 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"upper","names":false}) END )' self.run_cbq_query() self.query = 'create index idx18 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{}) END )' self.run_cbq_query() self.query = 'create index idx19 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"":""}) END )' self.run_cbq_query() self.query = 'create index idx20 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"case":"random"}) END )' self.run_cbq_query() self.query = 'create index idx21 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"names":"random"}) END )' self.run_cbq_query() self.query = 'create index idx22 on `beer-sample`( DISTINCT ARRAY v FOR v in tokens(description,{"specials":"random"}) END )' self.run_cbq_query() for i in xrange(1,22): index = 'idx{0}'.format(i) created_indexes.append(index) self.query = 'explain select name from `beer-sample` where any v in tokens(description) satisfies v = "golden" END limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue(actual_result['results']) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx2") self.assertTrue(str(plan['~children'][0]['~children'][0]['scan']['covers'][0]) == ('cover ((distinct (array `v` for `v` in tokens((`beer-sample`.`description`)) end)))')) self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(reverse(description)) satisfies v = "nedlog" END order by meta().id limit 10' expected_result = self.run_cbq_query() self.query = 'select name from `beer-sample` where any v in tokens(reverse(description)) satisfies v = "nedlog" END order by meta().id limit 10' actual_result = self.run_cbq_query() #self.assertTrue(str(actual_result['results'])=="[{u'name': u'21A IPA'}, {u'name': u'Amendment Pale Ale'}, {u'name': u'Double Trouble IPA'}, {u'name': u'South Park Blonde'}, {u'name': u'Restoration Pale Ale'}, {u'name': u'S.O.S'}, {u'name': u'Satsuma Harvest Wit'}, {u'name': u'Adnams Explorer'}, {u'name': u'Shock Top'}, {u'name': u'Anniversary Maibock'}]" ) self.assertTrue((actual_result['results'])== (expected_result['results'])) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{"case":"lower","names":true,"specials":false}) satisfies v = "brewery" END limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(str(plan['~children'][0]['~children'][0]['scan']['covers'][0]) == ('cover ((distinct (array `v` for `v` in tokens((`beer-sample`.`description`), {"case": "lower", "names": true, "specials": false}) end)))')) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx3") self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{"case":"lower","names":true,"specials":false}) satisfies v = "brewery" END order by meta().id limit 10' expected_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`idx15`) where any v in tokens(description,{"case":"lower","names":true,"specials":false}) satisfies v = "brewery" END order by meta().id limit 10' actual_result = self.run_cbq_query() self.assertTrue((actual_result['results'])== (expected_result['results']) ) self.query = 'explain select name from `beer-sample` use index(`idx14`) where any v in tokens(description,{"case":"upper","names":false,"specials":true}) satisfies v = "BREWERY" END order by meta().id limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(str(plan['~children'][0]['~children'][0]['scan']['covers'][0]) == ('cover ((distinct (array `v` for `v` in tokens((`beer-sample`.`description`), {"case": "upper", "names": false, "specials": true}) end)))')) self.assertTrue(str(plan['~children'][0]['~children'][0]['scan']['index']) == "idx4") self.query = 'select name from `beer-sample` use index(`idx16`) where any v in tokens(description,{"case":"upper","names":false,"specials":true}) satisfies v = "BREWERY" END order by meta().id limit 10' actual_result = self.run_cbq_query() self.assertTrue((actual_result['results'])== (expected_result['results'])) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{"case":"upper","names":false}) satisfies v = "GOLDEN" END limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx5") self.query = 'select name from `beer-sample` use index(`idx17`) where any v in tokens(description,{"case":"upper","names":false}) satisfies v = "GOLDEN" END limit 10' actual_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{"case":"upper","names":false}) satisfies v = "GOLDEN" END limit 10' expected_result = self.run_cbq_query() self.assertTrue(actual_result['results']==expected_result['results']) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{}) satisfies v = "golden" END limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx7") self.query = 'select name from `beer-sample` use index(`idx18`) where any v in tokens(description,{}) satisfies v = "golden" END limit 10' actual_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{}) satisfies v = "golden" END limit 10' expected_result = self.run_cbq_query() self.assertTrue(actual_result['results']==expected_result['results']) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{"":""}) satisfies v = "golden" END limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx8") self.query = 'select name from `beer-sample` use index(`idx19`) where any v in tokens(description,{"":""}) satisfies v = "golden" END order by name ' actual_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{"":""}) satisfies v = "golden" END order by name ' expected_result = self.run_cbq_query() self.assertTrue(actual_result['results']==expected_result['results']) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{"case":"random"}) satisfies v = "golden" END ' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['scan']['index'] == "idx9") self.query = 'select name from `beer-sample` use index(`idx20`) where any v in tokens(description,{"case":"random"}) satisfies v = "golden" END order by name ' actual_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{"case":"random"}) satisfies v = "golden" END order by name ' expected_result = self.run_cbq_query() self.assertTrue(actual_result['results']==expected_result['results']) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{"specials":"random"}) satisfies v = "brewery" END order by name' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx11") self.query = 'select name from `beer-sample` use index(`idx22`) where any v in tokens(description,{"specials":"random"}) satisfies v = "golden" END order by name' actual_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{"specials":"random"}) satisfies v = "golden" END order by name' expected_result = self.run_cbq_query() self.assertTrue(actual_result['results']==expected_result['results']) self.query = 'explain select name from `beer-sample` where any v in tokens(description,{"names":"random"}) satisfies v = "brewery" END limit 10' actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("covers" in str(plan)) self.assertTrue(plan['~children'][0]['~children'][0]['scan']['index'] == "idx10") self.query = 'select name from `beer-sample` use index(`idx21`) where any v in tokens(description,{"names":"random"}) satisfies v = "golden" END limit 10' actual_result = self.run_cbq_query() self.query = 'select name from `beer-sample` use index(`#primary`) where any v in tokens(description,{"names":"random"}) satisfies v = "golden" END limit 10' expected_result = self.run_cbq_query() self.assertTrue(actual_result['results']==expected_result['results']) for idx in created_indexes: self.query = "DROP INDEX %s.%s USING %s" % ("`beer-sample`", idx, self.index_type) actual_result = self.run_cbq_query() '''This test is specific to beer-sample bucket''' def test_tokens_simple_syntax(self): self.rest.load_sample("beer-sample") bucket_doc_map = {"beer-sample": 7303} bucket_status_map = {"beer-sample": "healthy"} self.wait_for_buckets_status(bucket_status_map, 5, 120) self.wait_for_bucket_docs(bucket_doc_map, 5, 120) self._wait_for_index_online("beer-sample", "beer_primary") self.sleep(10) created_indexes = [] try: idx1 = "idx_suffixes" idx2 = "idx_tokens" idx3 = "idx_pairs" idx4 = "idx_addresses" self.query = 'CREATE INDEX {0} ON `beer-sample`( DISTINCT SUFFIXES( name ) )'.format(idx1) self.run_cbq_query() self._wait_for_index_online("beer-sample", "beer_primary") created_indexes.append(idx1) self.query = "explain select * from `beer-sample` where name like '%Cafe%'" actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertEqual(plan['~children'][0]['scan']['index'], idx1) self.query = 'CREATE INDEX {0} ON `beer-sample`( DISTINCT TOKENS( description ) )'.format(idx2) self.run_cbq_query() self._wait_for_index_online("beer-sample", "beer_primary") created_indexes.append(idx2) self.query = "explain select * from `beer-sample` where contains_token(description,'Great')" actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertEqual(plan['~children'][0]['scan']['index'], idx2) self.query = "CREATE INDEX {0} ON `beer-sample`( DISTINCT PAIRS( SELF ) )".format(idx3) self.run_cbq_query() self._wait_for_index_online("beer-sample", "beer_primary") created_indexes.append(idx3) self.query = "explain select * from `beer-sample` where name like 'A%' and abv > 6" actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue("idx_suffixes" in str(plan)) self.query = "CREATE INDEX {0} ON `beer-sample`( ALL address )".format(idx4) self.run_cbq_query() self._wait_for_index_online("beer-sample", "beer_primary") created_indexes.append(idx4) self.query = "explain select min(addr) from `beer-sample` unnest address as addr" actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertEqual(plan['~children'][0]['index'], idx4) self.query = "explain select count(a) from `beer-sample` unnest address as a" actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertEqual(plan['~children'][0]['index'], idx4) self.query = "explain select * from `beer-sample` where any place in address satisfies " \ "place LIKE '100 %' end" actual_result = self.run_cbq_query() plan = self.ExplainPlanHelper(actual_result) self.assertTrue(idx4 in str(plan)) self.assertTrue(idx3 in str(plan)) finally: for idx in created_indexes: self.query = "DROP INDEX `beer-sample`.%s" % (idx) self.run_cbq_query() self.rest.delete_bucket("beer-sample") def test_dynamicindex_limit(self): self.rest.load_sample("beer-sample") self.sleep(20) created_indexes = [] try: idx1 = "idx_abv" idx2 = "dynamic" self.query = "CREATE INDEX idx_abv ON `beer-sample`( abv )" self.run_cbq_query() created_indexes.append(idx1) self.query = "CREATE INDEX dynamic ON `beer-sample`( DISTINCT PAIRS( SELF ) )" self.run_cbq_query() created_indexes.append(idx2) self.query = "Explain select * from `beer-sample` where abv > 5 LIMIT 10" res = self.run_cbq_query() plan = self.ExplainPlanHelper(res) self.assertTrue(plan['~children'][0]['~children'][0]['limit']=='10') finally: for idx in created_indexes: self.query = "DROP INDEX `beer-sample`.%s" % ( idx) self.run_cbq_query()
18,542
5,920
# AutoTransform # Large scale, component based code modification library # # Licensed under the MIT License <http://opensource.org/licenses/MIT> # SPDX-License-Identifier: MIT # Copyright (c) 2022-present Nathan Rockenbach <http://github.com/nathro> # @black_format """The implementation for the DirectoryInput.""" from __future__ import annotations from typing import ClassVar, Sequence from autotransform.input.base import Input, InputName from autotransform.item.base import Item class EmptyInput(Input): """An Input that simply returns an empty list. Used when a Transformer operates on the whole codebase, rather than on an individual Item/set of Items. Attributes: name (ClassVar[InputName]): The name of the component. """ name: ClassVar[InputName] = InputName.EMPTY def get_items(self) -> Sequence[Item]: """Returns an empty list of Items, useful for Transformers that operate on the whole codebase at once. Returns: Sequence[Item]: An empty list of Items. """ return []
1,074
311
""" tracker.py Copyright (c) 2013-2014 Snowplow Analytics Ltd. All rights reserved. This program is licensed to you under the Apache License Version 2.0, and you may not use this file except in compliance with the Apache License Version 2.0. You may obtain a copy of the Apache License Version 2.0 at http://www.apache.org/licenses/LICENSE-2.0. Unless required by applicable law or agreed to in writing, software distributed under the Apache License Version 2.0 is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the Apache License Version 2.0 for the specific language governing permissions and limitations there under. Authors: Anuj More, Alex Dean, Fred Blundun Copyright: Copyright (c) 2013-2014 Snowplow Analytics Ltd License: Apache License Version 2.0 """ import time import uuid import six from contracts import contract, new_contract from snowplow_tracker import payload, _version, SelfDescribingJson from snowplow_tracker import subject as _subject from snowplow_tracker.timestamp import Timestamp, TrueTimestamp, DeviceTimestamp """ Constants & config """ VERSION = "py-%s" % _version.__version__ DEFAULT_ENCODE_BASE64 = True BASE_SCHEMA_PATH = "iglu:com.snowplowanalytics.snowplow" SCHEMA_TAG = "jsonschema" CONTEXT_SCHEMA = "%s/contexts/%s/1-0-1" % (BASE_SCHEMA_PATH, SCHEMA_TAG) UNSTRUCT_EVENT_SCHEMA = "%s/unstruct_event/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG) FORM_NODE_NAMES = ("INPUT", "TEXTAREA", "SELECT") FORM_TYPES = ( "button", "checkbox", "color", "date", "datetime", "datetime-local", "email", "file", "hidden", "image", "month", "number", "password", "radio", "range", "reset", "search", "submit", "tel", "text", "time", "url", "week" ) """ Tracker class """ class Tracker: new_contract("not_none", lambda s: s is not None) new_contract("non_empty_string", lambda s: isinstance(s, six.string_types) and len(s) > 0) new_contract("string_or_none", lambda s: (isinstance(s, six.string_types) and len(s) > 0) or s is None) new_contract("payload", lambda s: isinstance(s, payload.Payload)) new_contract("tracker", lambda s: isinstance(s, Tracker)) new_contract("emitter", lambda s: hasattr(s, "input")) new_contract("self_describing_json", lambda s: isinstance(s, SelfDescribingJson)) new_contract("context_array", "list(self_describing_json)") new_contract("form_node_name", lambda s: s in FORM_NODE_NAMES) new_contract("form_type", lambda s: s.lower() in FORM_TYPES) new_contract("timestamp", lambda x: (isinstance(x, Timestamp))) new_contract("form_element", lambda x: Tracker.check_form_element(x)) @contract def __init__(self, emitters, subject=None, namespace=None, app_id=None, encode_base64=DEFAULT_ENCODE_BASE64): """ :param emitters: Emitters to which events will be sent :type emitters: list[>0](emitter) | emitter :param subject: Subject to be tracked :type subject: subject | None :param namespace: Identifier for the Tracker instance :type namespace: string_or_none :param app_id: Application ID :type app_id: string_or_none :param encode_base64: Whether JSONs in the payload should be base-64 encoded :type encode_base64: bool """ if subject is None: subject = _subject.Subject() if type(emitters) is list: self.emitters = emitters else: self.emitters = [emitters] self.subject = subject self.encode_base64 = encode_base64 self.standard_nv_pairs = { "tv": VERSION, "tna": namespace, "aid": app_id } self.timer = None @staticmethod @contract def get_uuid(): """ Set transaction ID for the payload once during the lifetime of the event. :rtype: string """ return str(uuid.uuid4()) @staticmethod @contract def get_timestamp(tstamp=None): """ :param tstamp: User-input timestamp or None :type tstamp: int | float | None :rtype: int """ if tstamp is None: return int(time.time() * 1000) elif isinstance(tstamp, (int, float, )): return int(tstamp) """ Tracking methods """ @contract def track(self, pb): """ Send the payload to a emitter :param pb: Payload builder :type pb: payload :rtype: tracker """ for emitter in self.emitters: emitter.input(pb.nv_pairs) return self @contract def complete_payload(self, pb, context, tstamp): """ Called by all tracking events to add the standard name-value pairs to the Payload object irrespective of the tracked event. :param pb: Payload builder :type pb: payload :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ pb.add("eid", Tracker.get_uuid()) if isinstance(tstamp, TrueTimestamp): pb.add("ttm", tstamp.value) if isinstance(tstamp, DeviceTimestamp): pb.add("dtm", Tracker.get_timestamp(tstamp.value)) elif isinstance(tstamp, (int, float, type(None))): pb.add("dtm", Tracker.get_timestamp(tstamp)) if context is not None: context_jsons = list(map(lambda c: c.to_json(), context)) context_envelope = SelfDescribingJson(CONTEXT_SCHEMA, context_jsons).to_json() pb.add_json(context_envelope, self.encode_base64, "cx", "co") pb.add_dict(self.standard_nv_pairs) pb.add_dict(self.subject.standard_nv_pairs) return self.track(pb) @contract def track_page_view(self, page_url, page_title=None, referrer=None, context=None, tstamp=None): """ :param page_url: URL of the viewed page :type page_url: non_empty_string :param page_title: Title of the viewed page :type page_title: string_or_none :param referrer: Referrer of the page :type referrer: string_or_none :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ pb = payload.Payload() pb.add("e", "pv") # pv: page view pb.add("url", page_url) pb.add("page", page_title) pb.add("refr", referrer) return self.complete_payload(pb, context, tstamp) @contract def track_page_ping(self, page_url, page_title=None, referrer=None, min_x=None, max_x=None, min_y=None, max_y=None, context=None, tstamp=None): """ :param page_url: URL of the viewed page :type page_url: non_empty_string :param page_title: Title of the viewed page :type page_title: string_or_none :param referrer: Referrer of the page :type referrer: string_or_none :param min_x: Minimum page x offset seen in the last ping period :type min_x: int | None :param max_x: Maximum page x offset seen in the last ping period :type max_x: int | None :param min_y: Minimum page y offset seen in the last ping period :type min_y: int | None :param max_y: Maximum page y offset seen in the last ping period :type max_y: int | None :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ pb = payload.Payload() pb.add("e", "pp") # pp: page ping pb.add("url", page_url) pb.add("page", page_title) pb.add("refr", referrer) pb.add("pp_mix", min_x) pb.add("pp_max", max_x) pb.add("pp_miy", min_y) pb.add("pp_may", max_y) return self.complete_payload(pb, context, tstamp) @contract def track_link_click(self, target_url, element_id=None, element_classes=None, element_target=None, element_content=None, context=None, tstamp=None): """ :param target_url: Target URL of the link :type target_url: non_empty_string :param element_id: ID attribute of the HTML element :type element_id: string_or_none :param element_classes: Classes of the HTML element :type element_classes: list(str) | tuple(str,*) | None :param element_content: The content of the HTML element :type element_content: string_or_none :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ properties = {} properties["targetUrl"] = target_url if element_id is not None: properties["elementId"] = element_id if element_classes is not None: properties["elementClasses"] = element_classes if element_target is not None: properties["elementTarget"] = element_target if element_content is not None: properties["elementContent"] = element_content event_json = SelfDescribingJson("%s/link_click/%s/1-0-1" % (BASE_SCHEMA_PATH, SCHEMA_TAG), properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_add_to_cart(self, sku, quantity, name=None, category=None, unit_price=None, currency=None, context=None, tstamp=None): """ :param sku: Item SKU or ID :type sku: non_empty_string :param quantity: Number added to cart :type quantity: int :param name: Item's name :type name: string_or_none :param category: Item's category :type category: string_or_none :param unit_price: Item's price :type unit_price: int | float | None :param currency: Type of currency the price is in :type currency: string_or_none :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ properties = {} properties["sku"] = sku properties["quantity"] = quantity if name is not None: properties["name"] = name if category is not None: properties["category"] = category if unit_price is not None: properties["unitPrice"] = unit_price if currency is not None: properties["currency"] = currency event_json = SelfDescribingJson("%s/add_to_cart/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG), properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_remove_from_cart(self, sku, quantity, name=None, category=None, unit_price=None, currency=None, context=None, tstamp=None): """ :param sku: Item SKU or ID :type sku: non_empty_string :param quantity: Number added to cart :type quantity: int :param name: Item's name :type name: string_or_none :param category: Item's category :type category: string_or_none :param unit_price: Item's price :type unit_price: int | float | None :param currency: Type of currency the price is in :type currency: string_or_none :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ properties = {} properties["sku"] = sku properties["quantity"] = quantity if name is not None: properties["name"] = name if category is not None: properties["category"] = category if unit_price is not None: properties["unitPrice"] = unit_price if currency is not None: properties["currency"] = currency event_json = SelfDescribingJson("%s/remove_from_cart/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG), properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_form_change(self, form_id, element_id, node_name, value, type_=None, element_classes=None, context=None, tstamp=None): """ :param form_id: ID attribute of the HTML form :type form_id: non_empty_string :param element_id: ID attribute of the HTML element :type element_id: string_or_none :param node_name: Type of input element :type node_name: form_node_name :param value: Value of the input element :type value: string_or_none :param type_: Type of data the element represents :type type_: non_empty_string, form_type :param element_classes: Classes of the HTML element :type element_classes: list(str) | tuple(str,*) | None :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ properties = dict() properties["formId"] = form_id properties["elementId"] = element_id properties["nodeName"] = node_name properties["value"] = value if type_ is not None: properties["type"] = type_ if element_classes is not None: properties["elementClasses"] = element_classes event_json = SelfDescribingJson("%s/change_form/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG), properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_form_submit(self, form_id, form_classes=None, elements=None, context=None, tstamp=None): """ :param form_id: ID attribute of the HTML form :type form_id: non_empty_string :param form_classes: Classes of the HTML form :type form_classes: list(str) | tuple(str,*) | None :param elements: Classes of the HTML form :type elements: list(form_element) | None :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ properties = dict() properties['formId'] = form_id if form_classes is not None: properties['formClasses'] = form_classes if elements is not None and len(elements) > 0: properties['elements'] = elements event_json = SelfDescribingJson("%s/submit_form/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG), properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_site_search(self, terms, filters=None, total_results=None, page_results=None, context=None, tstamp=None): """ :param terms: Search terms :type terms: seq[>=1](str) :param filters: Filters applied to the search :type filters: dict(str:str|bool) | None :param total_results: Total number of results returned :type total_results: int | None :param page_results: Total number of pages of results :type page_results: int | None :param context: Custom context for the event :type context: context_array | None :param tstamp: Optional user-provided timestamp for the event :type tstamp: timestamp | int | float | None :rtype: tracker """ properties = {} properties["terms"] = terms if filters is not None: properties["filters"] = filters if total_results is not None: properties["totalResults"] = total_results if page_results is not None: properties["pageResults"] = page_results event_json = SelfDescribingJson("%s/site_search/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG), properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_ecommerce_transaction_item(self, order_id, sku, price, quantity, name=None, category=None, currency=None, context=None, tstamp=None): """ This is an internal method called by track_ecommerce_transaction. It is not for public use. :param order_id: Order ID :type order_id: non_empty_string :param sku: Item SKU :type sku: non_empty_string :param price: Item price :type price: int | float :param quantity: Item quantity :type quantity: int :param name: Item name :type name: string_or_none :param category: Item category :type category: string_or_none :param currency: The currency the price is expressed in :type currency: string_or_none :param context: Custom context for the event :type context: context_array | None :rtype: tracker """ pb = payload.Payload() pb.add("e", "ti") pb.add("ti_id", order_id) pb.add("ti_sk", sku) pb.add("ti_nm", name) pb.add("ti_ca", category) pb.add("ti_pr", price) pb.add("ti_qu", quantity) pb.add("ti_cu", currency) return self.complete_payload(pb, context, tstamp) @contract def track_ecommerce_transaction(self, order_id, total_value, affiliation=None, tax_value=None, shipping=None, city=None, state=None, country=None, currency=None, items=None, context=None, tstamp=None): """ :param order_id: ID of the eCommerce transaction :type order_id: non_empty_string :param total_value: Total transaction value :type total_value: int | float :param affiliation: Transaction affiliation :type affiliation: string_or_none :param tax_value: Transaction tax value :type tax_value: int | float | None :param shipping: Delivery cost charged :type shipping: int | float | None :param city: Delivery address city :type city: string_or_none :param state: Delivery address state :type state: string_or_none :param country: Delivery address country :type country: string_or_none :param currency: The currency the price is expressed in :type currency: string_or_none :param items: The items in the transaction :type items: list(dict(str:*)) :param context: Custom context for the event :type context: context_array | None :rtype: tracker """ pb = payload.Payload() pb.add("e", "tr") pb.add("tr_id", order_id) pb.add("tr_tt", total_value) pb.add("tr_af", affiliation) pb.add("tr_tx", tax_value) pb.add("tr_sh", shipping) pb.add("tr_ci", city) pb.add("tr_st", state) pb.add("tr_co", country) pb.add("tr_cu", currency) tstamp = Tracker.get_timestamp(tstamp) self.complete_payload(pb, context, tstamp) for item in items: item["tstamp"] = tstamp item["order_id"] = order_id item["currency"] = currency self.track_ecommerce_transaction_item(**item) return self @contract def track_screen_view(self, name=None, id_=None, context=None, tstamp=None): """ :param name: The name of the screen view event :type name: string_or_none :param id_: Screen view ID :type id_: string_or_none :param context: Custom context for the event :type context: context_array | None :rtype: tracker """ screen_view_properties = {} if name is not None: screen_view_properties["name"] = name if id_ is not None: screen_view_properties["id"] = id_ event_json = SelfDescribingJson("%s/screen_view/%s/1-0-0" % (BASE_SCHEMA_PATH, SCHEMA_TAG), screen_view_properties) return self.track_unstruct_event(event_json, context, tstamp) @contract def track_struct_event(self, category, action, label=None, property_=None, value=None, context=None, tstamp=None): """ :param category: Category of the event :type category: non_empty_string :param action: The event itself :type action: non_empty_string :param label: Refer to the object the action is performed on :type label: string_or_none :param property_: Property associated with either the action or the object :type property_: string_or_none :param value: A value associated with the user action :type value: int | float | None :param context: Custom context for the event :type context: context_array | None :rtype: tracker """ pb = payload.Payload() pb.add("e", "se") pb.add("se_ca", category) pb.add("se_ac", action) pb.add("se_la", label) pb.add("se_pr", property_) pb.add("se_va", value) return self.complete_payload(pb, context, tstamp) @contract def track_unstruct_event(self, event_json, context=None, tstamp=None): """ :param event_json: The properties of the event. Has two field: A "data" field containing the event properties and A "schema" field identifying the schema against which the data is validated :type event_json: self_describing_json :param context: Custom context for the event :type context: context_array | None :param tstamp: User-set timestamp :type tstamp: timestamp | int | None :rtype: tracker """ envelope = SelfDescribingJson(UNSTRUCT_EVENT_SCHEMA, event_json.to_json()).to_json() pb = payload.Payload() pb.add("e", "ue") pb.add_json(envelope, self.encode_base64, "ue_px", "ue_pr") return self.complete_payload(pb, context, tstamp) # Alias track_self_describing_event = track_unstruct_event @contract def flush(self, is_async=False): """ Flush the emitter :param is_async: Whether the flush is done asynchronously. Default is False :type is_async: bool :rtype: tracker """ for emitter in self.emitters: if is_async: emitter.flush() else: emitter.sync_flush() return self @contract def set_subject(self, subject): """ Set the subject of the events fired by the tracker :param subject: Subject to be tracked :type subject: subject | None :rtype: tracker """ self.subject = subject return self @contract def add_emitter(self, emitter): """ Add a new emitter to which events should be passed :param emitter: New emitter :type emitter: emitter :rtype: tracker """ self.emitters.append(emitter) return self @staticmethod def check_form_element(element): """ PyContracts helper method to check that dictionary conforms element in sumbit_form and change_form schemas """ all_present = isinstance(element, dict) and 'name' in element and 'value' in element and 'nodeName' in element try: if element['type'] in FORM_TYPES: type_valid = True else: type_valid = False except KeyError: type_valid = True return all_present and element['nodeName'] in FORM_NODE_NAMES and type_valid
27,790
7,758
import logging from http import cookiejar as http_cookiejar from http.cookiejar import http2time # type: ignore from typing import Any # noqa from typing import Dict # noqa from urllib.parse import parse_qs from urllib.parse import urlsplit from urllib.parse import urlunsplit from oic.exception import UnSupported from oic.oauth2.exception import TimeFormatError from oic.utils.sanitize import sanitize logger = logging.getLogger(__name__) __author__ = "roland" URL_ENCODED = "application/x-www-form-urlencoded" JSON_ENCODED = "application/json" DEFAULT_POST_CONTENT_TYPE = URL_ENCODED PAIRS = { "port": "port_specified", "domain": "domain_specified", "path": "path_specified", } ATTRS = { "version": None, "name": "", "value": None, "port": None, "port_specified": False, "domain": "", "domain_specified": False, "domain_initial_dot": False, "path": "", "path_specified": False, "secure": False, "expires": None, "discard": True, "comment": None, "comment_url": None, "rest": "", "rfc2109": True, } # type: Dict[str, Any] def get_or_post( uri, method, req, content_type=DEFAULT_POST_CONTENT_TYPE, accept=None, **kwargs ): """ Construct HTTP request. :param uri: :param method: :param req: :param content_type: :param accept: :param kwargs: :return: """ if method in ["GET", "DELETE"]: if req.keys(): _req = req.copy() comp = urlsplit(str(uri)) if comp.query: _req.update(parse_qs(comp.query)) _query = str(_req.to_urlencoded()) path = urlunsplit( (comp.scheme, comp.netloc, comp.path, _query, comp.fragment) ) else: path = uri body = None elif method in ["POST", "PUT"]: path = uri if content_type == URL_ENCODED: body = req.to_urlencoded() elif content_type == JSON_ENCODED: body = req.to_json() else: raise UnSupported("Unsupported content type: '%s'" % content_type) header_ext = {"Content-Type": content_type} if accept: header_ext = {"Accept": accept} if "headers" in kwargs.keys(): kwargs["headers"].update(header_ext) else: kwargs["headers"] = header_ext else: raise UnSupported("Unsupported HTTP method: '%s'" % method) return path, body, kwargs def set_cookie(cookiejar, kaka): """ Place a cookie (a http_cookielib.Cookie based on a set-cookie header line) in the cookie jar. Always chose the shortest expires time. :param cookiejar: :param kaka: Cookie """ # default rfc2109=False # max-age, httponly for cookie_name, morsel in kaka.items(): std_attr = ATTRS.copy() std_attr["name"] = cookie_name _tmp = morsel.coded_value if _tmp.startswith('"') and _tmp.endswith('"'): std_attr["value"] = _tmp[1:-1] else: std_attr["value"] = _tmp std_attr["version"] = 0 attr = "" # copy attributes that have values try: for attr in morsel.keys(): if attr in ATTRS: if morsel[attr]: if attr == "expires": std_attr[attr] = http2time(morsel[attr]) else: std_attr[attr] = morsel[attr] elif attr == "max-age": if morsel[attr]: std_attr["expires"] = http2time(morsel[attr]) except TimeFormatError: # Ignore cookie logger.info( "Time format error on %s parameter in received cookie" % (sanitize(attr),) ) continue for att, spec in PAIRS.items(): if std_attr[att]: std_attr[spec] = True if std_attr["domain"] and std_attr["domain"].startswith("."): std_attr["domain_initial_dot"] = True if morsel["max-age"] == 0: try: cookiejar.clear( domain=std_attr["domain"], path=std_attr["path"], name=std_attr["name"], ) except ValueError: pass else: # Fix for Microsoft cookie error if "version" in std_attr: try: std_attr["version"] = std_attr["version"].split(",")[0] except (TypeError, AttributeError): pass new_cookie = http_cookiejar.Cookie(**std_attr) # type: ignore cookiejar.set_cookie(new_cookie) def match_to_(val, vlist): if isinstance(vlist, str): if vlist.startswith(val): return True else: for v in vlist: if v.startswith(val): return True return False def verify_header(reqresp, body_type): logger.debug("resp.headers: %s" % (sanitize(reqresp.headers),)) logger.debug("resp.txt: %s" % (sanitize(reqresp.text),)) if body_type == "": _ctype = reqresp.headers["content-type"] if match_to_("application/json", _ctype): body_type = "json" elif match_to_("application/jwt", _ctype): body_type = "jwt" elif match_to_(URL_ENCODED, _ctype): body_type = "urlencoded" else: body_type = "txt" # reasonable default ?? elif body_type == "json": if not match_to_("application/json", reqresp.headers["content-type"]): if match_to_("application/jwt", reqresp.headers["content-type"]): body_type = "jwt" else: raise ValueError( "content-type: %s" % (reqresp.headers["content-type"],) ) elif body_type == "jwt": if not match_to_("application/jwt", reqresp.headers["content-type"]): raise ValueError( "Wrong content-type in header, got: {} expected " "'application/jwt'".format(reqresp.headers["content-type"]) ) elif body_type == "urlencoded": if not match_to_(DEFAULT_POST_CONTENT_TYPE, reqresp.headers["content-type"]): if not match_to_("text/plain", reqresp.headers["content-type"]): raise ValueError("Wrong content-type") else: raise ValueError("Unknown return format: %s" % body_type) return body_type
6,597
1,992
import sys import yaml import numpy import random import logging import argparse from os import path, makedirs from datetime import datetime from src.finger_proxy.proxy_agent import ProxyAgent from src.utilities.logging_config_manager import setup_logging from src.display.touchscreendevice import TouchScreenDevice from src.vision.vision_agent import VisionAgent from src.finger.finger_agent import FingerAgent from src.proofread.proofread_agent import ProofreadAgent from src.supervisor.supervisor_agent import SupervisorAgent parser = argparse.ArgumentParser() # General parameters parser.add_argument("--all", action="store_true", default=False, help="train/test all the agents [vision, finger, proofread, supervisor]") parser.add_argument("--vision", action="store_true", default=False, help="train/test only the vision agent") parser.add_argument("--finger", action="store_true", default=False, help="train/test only the finger agent") parser.add_argument("--proofread", action="store_true", default=False, help="train/test only the proofread agent") parser.add_argument("--supervisor", action="store_true", default=False, help="train/test only the supervisor agent") parser.add_argument("--train", action="store_true", default=False, help="run model in train mode") parser.add_argument("--config", required=True, help="name of the configuration file (REQUIRED)") parser.add_argument("--seed", type=int, default=datetime.now().microsecond, help="random seed default: current time") parser.add_argument("--type", default=">", help="sentence to type for the agent.") parser.add_argument("--batch", action="store_true", default=False, help="evaluate a batch of sentences.") parser.add_argument("--users", type=int, default=1, help="number of users to simulate") parser.add_argument("--twofinger", action="store_true", default=False, help="enable typing with two finger.") parser.add_argument("--verbose", action="store_true", default=False, help="print tqdm step in new line.") # get user command line arguments. args = parser.parse_args() # Initialise random seed. numpy.random.seed(args.seed) random.seed(args.seed) # Setup Logger. if not path.isdir("logs"): # if logs folder doesn't exist create one. makedirs("logs") setup_logging(default_path=path.join("configs", "logging.yml")) logger = logging.getLogger(__name__) logger.info("logger is set.") # load app config. if path.exists(path.join("configs", args.config)): with open(path.join("configs", args.config), 'r') as file: config_file = yaml.load(file, Loader=yaml.FullLoader) logger.info("App Configurations loaded.") else: logger.error("File doesn't exist: Failed to load %s file under configs folder." % str(args.config)) sys.exit(0) if args.train: if path.exists(path.join("configs", config_file['training_config'])): with open(path.join("configs", config_file['training_config']), 'r') as file: train_config = yaml.load(file, Loader=yaml.FullLoader) logger.info("Training Configurations loaded.") else: logger.error("File doesn't exist: Failed to load %s file under configs folder." % config_file['training_config']) sys.exit(0) if args.vision or args.all: logger.info("Initiating Vision Agent Training.") vision_agent = VisionAgent(config_file['device_config'], train_config['vision'], args.verbose) vision_agent.train(vision_agent.episodes) if args.finger or args.all: logger.info("Initiating Finger Agent Training.") finger_agent = FingerAgent(config_file['device_config'], train_config['finger'], 0, True, args.verbose) finger_agent.train(finger_agent.episodes) if args.proofread or args.all: logger.info("Initiating Proofread Agent Training.") proofread_agent = ProofreadAgent(config_file['device_config'], train_config['proofread'], args.verbose) proofread_agent.train(proofread_agent.episodes) if args.supervisor or args.all: logger.info("Initiating Supervisor Agent Training.") if args.twofinger: supervisor_agent = SupervisorAgent(config_file['device_config'], train_config, True, True, args.verbose) else: supervisor_agent = SupervisorAgent(config_file['device_config'], train_config, True, False, args.verbose) print(type(supervisor_agent.episodes)) supervisor_agent.train(supervisor_agent.episodes) else: if path.exists(path.join("configs", config_file['testing_config'])): with open(path.join("configs", config_file['testing_config']), 'r') as file: test_config = yaml.load(file, Loader=yaml.FullLoader) logger.info("Training Configurations loaded.") else: logger.error("File doesn't exist: Failed to load %s file under configs folder." % config_file['testing_config']) sys.exit(0) if args.vision or args.all: logger.info("Initiating Vision Agent Evaluation.") vision_agent = VisionAgent(config_file['device_config'], test_config['vision']) vision_agent.evaluate(args.type) if args.finger or args.all: logger.info("Initiating Finger Agent Evaluation.") finger_agent = FingerAgent(config_file['device_config'], test_config['finger'], 0, False) finger_agent.evaluate(args.type, sat_desired=test_config['finger']['typing_accuracy']) if args.supervisor or args.all: logger.info("Initiating Supervisor Agent Evaluation.") if args.twofinger: supervisor_agent = SupervisorAgent(config_file['device_config'], test_config, False, True, args.verbose) else: supervisor_agent = SupervisorAgent(config_file['device_config'], test_config, False, False, args.verbose) supervisor_agent.evaluate(args.type, args.batch, args.users)
5,890
1,721
#!/usr/bin/python # # Decoding a legacy chain ric # import pyrfa p = pyrfa.Pyrfa() p.createConfigDb("./pyrfa.cfg") p.acquireSession("Session1") p.createOMMConsumer() p.login() p.directoryRequest() p.dictionaryRequest() p.setInteractionType("snapshot") def snapshotRequest(chainRIC): p.marketPriceRequest(chainRIC) snapshots = p.dispatchEventQueue(1000) if snapshots: for snapshot in snapshots: if snapshot['MTYPE'] == 'IMAGE': return snapshot return () fids = ['LINK_1', 'LINK_2', 'LINK_3', 'LINK_4', 'LINK_5', 'LINK_6', 'LINK_7', 'LINK_8', 'LINK_9', 'LINK_10', 'LINK_11', 'LINK_12', 'LINK_13', 'LINK_14', 'LONGLINK1', 'LONGLINK2', 'LONGLINK3', 'LONGLINK4', 'LONGLINK5', 'LONGLINK6', 'LONGLINK7', 'LONGLINK8', 'LONGLINK9', 'LONGLINK10', 'LONGLINK11', 'LONGLINK12', 'LONGLINK13', 'LONGLINK14', 'BR_LINK1', 'BR_LINK2', 'BR_LINK3', 'BR_LINK4', 'BR_LINK5', 'BR_LINK6', 'BR_LINK7', 'BR_LINK8', 'BR_LINK9', 'BR_LINK10', 'BR_LINK11', 'BR_LINK12', 'BR_LINK13', 'BR_LINK14'] def expandChainRIC(chainRIC): expanded = [] done = False snapshot = snapshotRequest(chainRIC) while not done: if not snapshot: break for fid in fids: if snapshot.has_key(fid) and snapshot[fid]: expanded.append(snapshot[fid]) if snapshot.has_key('NEXT_LR') and snapshot['NEXT_LR']: snapshot = snapshotRequest(snapshot['NEXT_LR']) elif snapshot.has_key('LONGNEXTLR') and snapshot['LONGNEXTLR']: snapshot = snapshotRequest(snapshot['LONGNEXTLR']) elif snapshot.has_key('BR_NEXTLR') and snapshot['BR_NEXTLR']: snapshot = snapshotRequest(snapshot['BR_NEXTLR']) else: done = True return expanded rics = expandChainRIC("0#.FTSE") print(rics)
1,854
736
#!/usr/bin/python # -*- coding: utf-8 -*- # launchctl unload /System/Library/LaunchDaemons/com.apple.syslogd.plist # launchctl load /System/Library/LaunchDaemons/com.apple.syslogd.plist from twisted.internet import reactor, stdio, defer from twisted.internet.protocol import Protocol, Factory from twisted.protocols.basic import LineReceiver import time, re, math, json #<22>Nov 1 00:12:04 gleicon-vm1 postfix/smtpd[4880]: connect from localhost[127.0.0.1] severity = ['emerg', 'alert', 'crit', 'err', 'warn', 'notice', 'info', 'debug', ] facility = ['kern', 'user', 'mail', 'daemon', 'auth', 'syslog', 'lpr', 'news', 'uucp', 'cron', 'authpriv', 'ftp', 'ntp', 'audit', 'alert', 'at', 'local0', 'local1', 'local2', 'local3', 'local4', 'local5', 'local6', 'local7',] fs_match = re.compile("<(.+)>(.*)", re.I) class SyslogdProtocol(LineReceiver): delimiter = '\n' def connectionMade(self): print 'Connection from %r' % self.transport def lineReceived(self, line): k = {} k['line'] = line.strip() (fac, sev) = self._calc_lvl(k['line']) k['host'] = self.transport.getHost().host k['tstamp'] = time.time() k['facility'] = fac k['severity'] = sev print json.dumps(k) def _calc_lvl(self, line): lvl = fs_match.split(line) if lvl and len(lvl) > 1: i = int(lvl[1]) fac = int(math.floor(i / 8)) sev = i - (fac * 8) return (facility[fac], severity[sev]) return (None, None) class SyslogdFactory(Factory): protocol = SyslogdProtocol def main(): factory = SyslogdFactory() reactor.listenTCP(25000, factory, 10) reactor.run() if __name__ == '__main__': main()
1,746
655
import requests as reqlib import os import re import random import time import pickle import abc import hashlib import threading from urllib.parse import urlparse from purifier import TEAgent from purifier.logb import getLogger from enum import IntEnum from typing import Tuple, List, Dict, Optional class ScraperTimeout(Exception): def __init__(self, ex): self.ex = ex def __str__(self): return f"Timeout: {self.ex}" class ScraperNot200(Exception): def __init__(self, sc): self.sc = sc def __str__(self): return f"Unexpected Status Code={self.sc}!" class UnsupportedMIME(Exception): def __init__(self, mime): self.mime = mime def __str__(self): return f"Unsupported MIME={self.mime}!" class Scraper(metaclass=abc.ABCMeta): @abc.abstractmethod def get(self, url): pass class ReqScraper(object): def __init__(self, page_cache_path="page_caches", headers={'User-Agent': 'Mozilla/5.0'}, skip_cache=False, supported_mime_set={"text/html"}): self.page_cache_path = page_cache_path if not os.path.isdir(self.page_cache_path): os.makedirs(self.page_cache_path) self.headers = headers self.logger = getLogger(os.path.basename(self.__class__.__name__)) self.skip_cache = skip_cache self.supported_mime_set = supported_mime_set def _get_cache_path(self, url): test_url_host = urlparse(url).netloc url_md5 = hashlib.md5(url.encode('utf-8')).hexdigest() cache_file_name = f"{test_url_host}_{url_md5}.txt" cache_file_path = os.path.join(self.page_cache_path, cache_file_name) return cache_file_path def _del_from_cache(self, url): cache_file_path = self._get_cache_path(url) if os.path.isfile(cache_file_path): self.logger.warning("Removing cache file={cache_file_path}...") os.remove(cache_file_path) def _get_from_cache(self, url): cache_file_path = self._get_cache_path(url) if os.path.isfile(cache_file_path): self.logger.debug(f"Return content of {url} from cache...") with open(cache_file_path, 'r', encoding='utf8') as fo: return fo.read() return None def _save2cache(self, url, html_content): cache_file_path = self._get_cache_path(url) with open(cache_file_path, 'w', encoding='utf8') as fw: fw.write(html_content) def get(self, url): if not self.skip_cache: cache_text = self._get_from_cache(url) if cache_text is not None: return cache_text self.logger.debug(f"Crawling {url}...") try: resp = reqlib.get(url, headers=self.headers, timeout=(5, 10)) if resp.ok: mime = resp.headers['content-type'].split(';')[0].strip() self.logger.debug(f"URL={url} with MIME={mime}...") if mime.lower() not in self.supported_mime_set: raise UnsupportedMIME(mime) self._save2cache(url, resp.text) return resp.text else: raise ScraperNot200(resp.status_code) except Exception as e: raise ScraperTimeout(e) class bcolors: HEADER = '\033[95m' OKBLUE = '\033[94m' OKCYAN = '\033[96m' OKGREEN = '\033[92m' WARNING = '\033[93m' FAIL = '\033[91m' ENDC = '\033[0m' BOLD = '\033[1m' UNDERLINE = '\033[4m' class ThreadState(IntEnum): STOPPED = 0 RUNNING = 1 STOPPING = 2 class CrawlAgent(object): def __init__(self, name, throttling_range=(1, 2)): self.rs = ReqScraper(page_cache_path=f"{name}_cache") self.et = TEAgent( policy_path="policy", disable_policy=True, ext_title=True ) self.logger = getLogger(os.path.basename(self.__class__.__name__)) self.throttling_range = throttling_range def obsolete_cache(self, url): self.rs._del_from_cache(url) def handle(self, url:str, skip_throttling:bool=False) -> Tuple[str, str, List[str]]: try: if skip_throttling: wait_in_sec = random.uniform(*self.throttling_range) self.logger.debug(f"throttling wait {wait_in_sec}s...") time.sleep(wait_in_sec) url_content_html = self.rs.get(url) is_succ, rst, handler = self.et.parse( "text/html", url, url_content_html, do_ext_link=True ) if is_succ: return (rst['title'], rst['text'], rst['all_links']) else: return (rst['title'], rst['text'], rst['all_links']) except ScraperNot200 as e: self.logger.warning(f"Fail to handle URL={url}: {str(e)}") return None, None, None except UnsupportedMIME as e: self.logger.warning(f"Fail to handle URL={url}: {str(e)}") return None, None, None except ScraperTimeout as e: time.sleep(2) self.logger.warning(f"Fail to handle URL={url}: {str(e)}") return None, None, None class ExplorerWorker(threading.Thread): def __init__( self, name:str, url_ptn:str, src_url:str, test_run:int=-1, page_saved_dir:Optional[str]=None): super(ExplorerWorker, self ).__init__(name = name) self.name = name self.url_ptn = url_ptn self.src_url = src_url self.test_run = test_run self.ca = CrawlAgent(name) self.pc_dict = self._get_pc_dict() ''' Processed result cache: Key as URL; value as bool (True means this URL is crawled successfully)''' self.state = ThreadState.STOPPED ''' Thread state: 0-> stopped; 1-> running; 2-> stopping''' self.logger = getLogger(os.path.basename(self.__class__.__name__)) ''' Logger object ''' self.page_saved_dir = page_saved_dir if page_saved_dir is not None else f"{self.name}_pages_output" ''' Path or directory to save dump page''' self.stop_signal = f"STOP_{self.name}" ''' Stop signal file ''' if not os.path.isdir(self.page_saved_dir): os.makedirs(self.page_saved_dir) def _get_output_page_path(self, url): url_host = urlparse(url).netloc url_md5 = hashlib.md5(url.encode('utf-8')).hexdigest() page_file_name = f"{url_host}_{url_md5}.txt" page_file_path = os.path.join(self.page_saved_dir, page_file_name) return page_file_path def _get_pc_serialized_file(self) -> str: return f"{self.name}_pc_dict.pkl" def _get_pc_dict(self) -> Dict[str, bool]: pkl_file = self._get_pc_serialized_file() if os.path.isfile(pkl_file): with open(pkl_file, 'rb') as fo: return pickle.load(fo) else: return {} def _serialized(self): pkl_file = self._get_pc_serialized_file() with open(pkl_file, 'wb') as fo: pickle.dump(self.pc_dict, fo) def run(self): self.state = ThreadState.RUNNING url_queue = [self.src_url] pc = sc = fc = oc = 0 while self.state == ThreadState.RUNNING and url_queue: if os.path.isfile(self.stop_signal): os.remove(self.stop_signal) self.logger.warning("Receive STOP signal!") break url = url_queue.pop(0) pc += 1 if url not in self.pc_dict: # New URL self.logger.debug(f"Handling URL={url}...") title, content, collected_urls = self.ca.handle(url) if content is None: self.pc_dict[url] = False fc += 1 else: if url != self.src_url: self.pc_dict[url] = True sc += 1 self.logger.info(bcolors.BOLD + f"Completed URL={url} ({len(url_queue):,d}/{pc:,d})" + bcolors.ENDC) next_level_urls = list(filter(lambda u: re.match(self.url_ptn, u) is not None and "#" not in u, collected_urls)) if next_level_urls: self.logger.debug(f"\tCollected {len(next_level_urls)} next level URL(s)") url_queue.extend(list(set(next_level_urls) - set(url_queue))) if content and "?" not in url: page_output_path = self._get_output_page_path(url) with open(page_output_path, 'w', encoding='utf8') as fw: fw.write(f"{url}\n\n") fw.write(f"{title}\n\n") fw.write(f"{content}") self.logger.debug(f"\tSaved page to {page_output_path}!") else: # Old URL if not self.pc_dict[url]: self.logger.info(f"Skip broken URL={url} in the past...") continue title, content, collected_urls = self.ca.handle(url, skip_throttling=True) if collected_urls: next_level_urls = list(filter(lambda u: re.match(self.url_ptn, u) is not None, collected_urls)) url_queue.extend(list(set(next_level_urls) - set(url_queue))) oc += 1 self.logger.info(f"URL={url} is already handled...({len(url_queue):,d}/{pc:,d})") continue if self.test_run > 0: if (sc + fc) > self.test_run: self.logger.info(f"Exceed test_run={self.test_run} and therefore stop running...") break if pc % 1000 == 0: self.logger.info(bcolors.OKBLUE + bcolors.BOLD + f"{pc} URL completed: sc={sc:,d}; fc={fc:,d}; oc={oc:,d}\n" + bcolors.ENDC) self._serialized() self.ca.obsolete_cache(self.src_url) url_queue.append(self.src_url) self.logger.warning(f"Serialized explorer result (name={self.name})...") self._serialized() self.logger.warning(f"Explorer is stopped! (name={self.name})...") self.state = ThreadState.STOPPED def stop(self): self.logger.warning(f"Stopping explorer worker (name={self.name})...") if self.state == ThreadState.RUNNING: self.state = ThreadState.STOPPING while self.state != ThreadState.STOPPED: time.sleep(1)
11,371
3,755
# CPU实时监控 # 作者:Charles # 公众号:Charles的皮卡丘 import matplotlib.pyplot as plt import matplotlib.font_manager as font_manager import psutil as p POINTS = 300 fig, ax = plt.subplots() ax.set_ylim([0, 100]) ax.set_xlim([0, POINTS]) ax.set_autoscale_on(False) ax.set_xticks([]) ax.set_yticks(range(0, 101, 10)) ax.grid(True) # 执行用户进程的时间百分比 user = [None] * POINTS # 执行内核进程和中断的时间百分比 sys = [None] * POINTS # CPU处于空闲状态的时间百分比 idle = [None] * POINTS l_user, = ax.plot(range(POINTS), user, label='User %') l_sys, = ax.plot(range(POINTS), sys, label='Sys %') l_idle, = ax.plot(range(POINTS), idle, label='Idle %') ax.legend(loc='upper center', ncol=4, prop=font_manager.FontProperties(size=10)) bg = fig.canvas.copy_from_bbox(ax.bbox) def cpu_usage(): t = p.cpu_times() return [t.user, t.system, t.idle] before = cpu_usage() def get_cpu_usage(): global before now = cpu_usage() delta = [now[i] - before[i] for i in range(len(now))] total = sum(delta) before = now return [(100.0*dt)/(total+0.1) for dt in delta] def OnTimer(ax): global user, sys, idle, bg tmp = get_cpu_usage() user = user[1:] + [tmp[0]] sys = sys[1:] + [tmp[1]] idle = idle[1:] + [tmp[2]] l_user.set_ydata(user) l_sys.set_ydata(sys) l_idle.set_ydata(idle) while True: try: ax.draw_artist(l_user) ax.draw_artist(l_sys) ax.draw_artist(l_idle) break except: pass ax.figure.canvas.draw() def start_monitor(): timer = fig.canvas.new_timer(interval=100) timer.add_callback(OnTimer, ax) timer.start() plt.show() if __name__ == '__main__': start_monitor()
1,554
763
import datetime from rest_framework import viewsets from rest_framework.response import Response from leaderboards import LEADERBOARD_MAX_PER_PAGE from leaderboards.models import LeaderboardEntry, Medal, LeaderboardSpan from leaderboards.serializers import (LeaderboardEntrySerializer, MedalSerializer, LeaderboardSerializer, LeaderboardSpanSerializer, LeaderboardEntrySpanSerializer) from leaderboards import service as leaderboards_service from users import service as users_service from channels import service as channels_service from shows import service as shows_service from utilities.api import APIObject class LeaderboardEntryAPIObject(APIObject): field_list = ['id', 'user_id', 'points', 'wins'] def __init__(self, leaderboard_entry, **kwargs): super(LeaderboardEntryAPIObject, self).__init__(leaderboard_entry, **kwargs) self.show = leaderboard_entry.show_id self.channel_name = leaderboard_entry.channel.name self.medals = leaderboards_service.fetch_medal_ids_by_leaderboard_entry(leaderboard_entry.id) # If a user is attached to the entry if leaderboard_entry.user_id: # Get their username user_profile = users_service.fetch_user_profile(leaderboard_entry.user_id) self.username = user_profile.safe_username # Get the number of suggestions submitted self.suggestions = shows_service.fetch_suggestions(user_id=leaderboard_entry.user_id, show_id=leaderboard_entry.show_id, count=True) class LeaderboardEntrySpanAPIObject(APIObject): def __init__(self, leaderboard_span_entry, **kwargs): super(LeaderboardEntrySpanAPIObject, self).__init__(leaderboard_span_entry, **kwargs) self.channel_name = kwargs.get('channel_name') self.user_id = leaderboard_span_entry['user_id'] user_profile = users_service.fetch_user_profile(self.user_id) self.username = user_profile.safe_username self.points = leaderboard_span_entry['points'] self.show_wins = leaderboard_span_entry['show_wins'] self.suggestion_wins = leaderboard_span_entry['suggestion_wins'] class LeaderboardAPIObject(APIObject): field_list = ['id', 'points', 'suggestion_wins', 'show_wins'] def __init__(self, channel_user, **kwargs): super(LeaderboardAPIObject, self).__init__(channel_user, **kwargs) self.channel_name = kwargs.get('channel_name') user_profile = users_service.fetch_user_profile(channel_user.user_id) self.user_id = channel_user.user_id self.username = user_profile.safe_username class LeaderboardEntryViewSet(viewsets.ViewSet): """ API endpoint that returns leaderboard entries """ def list(self, request): kwargs = {} user_id = self.request.query_params.get('user_id') channel_id = self.request.query_params.get('channel_id') show_id = self.request.query_params.get('show_id') limit = self.request.query_params.get('limit') # Pagination page = int(self.request.query_params.get('page', 1)) offset = LEADERBOARD_MAX_PER_PAGE * (page - 1) order_by_show_date = self.request.query_params.get('order_by_show_date') if user_id: kwargs['user'] = user_id if channel_id: kwargs['channel'] = channel_id if show_id: kwargs['show'] = show_id # Make sure we exclude any entries that don't have users attached queryset = LeaderboardEntry.objects.filter(**kwargs).exclude(user=None) # If we are ordering by when the show happened if order_by_show_date is not None: queryset = queryset.order_by('-show_date') # Order by suggestion wins, then points for a show or span if show_id: queryset = queryset.order_by('-wins', '-points') # If there is a limit to the results returned if limit: queryset = queryset[:int(limit)] # Start from the page offset try: queryset = queryset[offset:offset+LEADERBOARD_MAX_PER_PAGE] except IndexError: try: queryset = queryset[offset:] except IndexError: pass leaderboard_entry_list = [LeaderboardEntryAPIObject(item) for item in queryset] serializer = LeaderboardEntrySerializer(leaderboard_entry_list, many=True) return Response(serializer.data) class LeaderboardEntrySpanViewSet(viewsets.ViewSet): """ API endpoint that returns leaderboard entries by span dates """ def list(self, request): kwargs = {} kwargs['channel'] = self.request.query_params.get('channel_id') channel = channels_service.channel_or_404(kwargs['channel'], channel_id=True) start = self.request.query_params.get('start') end = self.request.query_params.get('end') # Convert start and end to datetimes start_time = datetime.datetime.strptime(start, "%Y%m%d") end_time = datetime.datetime.strptime(end, "%Y%m%d") # Add them to the queryset params kwargs['show_date__gte'] = start_time kwargs['show_date__lte'] = end_time # Make sure we exclude any entries that don't have users attached queryset = LeaderboardEntry.objects.filter(**kwargs).exclude(user=None) # Aggregate all the leaderboard entries by the user leaderboard_aggregate = leaderboards_service.aggregate_leaderboard_entries_by_user( queryset) api_kwargs = {'channel_name': channel.name} # Make an api object out of the user aggregates leaderboard_entry_list = [LeaderboardEntrySpanAPIObject(item, **api_kwargs) for item in leaderboard_aggregate] serializer = LeaderboardEntrySpanSerializer(leaderboard_entry_list, many=True) return Response(serializer.data) class LeaderboardViewSet(viewsets.ViewSet): """ API endpoint for show, channel, or combined leaderboards """ def list(self, request): channel_id = self.request.query_params.get('channel_id') page = int(self.request.query_params.get('page', 1)) offset = LEADERBOARD_MAX_PER_PAGE * (page - 1) queryset = channels_service.fetch_channel_users(channel_id, leaderboard_sort=True) # Start from the page offset try: queryset = queryset[offset:offset+LEADERBOARD_MAX_PER_PAGE] except IndexError: try: queryset = queryset[offset:] except IndexError: pass leaderboard_list = [LeaderboardAPIObject(item) for item in queryset] serializer = LeaderboardSerializer(leaderboard_list, many=True) return Response(serializer.data) class LeaderboardSpanViewSet(viewsets.ViewSet): """ API endpoint that allows leaderboard spans to be viewed """ def retrieve(self, request, pk=None): leaderboard_span = leaderboards_service.leaderboard_span_or_404(pk) serializer = LeaderboardSpanSerializer(leaderboard_span) return Response(serializer.data) def list(self, request): """ This view should return a list of all leaderboard spans """ queryset = LeaderboardSpan.objects.all() channel_id = self.request.query_params.get('channel_id') if channel_id: queryset = queryset.filter(channel=channel_id) serializer = LeaderboardSpanSerializer(queryset, many=True) return Response(serializer.data) class MedalViewSet(viewsets.ModelViewSet): """ API endpoint that allows medals to be viewed """ model = Medal serializer_class = MedalSerializer queryset = Medal.objects.all()
8,128
2,272
from ... pyaz_utils import _call_az def show(name, resource_group): ''' Get the details for an instance pool. Required Parameters: - name -- Instance Pool Name - resource_group -- Name of resource group. You can configure the default group using `az configure --defaults group=<name>` ''' return _call_az("az sql instance-pool show", locals()) def list(resource_group=None): ''' List available instance pools. Optional Parameters: - resource_group -- Name of resource group. You can configure the default group using `az configure --defaults group=<name>` ''' return _call_az("az sql instance-pool list", locals()) def update(name, resource_group, add=None, force_string=None, remove=None, set=None, tags=None): ''' Update an instance pool. Required Parameters: - name -- Instance Pool Name - resource_group -- Name of resource group. You can configure the default group using `az configure --defaults group=<name>` Optional Parameters: - add -- Add an object to a list of objects by specifying a path and key value pairs. Example: --add property.listProperty <key=value, string or JSON string> - force_string -- When using 'set' or 'add', preserve string literals instead of attempting to convert to JSON. - remove -- Remove a property or an element from a list. Example: --remove property.list <indexToRemove> OR --remove propertyToRemove - set -- Update an object by specifying a property path and value to set. Example: --set property1.property2=<value> - tags -- space-separated tags: key[=value] [key[=value] ...]. Use '' to clear existing tags. ''' return _call_az("az sql instance-pool update", locals()) def delete(name, resource_group, no_wait=None, yes=None): ''' Delete an instance pool. Required Parameters: - name -- Instance Pool Name - resource_group -- Name of resource group. You can configure the default group using `az configure --defaults group=<name>` Optional Parameters: - no_wait -- Do not wait for the long-running operation to finish. - yes -- Do not prompt for confirmation. ''' return _call_az("az sql instance-pool delete", locals()) def create(capacity, family, location, name, resource_group, subnet, tier, license_type=None, no_wait=None, tags=None, vnet_name=None): ''' Create an instance pool. Required Parameters: - capacity -- Capacity of the instance pool in vcores. - family -- The compute generation component of the sku. Allowed value: Gen5 - location -- Location. Values from: `az account list-locations`. You can configure the default location using `az configure --defaults location=<location>`. - name -- Instance Pool Name - resource_group -- Name of resource group. You can configure the default group using `az configure --defaults group=<name>` - subnet -- Name or ID of the subnet that allows access to an Instance Pool. If subnet name is provided, --vnet-name must be provided. - tier -- The edition component of the sku. Allowed value: GeneralPurpose. Optional Parameters: - license_type -- The license type to apply for this instance pool. - no_wait -- Do not wait for the long-running operation to finish. - tags -- space-separated tags: key[=value] [key[=value] ...]. Use '' to clear existing tags. - vnet_name -- The virtual network name ''' return _call_az("az sql instance-pool create", locals()) def wait(name, resource_group, created=None, custom=None, deleted=None, exists=None, interval=None, timeout=None, updated=None): ''' Wait for an instance pool to reach a desired state. Required Parameters: - name -- Instance Pool Name - resource_group -- Name of resource group. You can configure the default group using `az configure --defaults group=<name>` Optional Parameters: - created -- wait until created with 'provisioningState' at 'Succeeded' - custom -- Wait until the condition satisfies a custom JMESPath query. E.g. provisioningState!='InProgress', instanceView.statuses[?code=='PowerState/running'] - deleted -- wait until deleted - exists -- wait until the resource exists - interval -- polling interval in seconds - timeout -- maximum wait in seconds - updated -- wait until updated with provisioningState at 'Succeeded' ''' return _call_az("az sql instance-pool wait", locals())
4,450
1,209
aluno = {} aluno['nome'] = str(input('Digite o nome do aluno: ')) aluno['media'] = float(input('Digite a média desse aluno: ')) if aluno['media'] >= 5: aluno['situação'] = '\033[32mAprovado\033[m' else: aluno['situação'] = '\033[31mReprovado\033[m' for k, v in aluno.items(): print(f'{k} do aluno é {v}')
316
147
# -*- coding: utf-8 -*- """ Created on Tue Aug 6 10:57:41 2019 Creates train and test splits at the IMAGE LEVEL to prep for thumbnail extraction in countr_cnn_1 @author: smcguire """ import pandas as pd import numpy as np import matplotlib.pyplot as plt ## read dataframe of unique images with annotation info df_unique = pd.read_pickle('./df_unique.pkl') # create df_test from every 4th image df_test = df_unique[df_unique.index % 4 == 0] # create df_train_val from every image not in df_test df_train_val = df_unique[df_unique.index % 4 != 0] # reset indexes df_test = df_test.reset_index(drop=True) df_train_val = df_train_val.reset_index(drop=True) # pickle dataframes df_test.to_pickle('./df_test.pkl') df_train_val.to_pickle('./df_train_val.pkl')
813
330
## @ingroup Attributes-Propellants # Aviation_Gasoline.py # # Created: Unk 2013, SUAVE TEAM # Modified: Apr 2015, SUAVE TEAM # ---------------------------------------------------------------------- # Imports # ---------------------------------------------------------------------- from .Propellant import Propellant # ---------------------------------------------------------------------- # Aviation_Gasoline Propellant Class # ---------------------------------------------------------------------- ## @ingroup Attributes-Propellants class Aviation_Gasoline(Propellant): """Contains density and specific energy values for this propellant Assumptions: None Source: None """ def __defaults__(self): """This sets the default values. Assumptions: None Source: Values commonly available Inputs: None Outputs: None Properties Used: None """ self.tag='Aviation Gasoline' self.density = 721.0 # kg/m^3 self.specific_energy = 43.71e6 # J/kg
1,144
315
import os, glob, shutil class Write_grok(): def __init__(self, grok_name, grok_directory): ''' modify a given grok file''' self.grok_directory = grok_directory self.grok_path = os.path.join(grok_directory,grok_name) if not os.path.exists(self.grok_path): print("File not found: {0}".format(self.grok_path)) return None self.backup_copy = self.grok_path + '.preunsat_original' if not os.path.exists(self.backup_copy): shutil.copy(self.grok_path, self.backup_copy) shutil.move(self.grok_path, self.grok_path + '.backup') def search_path(self, file_name, ldebug=False): ''' search for file name in the grok folder''' filename= glob.glob(os.path.join(self.grok_directory,'**',file_name), recursive=True) if not filename: raise IOError(file_name) if len(filename)>1: print('Warning: more than one match was found. {}'.format(filename)) if ldebug: print (filename) file_rela = (os.path.relpath(filename[0], self.grok_directory)) if ldebug: print (file_rela) return(file_rela) def add_gw_wells(self, inc_file_name, overwrite=False): '''add include file below the ref_line''' ref_line = 'Data Output' # get the relative path of the include file inc_file_path = self.search_path(inc_file_name) with open(self.backup_copy) as fhand, open(self.grok_path, 'w') as fcopy: line = fhand.readline() while line: if ref_line in line: # locate where command line need to be inserted output_flag=True while output_flag: if inc_file_name in line: print('file {} is already included'.format(inc_file_name)) output_flag = False elif 'Simulation Control Parameters' in line: output_flag = False fcopy.write('include .\{} \n'.format(inc_file_path)) fcopy.write(line) line=fhand.readline() fcopy.write(line) line=fhand.readline() if overwrite: shutil.copy(self.grok_path, self.backup_copy) def add_target_output_times(self,out_times,target_times=None,overwrite=False): with open(self.backup_copy) as fhand, open(self.grok_path, 'w') as fcopy: line = fhand.readline() while line: if line.strip().lower().startswith('output times'): o_time_flag = True fcopy.write(line) for t in out_times: fcopy.write('{0} \n'.format(t)) print('{0} \n'.format(t)) while o_time_flag: if line.strip().lower().startswith('end'): o_time_flag = False fcopy.write(line) line=fhand.readline() fcopy.write(line) line=fhand.readline() if overwrite: shutil.copy(self.grok_path, self.backup_copy) with open(self.backup_copy) as fhand, open(self.grok_path, 'w') as fcopy: line = fhand.readline() while line: if target_times: if line.strip().lower().startswith('target times'): tar_time_flag = True fcopy.write(line) for t in target_times: fcopy.write('{0} \n'.format(t)) while tar_time_flag: if line.strip().lower().startswith('end'): tar_time_flag = False fcopy.write(line) line=fhand.readline() fcopy.write(line) line=fhand.readline() if overwrite: shutil.copy(self.grok_path, self.backup_copy)
4,108
1,162
import decimal import hashlib import json import requests import tempfile import uuid import os from tqdm import tqdm from requests_toolbelt import MultipartEncoder, MultipartEncoderMonitor def sha256_for_file(f, buf_size=65536): pos = f.tell() dgst = hashlib.sha256() while True: data = f.read(buf_size) if not data: break dgst.update(data) size = f.tell() - pos f.seek(pos) return size, dgst.hexdigest() namespace = "default" fission_url = os.environ["FISSION_URL"] def post(rel_url, data): response = requests.post( "%s%s" % (fission_url, rel_url), data=json.dumps(data), headers={"Content-Type": "application/json"}) # print("POST", rel_url) # print(response, response.text) if response.status_code in [404, 409]: return response.status_code, None if response.status_code == 500: raise Exception(response.text) return response.status_code, response.json() def get(rel_url, params=None): response = requests.get( "%s%s" % (fission_url, rel_url), params=params) if response.status_code == 404: return response.status_code, None if response.status_code == 500: raise Exception(response.text) return response.status_code, response.json() def format_bytes(count): label_ix = 0 labels = ["B", "KiB", "MiB", "GiB"] while label_ix < len(labels) and count / 1024. > 1: count = count / 1024. label_ix += 1 count = decimal.Decimal(count) count = count.to_integral() if count == count.to_integral() else round(count.normalize(), 2) return "%s %s" % (count, labels[label_ix]) def lazily_define_package(environment, file): filesize, archive_sha256 = sha256_for_file(file) base_archive_url = "%s/proxy/storage/v1/archive" % fission_url status_code, response = get("/v2/packages/%s" % archive_sha256) if status_code == 200: print("Already uploaded", flush=True) return archive_sha256, response progress = tqdm( total=filesize, desc="Uploading", unit="B", unit_scale=True, unit_divisor=1024, leave=True) last_bytes_read = 0 def update_progress(monitor): # Your callback function nonlocal last_bytes_read progress.update(monitor.bytes_read - last_bytes_read) last_bytes_read = monitor.bytes_read e = MultipartEncoder(fields={'uploadfile': ('uploaded', file, 'text/plain')}) m = MultipartEncoderMonitor(e, update_progress) archive_response = requests.post(base_archive_url, data=m, headers={ "X-File-Size": str(filesize), 'Content-Type': m.content_type}) archive_id = archive_response.json()['id'] print(" done", flush=True) archive_url = "%s?id=%s" % (base_archive_url, archive_id) package = { "metadata": { "name": archive_sha256, "namespace": namespace, }, "spec": { "environment": environment, "deployment": { "type": "url", "url": archive_url, "checksum": { "type": "sha256", "sum": archive_sha256, }, }, }, "status": { "buildstatus": "succeeded", }, } return archive_sha256, post("/v2/packages", package)[1] def lazily_define_function(environment, f): archive_sha256, package_ref = lazily_define_package(environment, f) print("Registering ...", end='', flush=True) function_name = archive_sha256[:8] status_code, response = get("/v2/functions/%s" % function_name) if status_code == 200: return function_name status_code, r = post("/v2/functions", { "metadata": { "name": function_name, "namespace": namespace, }, "spec": { "environment": environment, "package": { "functionName": function_name, "packageref": package_ref, }, }, }) if status_code == 409 or status_code == 201: print(" done", flush=True) return function_name print(" error", flush=True) raise Exception(r.text) def lazily_define_trigger2(function_name, http_method, host, relativeurl): trigger_name = "%s-%s-%s" % ( host.replace('.', '-'), relativeurl.replace(':.*', '').replace('{', '').replace('}', '').replace('/', '-'), http_method.lower()) status_code, response = get("/v2/triggers/http/%s" % trigger_name) if status_code == 200: return status_code, r = post("/v2/triggers/http", { "metadata": { "name": trigger_name, "namespace": namespace, }, "spec": { "host": host, "relativeurl": relativeurl, "method": http_method, "functionref": { "Type": "name", "Name": function_name, }, }, }) if status_code == 409 or status_code == 201: return raise Exception(r.text) def publish(environment_name, f): environment = { "namespace": namespace, "name": environment_name, } function_name = lazily_define_function(environment, f) host = "%s.tfi.gcp.tesserai.com" % function_name lazily_define_trigger2(function_name, "POST", host, "/{path-info:.*}") lazily_define_trigger2(function_name, "GET", host, "/{path-info:.*}") lazily_define_trigger2(function_name, "GET", host, "/") return "http://%s" % host
5,752
1,848
from __future__ import annotations import sys from skeema.intermediate.compiler.parser import Parser from skeema import ModelMeta from skeema import util def private(name): return f'_{name}' class ClassBuilder: """ ClassBuilder """ @staticmethod def set_class_module(klass, module_name: str): klass.__module__ = module_name module = sys.modules[module_name] module.__dict__[klass.__name__] = klass @staticmethod def create_class(class_name: str, base_classes: [str], parameters: [dict], data_members: [dict]): module_name = 'skeema' # Populate a dictionary of property accessors cls_dict = dict() # Parsing for json def parse(cls, json_str: str): return Parser.parse(cls, json_str) cls_dict['parse'] = classmethod(parse) def decorate(annotation: str, array: bool) -> str: if array: return f'[{annotation}]' else: return annotation parameter_annotation_dict = { name: decorate(annotation, array) for name, annotation, array in ((parameter['name'], parameter['class'], parameter['array']) for parameter in parameters) } data_member_dict = { name: decorate(annotation, array) for name, annotation, array in ((data_member['name'], data_member['class'], data_member['array']) for data_member in data_members) } cls = ModelMeta( class_name, tuple(util.class_lookup(module_name, base_class) for base_class in base_classes), cls_dict, parameters=parameter_annotation_dict, data_members=data_member_dict ) ClassBuilder.set_class_module(cls, module_name) return cls
1,824
504
""" From http://jimorlin.wordpress.com/2009/02/17/colored-letters-labeled-dice-a-logic-puzzle/ There are 13 words as follows: buoy, cave, celt, flub, fork, hemp, judy, junk, limn, quip, swag, visa. There are 24 different letters that appear in the 13 words. The question is: can one assign the 24 letters to 4 different cubes so that the four letters of each word appears on different cubes. There is one letter from each word on each cube. The puzzle was created by Humphrey Dudley. Execution: python3 LabeledDice.py """ from pycsp3 import * words = ["buoy", "cave", "celt", "flub", "fork", "hemp", "judy", "junk", "limn", "quip", "swag", "visa"] # x[i] is the cube where the ith letter of the alphabet is put x = VarArray(size=26, dom=lambda i: range(1, 5) if i in alphabet_positions("".join(words)) else None) satisfy( # the four letters of each word appear on different cubes [AllDifferent(x[i] for i in alphabet_positions(w)) for w in words], # each cube is assigned 6 letters Cardinality(x, occurrences={i: 6 for i in range(1, 5)}) )
1,064
387
from JumpScale import j base = j.tools.cuisine._getBaseClass() class CuisineCockpit(base): def __init__(self, executor, cuisine): self._executor = executor self._cuisine = cuisine def configure(self): C = """ [mode] prod = false [oauth] client_secret = "IqFcD2TYC7WrPpoH2Oi6YTpN102Hj-sdfsdfsdfsdf-sMBcydIacGI" redirect_uri = "https://f0919f6e.ngrok.io/oauth/callback" client_id = "OrgTest" organization = "OrgTest" jwt_key = "" [api] [api.ays] port = 5000 host = "0.0.0.0" debug = true """ self._cuisine.core.dir_ensure("$cfgDir/cockpit_api") j.tools.cuisine.local.core.file_write("$cfgDir/cockpit_api/config.toml", C) def install(self, start=True, branch="master"): self.install_deps() self._cuisine.development.git.pullRepo('https://github.com/Jumpscale/jscockpit', branch=branch) dir_paths = self._cuisine.core.dir_paths self._cuisine.core.dir_ensure('%s/ays_api/' % dir_paths['appDir']) self._cuisine.core.file_link('%s/github/jumpscale/jscockpit/api_server' % dir_paths['codeDir'], '%s/ays_api/api_server' % dir_paths['appDir']) self._cuisine.core.file_link('%s/github/jumpscale/jscockpit/ays_api' % dir_paths['codeDir'], '%s/ays_api/ays_api' % dir_paths['appDir']) self.configure() if start: self.start() def install_all_in_one(self, start=True, branch="master", reset=True, ip="localhost"): """ This will install the all the component of the cockpit in one command. (mongodb, portal, ays_api, ays_daemon) Make sure that you don't have uncommitted code in any code repository cause this method will discard them !!! """ # install mongodb, required for portal self._cuisine.apps.mongodb.build(install=False, start=start, reset=reset) self._cuisine.apps.mongodb.install(start=start) # install portal self._cuisine.apps.portal.install(start=False, installdeps=True, branch=branch) # add link from portal to API # 1- copy the nav to the portalbase and then edit it content = self._cuisine.core.file_read('$codeDir/github/jumpscale/jumpscale_portal8/apps/portalbase/AYS81/.space/nav.wiki') # 2- fix the ays api endpoint. if 'REST API:/api' not in content: content += 'AYS API:http://{ip}:5000/apidocs/index.html?raml=api.raml'.format(ip=ip) self._cuisine.core.file_write('$appDir/portals/main/base/AYS81/.space/nav.wiki', content=content) self._cuisine.apps.portal.configure(production=False) self._cuisine.apps.portal.start() # install REST API AND ays daemon self.install(start=start, branch=branch) # configure base URI for api-console raml = self._cuisine.core.file_read('$appDir/ays_api/ays_api/apidocs/api.raml') raml = raml.replace('baseUri: https://localhost:5000', "baseUri: http://{ip}:5000".format(ip=ip)) self._cuisine.core.file_write('$appDir/ays_api/ays_api/apidocs/api.raml', raml) if start: # start API and daemon self.start() def start(self, name='main'): # start AYS REST API cmd = 'jspython api_server' dir_paths = self._cuisine.core.dir_paths self._cuisine.processmanager.ensure(cmd=cmd, name='cockpit_%s' % name, path='%s/ays_api' % dir_paths['appDir']) # start daemon cmd = 'ays start' self._cuisine.processmanager.ensure(cmd=cmd, name='cockpit_daemon_%s' % name) def stop(self, name='main'): self._cuisine.processmanager.stop('cockpit_%s' % name,) self._cuisine.processmanager.stop('cockpit_daemon_%s' % name,) def install_deps(self): self._cuisine.package.mdupdate() self._cuisine.package.install('libssl-dev') deps = """ cryptography python-jose wtforms_json flask_wtf python-telegram-bot """ self._cuisine.development.pip.multiInstall(deps, upgrade=True)
4,220
1,443
""" Finetuning BERT using DeepSpeed's ZeRO-Offload """ import json import dataclasses from pathlib import Path from functools import partial import nlp import torch import typer import deepspeed import numpy as np from transformers import BertTokenizerFast from transformers import BertForSequenceClassification from sklearn.model_selection import train_test_split from pytorch_helper_bot import ( DeepSpeedBot, MovingAverageStatsTrackerCallback, CheckpointCallback, LearningRateSchedulerCallback, MultiStageScheduler, Top1Accuracy, LinearLR, CosineAnnealingScheduler ) CACHE_DIR = Path("cache/") CACHE_DIR.mkdir(exist_ok=True) APP = typer.Typer() class SST2Dataset(torch.utils.data.Dataset): def __init__(self, entries_dict): super().__init__() self.entries_dict = entries_dict def __len__(self): return len(self.entries_dict["label"]) def __getitem__(self, idx): return ( self.entries_dict["input_ids"][idx], self.entries_dict["attention_mask"][idx], self.entries_dict["token_type_ids"][idx], self.entries_dict["label"][idx] ) @dataclasses.dataclass class SST2Bot(DeepSpeedBot): log_dir = CACHE_DIR / "logs" def __post_init__(self): super().__post_init__() self.loss_format = "%.6f" @staticmethod def extract_prediction(output): return output[0] class Object(object): pass def convert_to_features(tokenizer, example_batch): # Tokenize contexts and questions (as pairs of inputs) encodings = tokenizer.batch_encode_plus( example_batch['sentence'], padding='max_length', max_length=64, truncation=True) return encodings @APP.command( context_settings={"allow_extra_args": True, "ignore_unknown_options": True} ) def main(arch="bert-base-uncased", config="gpu.json"): # Reference: # # * https://github.com/huggingface/nlp/blob/master/notebooks/Overview.ipynb with open(config) as fin: config_params = json.load(fin) dataset = nlp.load_dataset('glue', "sst2") print(set([x['label'] for x in dataset["train"]])) tokenizer = BertTokenizerFast.from_pretrained(arch) # Format our dataset to outputs torch.Tensor to train a pytorch model columns = ['input_ids', 'token_type_ids', 'attention_mask', "label"] for subset in ("train", "validation"): dataset[subset] = dataset[subset].map( partial(convert_to_features, tokenizer), batched=True) dataset[subset].set_format(type='torch', columns=columns) print(tokenizer.decode(dataset['train'][6]["input_ids"].numpy())) print(dataset['train'][0]["attention_mask"]) valid_idx, test_idx = train_test_split( list(range(len(dataset["validation"]))), test_size=0.5, random_state=42) train_dict = { "input_ids": dataset['train']["input_ids"], "attention_mask": dataset['train']["attention_mask"], "token_type_ids": dataset['train']["token_type_ids"], "label": dataset['train']["label"] } valid_dict = { "input_ids": dataset['validation']["input_ids"][valid_idx], "attention_mask": dataset['validation']["attention_mask"][valid_idx], "token_type_ids": dataset['validation']["token_type_ids"][valid_idx], "label": dataset['validation']["label"][valid_idx] } test_dict = { "input_ids": dataset['validation']["input_ids"][test_idx], "attention_mask": dataset['validation']["attention_mask"][test_idx], "token_type_ids": dataset['validation']["token_type_ids"][test_idx], "label": dataset['validation']["label"][test_idx] } # Instantiate a PyTorch Dataloader around our dataset train_loader = torch.utils.data.DataLoader( SST2Dataset(train_dict), batch_size=config_params["train_batch_size"], shuffle=True) valid_loader = torch.utils.data.DataLoader( SST2Dataset(valid_dict), batch_size=config_params["train_batch_size"], drop_last=False) test_loader = torch.utils.data.DataLoader( SST2Dataset(test_dict), batch_size=config_params["train_batch_size"], drop_last=False) model = BertForSequenceClassification.from_pretrained(arch) # torch.nn.init.kaiming_normal_(model.classifier.weight) # torch.nn.init.constant_(model.classifier.bias, 0) # torch.nn.init.kaiming_normal_(model.bert.pooler.dense.weight) # torch.nn.init.constant_(model.bert.pooler.dense.bias, 0); args = Object() setattr(args, "local_rank", 0) setattr(args, "deepspeed_config", config) if config[:3] == "cpu": if "optimizer" in config_params: model, optimizer, _, _ = deepspeed.initialize( args=args, model=model, model_parameters=model.parameters() ) else: from deepspeed.ops.adam import DeepSpeedCPUAdam optimizer = DeepSpeedCPUAdam(model.parameters(), lr=2e-5) model, optimizer, _, _ = deepspeed.initialize( args=args, model=model, model_parameters=model.parameters(), optimizer=optimizer ) else: model, optimizer, _, _ = deepspeed.initialize( args=args, model=model, model_parameters=model.parameters() # optimizer=optimizer ) total_steps = len(train_loader) * 3 # checkpoints = CheckpointCallback( # keep_n_checkpoints=1, # checkpoint_dir=CACHE_DIR / "model_cache/", # monitor_metric="accuracy" # ) lr_durations = [ int(total_steps*0.2), int(np.ceil(total_steps*0.8)) ] break_points = [0] + list(np.cumsum(lr_durations))[:-1] callbacks = [ MovingAverageStatsTrackerCallback( avg_window=len(train_loader) // 8, log_interval=len(train_loader) // 10 ), LearningRateSchedulerCallback( MultiStageScheduler( [ LinearLR(optimizer, 0.01, lr_durations[0]), CosineAnnealingScheduler(optimizer, lr_durations[1]) ], start_at_epochs=break_points ) ), # checkpoints ] bot = SST2Bot( model=model, train_loader=train_loader, valid_loader=valid_loader, clip_grad=10., optimizer=optimizer, echo=True, criterion=torch.nn.CrossEntropyLoss(), callbacks=callbacks, pbar=False, use_tensorboard=False, # use_amp=APEX_AVAILABLE, metrics=(Top1Accuracy(),) ) print(total_steps) bot.train( total_steps=total_steps, checkpoint_interval=len(train_loader) // 2 ) # bot.load_model(checkpoints.best_performers[0][1]) # checkpoints.remove_checkpoints(keep=0) # TARGET_DIR = CACHE_DIR / "sst2_bert_uncased" # TARGET_DIR.mkdir(exist_ok=True) # bot.model.save_pretrained(TARGET_DIR) bot.eval(valid_loader) bot.eval(test_loader) if __name__ == "__main__": APP()
7,103
2,321
# -*- coding: utf-8 -*- """ Auto Encoder Example. Using an auto encoder on MNIST handwritten digits. References: Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. "Gradient-based learning applied to document recognition." Proceedings of the IEEE, 86(11):2278-2324, November 1998. Links: [MNIST Dataset] http://yann.lecun.com/exdb/mnist/ """ from __future__ import division, print_function, absolute_import import numpy as np import matplotlib.pyplot as plt import tflearn # Data loading and preprocessing import tflearn.datasets.mnist as mnist X, Y, testX, testY = mnist.load_data(one_hot=True) # Building the encoder encoder = tflearn.input_data(shape=[None, 784]) encoder = tflearn.fully_connected(encoder, 256) encoder = tflearn.fully_connected(encoder, 64) # Building the decoder decoder = tflearn.fully_connected(encoder, 256) decoder = tflearn.fully_connected(decoder, 784, activation='sigmoid') # Regression, with mean square error net = tflearn.regression(decoder, optimizer='adam', learning_rate=0.001, loss='mean_square', metric=None) # Training the auto encoder model = tflearn.DNN(net, tensorboard_verbose=0) model.fit(X, X, n_epoch=20, validation_set=(testX, testX), run_id="auto_encoder", batch_size=256) # Encoding X[0] for test print("\nTest encoding of X[0]:") # New model, re-using the same session, for weights sharing encoding_model = tflearn.DNN(encoder, session=model.session) print(encoding_model.predict([X[0]])) # Testing the image reconstruction on new data (test set) print("\nVisualizing results after being encoded and decoded:") testX = tflearn.data_utils.shuffle(testX)[0] # Applying encode and decode over test set encode_decode = model.predict(testX) # Compare original images with their reconstructions f, a = plt.subplots(2, 10, figsize=(10, 2)) for i in range(10): temp = [[ii, ii, ii] for ii in list(testX[i])] a[0][i].imshow(np.reshape(temp, (28, 28, 3))) temp = [[ii, ii, ii] for ii in list(encode_decode[i])] a[1][i].imshow(np.reshape(temp, (28, 28, 3))) f.show() plt.draw() plt.waitforbuttonpress()
2,118
780