text
string
size
int64
token_count
int64
# --------------------------------------------------------------------- # Vendor: Cisco # OS: SCOS # --------------------------------------------------------------------- # Copyright (C) 2007-2012 The NOC Project # See LICENSE for details # --------------------------------------------------------------------- # NOC modules from noc.core.profile.base import BaseProfile class Profile(BaseProfile): name = "Cisco.SCOS" pattern_more = [(r"--More--", " "), (r"\?\s*\[confirm\]", "\n")] pattern_unprivileged_prompt = r"^\S+?>" pattern_syntax_error = r"% invalid input |% Ambiguous command:|% Incomplete command." # command_disable_pager = "terminal length 0" command_super = "enable" command_enter_config = "configure" command_leave_config = "exit" command_save_config = "copy running-config startup-config\n" pattern_prompt = r"^(?P<hostname>[a-zA-Z0-9]\S*?)(?:-\d+)?(?:\(config[^\)]*\))?#" requires_netmask_conversion = True convert_mac = BaseProfile.convert_mac_to_cisco def convert_interface_name(self, interface): if interface.startswith("Fast"): return "Fa " + interface[12:].strip() elif interface.startswith("Giga"): return "Gi " + interface[15:].strip() elif interface.startswith("Ten"): return "Te " + interface[18:].strip() else: return interface
1,398
432
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. """ This DAG will not work unless you create an Amazon EMR cluster running Apache Hive and copy data into it following steps 1-4 (inclusive) here: https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EMRforDynamoDB.Tutorial.html """ import os from datetime import datetime from airflow import DAG from airflow.decorators import task from airflow.models import Connection from airflow.providers.amazon.aws.hooks.dynamodb import DynamoDBHook from airflow.providers.amazon.aws.transfers.hive_to_dynamodb import HiveToDynamoDBOperator from airflow.utils import db DYNAMODB_TABLE_NAME = 'example_hive_to_dynamodb_table' HIVE_CONNECTION_ID = os.getenv('HIVE_CONNECTION_ID', 'hive_on_emr') HIVE_HOSTNAME = os.getenv('HIVE_HOSTNAME', 'ec2-123-45-67-890.compute-1.amazonaws.com') # These values assume you set up the Hive data source following the link above. DYNAMODB_TABLE_HASH_KEY = 'feature_id' HIVE_SQL = 'SELECT feature_id, feature_name, feature_class, state_alpha FROM hive_features' @task def create_dynamodb_table(): client = DynamoDBHook(client_type='dynamodb').conn client.create_table( TableName=DYNAMODB_TABLE_NAME, KeySchema=[ {'AttributeName': DYNAMODB_TABLE_HASH_KEY, 'KeyType': 'HASH'}, ], AttributeDefinitions=[ {'AttributeName': DYNAMODB_TABLE_HASH_KEY, 'AttributeType': 'N'}, ], ProvisionedThroughput={'ReadCapacityUnits': 20, 'WriteCapacityUnits': 20}, ) # DynamoDB table creation is nearly, but not quite, instantaneous. # Wait for the table to be active to avoid race conditions writing to it. waiter = client.get_waiter('table_exists') waiter.wait(TableName=DYNAMODB_TABLE_NAME, WaiterConfig={'Delay': 1}) @task def get_dynamodb_item_count(): """ A DynamoDB table has an ItemCount value, but it is only updated every six hours. To verify this DAG worked, we will scan the table and count the items manually. """ table = DynamoDBHook(resource_type='dynamodb').conn.Table(DYNAMODB_TABLE_NAME) response = table.scan(Select='COUNT') item_count = response['Count'] while 'LastEvaluatedKey' in response: response = table.scan(Select='COUNT', ExclusiveStartKey=response['LastEvaluatedKey']) item_count += response['Count'] print(f'DynamoDB table contains {item_count} items.') # Included for sample purposes only; in production you wouldn't delete # the table you just backed your data up to. Using 'all_done' so even # if an intermediate step fails, the DAG will clean up after itself. @task(trigger_rule='all_done') def delete_dynamodb_table(): DynamoDBHook(client_type='dynamodb').conn.delete_table(TableName=DYNAMODB_TABLE_NAME) # Included for sample purposes only; in production this should # be configured in the environment and not be part of the DAG. # Note: The 'hiveserver2_default' connection will not work if Hive # is hosted on EMR. You must set the host name of the connection # to match your EMR cluster's hostname. @task def configure_hive_connection(): db.merge_conn( Connection( conn_id=HIVE_CONNECTION_ID, conn_type='hiveserver2', host=HIVE_HOSTNAME, port=10000, ) ) with DAG( dag_id='example_hive_to_dynamodb', schedule_interval=None, start_date=datetime(2021, 1, 1), tags=['example'], catchup=False, ) as dag: # Add the prerequisites docstring to the DAG in the UI. dag.doc_md = __doc__ # [START howto_transfer_hive_to_dynamodb] backup_to_dynamodb = HiveToDynamoDBOperator( task_id='backup_to_dynamodb', hiveserver2_conn_id=HIVE_CONNECTION_ID, sql=HIVE_SQL, table_name=DYNAMODB_TABLE_NAME, table_keys=[DYNAMODB_TABLE_HASH_KEY], ) # [END howto_transfer_hive_to_dynamodb] ( configure_hive_connection() >> create_dynamodb_table() >> backup_to_dynamodb >> get_dynamodb_item_count() >> delete_dynamodb_table() )
4,825
1,579
__version__ = "0.1.0" from tle_download.main import get_tles, write_tles
74
32
import numpy as np import random def align(pair,sequences,scoring_matrix,penalties,aa_dict): seq1, seq2 = sequences[pair[0]], sequences[pair[1]] #Create matrix F (containing maximum scores up to cell F(i,j)) and tracking matrix P (containing the tracking flag describing whether the step into the cell was a match or a gap, or if the cell is the end of a local alignment. F = np.zeros((len(seq1),len(seq2))) P = np.zeros((len(seq1),len(seq2))) F, P = fill_scoring_matrix(F,P,seq1,seq2,scoring_matrix,aa_dict,penalties) return traceback(F,P,seq1,seq2) def fill_scoring_matrix(F,P,seq1,seq2,scoring_matrix,aa_dict,penalties): d,e = penalties it = np.nditer(F, flags=['multi_index']) while not it.finished: i,j = it.multi_index aa1_index = aa_dict[seq1[i]] aa2_index = aa_dict[seq2[j]] #Set penalties: opening a gap has penalty d, extending a gap has penalty e, and we do not allow a gap to be opened introduced in one sequence after a gap has occured in the other (because this would effectively allow for jumps from local alignment to local alignment). #If the cell directly above was arrived at by introducing a gap in seq1, then do not allow a gap in seq2 if P[i-1,j] == 1: gap_in_seq1,gap_in_seq2 = e,float("inf") #If the cell directly to the left was arrived at by introducing a gap in seq2, then do not allow a gap in seq1 elif P[i,j-1] == -1: gap_in_seq1,gap_in_seq2 = float("inf"),e else: gap_in_seq1,gap_in_seq2 = d,d possible_scores = [0, F[i-1,j-1]+scoring_matrix[aa1_index][aa2_index], F[i,j-1]-gap_in_seq1, F[i-1,j]-gap_in_seq2 ] #Find maximum of possible_scores to decide on move F[i,j] = max(possible_scores) if F[i,j] == possible_scores[1]: P[i,j] = 0 #If match, set tracking flag to 0 elif F[i,j] == possible_scores[2]: P[i,j] = 1 #If gap in seq1, set tracking flag to 1 else : P[i,j] = -1 #If gap in seq2 (or if max was 0, and alignment ended), set tracking flag to -1 it.iternext() F[:,0] = 0 F[0,:] = 0 return F,P def traceback(F,P,seq1,seq2): #Find max value max_index = F.argmax() i,j = np.unravel_index(max_index, F.shape) score = F[i,j] alignment = ('','') while F[i,j] > 0 and i is not 0 and j is not 0: if P[i,j] == 0: #There must have been a match of the amino acid at seq1[i] and seq2[j] alignment = seq1[i]+alignment[0],seq2[j]+alignment[1] i -= 1 j -= 1 elif P[i,j] == 1: #There must have been a gap in seq1 alignment = "_"+alignment[0],seq2[j]+alignment[1] j -= 1 elif P[i,j] == -1: #There must have been a gap in seq2 alignment = seq1[i]+alignment[0],"_"+alignment[1] i -= 1 return alignment, score def get_fp_rate(pos_scores,neg_scores): #We want to set our threshold as the value of the element in the pos_scores list that is greater than or equal to exactly 30% of the scores in the list. Thus, we sort the pos_scores, and take the value of the 15th element out of the 50 in the list to be the threshold. pos_scores.sort() threshold = pos_scores[14] fp = 0 for score in neg_scores: if score > threshold: fp+=1 fp_rate = fp/50.0 return threshold, fp_rate def get_tp_rate(pos_scores,neg_scores,fpr): neg_scores.sort() index = int(50*(1-fpr))-1 threshold = neg_scores[index] tp = 0 for score in pos_scores: if score > threshold: tp+=1 tp_rate = tp/50.0 return tp_rate def calc_score(alignments,scoremat,penalties,aa_dict): d,e = penalties scores = [] for alignment in alignments: score = 0 for index in range(len(alignment[0])): score_flag = False aa1 = alignment[0][index] aa2 = alignment[1][index] if aa1 == '_': if ext_flag == True: score -= e score_flag = True else: score -= d ext_flag = True score_flag = True else: i = aa_dict[aa1] if aa2 == '_': if ext_flag == True: score -= e score_flag = True else: score -= d ext_flag = True score_flag = True else: j = aa_dict[aa2] if (score_flag == False): score += scoremat[i,j] ext_flag = False scores.append(score) return scores
4,125
1,822
"""Support for the Netatmo climate schedule selector.""" from __future__ import annotations import logging from homeassistant.components.select import SelectEntity from homeassistant.config_entries import ConfigEntry from homeassistant.core import HomeAssistant, callback from homeassistant.helpers.dispatcher import async_dispatcher_connect from homeassistant.helpers.entity_platform import AddEntitiesCallback from .const import ( CONF_URL_ENERGY, DATA_SCHEDULES, DOMAIN, EVENT_TYPE_SCHEDULE, MANUFACTURER, NETATMO_CREATE_SELECT, ) from .data_handler import HOME, SIGNAL_NAME, NetatmoHome from .netatmo_entity_base import NetatmoBase _LOGGER = logging.getLogger(__name__) async def async_setup_entry( hass: HomeAssistant, entry: ConfigEntry, async_add_entities: AddEntitiesCallback ) -> None: """Set up the Netatmo energy platform schedule selector.""" print("Setup of select platform") @callback def _create_entity(netatmo_home: NetatmoHome) -> None: entity = NetatmoScheduleSelect(netatmo_home) _LOGGER.debug("Adding climate select %s", entity) async_add_entities([entity]) entry.async_on_unload( async_dispatcher_connect(hass, NETATMO_CREATE_SELECT, _create_entity) ) class NetatmoScheduleSelect(NetatmoBase, SelectEntity): """Representation a Netatmo thermostat schedule selector.""" def __init__( self, netatmo_home: NetatmoHome, ) -> None: """Initialize the select entity.""" print("NetatmoScheduleSelect", netatmo_home.signal_name) SelectEntity.__init__(self) super().__init__(netatmo_home.data_handler) self._home = netatmo_home.home self._home_id = self._home.entity_id self._signal_name = netatmo_home.signal_name self._publishers.extend( [ { "name": HOME, "home_id": self._home.entity_id, SIGNAL_NAME: self._signal_name, }, ] ) self._device_name = self._home.name self._attr_name = f"{MANUFACTURER} {self._device_name}" self._model: str = "NATherm1" self._config_url = CONF_URL_ENERGY self._attr_unique_id = f"{self._home_id}-schedule-select" self._attr_current_option = getattr(self._home.get_selected_schedule(), "name") self._attr_options = [ schedule.name for schedule in self._home.schedules.values() ] async def async_added_to_hass(self) -> None: """Entity created.""" await super().async_added_to_hass() self.data_handler.config_entry.async_on_unload( async_dispatcher_connect( self.hass, f"signal-{DOMAIN}-webhook-{EVENT_TYPE_SCHEDULE}", self.handle_event, ) ) @callback def handle_event(self, event: dict) -> None: """Handle webhook events.""" data = event["data"] if self._home_id != data["home_id"]: return if data["event_type"] == EVENT_TYPE_SCHEDULE and "schedule_id" in data: self._attr_current_option = getattr( self.hass.data[DOMAIN][DATA_SCHEDULES][self._home_id].get( data["schedule_id"] ), "name", ) self.async_write_ha_state() async def async_select_option(self, option: str) -> None: """Change the selected option.""" for sid, schedule in self.hass.data[DOMAIN][DATA_SCHEDULES][ self._home_id ].items(): if schedule.name != option: continue _LOGGER.debug( "Setting %s schedule to %s (%s)", self._home_id, option, sid, ) await self._home.async_switch_schedule(schedule_id=sid) break @callback def async_update_callback(self) -> None: """Update the entity's state.""" self._attr_current_option = getattr(self._home.get_selected_schedule(), "name") self.hass.data[DOMAIN][DATA_SCHEDULES][self._home_id] = self._home.schedules self._attr_options = [ schedule.name for schedule in self._home.schedules.values() ]
4,346
1,330
# Ivan Carvalho # Solution to https://www.urionlinejudge.com.br/judge/problems/view/1943 #!/usr/bin/env python2.7 # encoding : utf-8 e = int(raw_input()) if e == 1: print "Top 1" elif e <= 3: print "Top 3" elif e <= 5: print "Top 5" elif e <= 10: print "Top 10" elif e <= 25: print "Top 25" elif e <= 50: print "Top 50" else: print "Top 100"
349
174
from dataclasses import dataclass, field from typing import Optional __NAMESPACE__ = "sdformat/v1.5/physics.xsd" @dataclass class Physics: """ The physics tag specifies the type and properties of the dynamics engine. Parameters ---------- max_step_size: Maximum time step size at which every system in simulation can interact with the states of the world. (was physics.sdf's dt). real_time_factor: target simulation speedup factor, defined by ratio of simulation time to real-time. real_time_update_rate: Rate at which to update the physics engine (UpdatePhysics calls per real-time second). (was physics.sdf's update_rate). max_contacts: Maximum number of contacts allowed between two entities. This value can be over ridden by a max_contacts element in a collision element. gravity: The gravity vector in m/s^2, expressed in a coordinate frame defined by the spherical_coordinates tag. magnetic_field: The magnetic vector in Tesla, expressed in a coordinate frame defined by the spherical_coordinates tag. simbody: Simbody specific physics properties bullet: Bullet specific physics properties ode: ODE specific physics properties name: The name of this set of physics parameters. default: If true, this physics element is set as the default physics profile for the world. If multiple default physics elements exist, the first element marked as default is chosen. If no default physics element exists, the first physics element is chosen. type: The type of the dynamics engine. Current options are ode, bullet, simbody and rtql8. Defaults to ode if left unspecified. """ class Meta: name = "physics" max_step_size: float = field( default=0.001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) real_time_factor: float = field( default=1.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) real_time_update_rate: float = field( default=1000.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) max_contacts: int = field( default=20, metadata={ "type": "Element", "namespace": "", "required": True, }, ) gravity: str = field( default="0 0 -9.8", metadata={ "type": "Element", "namespace": "", "required": True, "pattern": r"(\s*(-|\+)?(\d+(\.\d*)?|\.\d+|\d+\.\d+[eE][-\+]?[0-9]+)\s+){2}((-|\+)?(\d+(\.\d*)?|\.\d+|\d+\.\d+[eE][-\+]?[0-9]+))\s*", }, ) magnetic_field: str = field( default="5.5645e-6 22.8758e-6 -42.3884e-6", metadata={ "type": "Element", "namespace": "", "required": True, "pattern": r"(\s*(-|\+)?(\d+(\.\d*)?|\.\d+|\d+\.\d+[eE][-\+]?[0-9]+)\s+){2}((-|\+)?(\d+(\.\d*)?|\.\d+|\d+\.\d+[eE][-\+]?[0-9]+))\s*", }, ) simbody: Optional["Physics.Simbody"] = field( default=None, metadata={ "type": "Element", "namespace": "", }, ) bullet: Optional["Physics.Bullet"] = field( default=None, metadata={ "type": "Element", "namespace": "", }, ) ode: Optional["Physics.Ode"] = field( default=None, metadata={ "type": "Element", "namespace": "", }, ) name: str = field( default="default_physics", metadata={ "type": "Attribute", }, ) default: bool = field( default=False, metadata={ "type": "Attribute", }, ) type: Optional[str] = field( default=None, metadata={ "type": "Attribute", "required": True, }, ) @dataclass class Simbody: """ Simbody specific physics properties. Parameters ---------- min_step_size: (Currently not used in simbody) The time duration which advances with each iteration of the dynamics engine, this has to be no bigger than max_step_size under physics block. If left unspecified, min_step_size defaults to max_step_size. accuracy: Roughly the relative error of the system. -LOG(accuracy) is roughly the number of significant digits. max_transient_velocity: Tolerable "slip" velocity allowed by the solver when static friction is supposed to hold object in place. contact: Relationship among dissipation, coef. restitution, etc. d = dissipation coefficient (1/velocity) vc = capture velocity (velocity where e=e_max) vp = plastic velocity (smallest v where e=e_min) &amp;gt; vc Assume real COR=1 when v=0. e_min = given minimum COR, at v &amp;gt;= vp (a.k.a. plastic_coef_restitution) d = slope = (1-e_min)/vp OR, e_min = 1 - d*vp e_max = maximum COR = 1-d*vc, reached at v=vc e = 0, v &amp;lt;= vc = 1 - d*v, vc &amp;lt; v &amp;lt; vp = e_min, v &amp;gt;= vp dissipation factor = d*min(v,vp) [compliant] cor = e [rigid] Combining rule e = 0, e1==e2==0 = 2*e1*e2/(e1+e2), otherwise """ min_step_size: float = field( default=0.0001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) accuracy: float = field( default=0.001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) max_transient_velocity: float = field( default=0.01, metadata={ "type": "Element", "namespace": "", "required": True, }, ) contact: Optional["Physics.Simbody.Contact"] = field( default=None, metadata={ "type": "Element", "namespace": "", }, ) @dataclass class Contact: """Relationship among dissipation, coef. restitution, etc. d = dissipation coefficient (1/velocity) vc = capture velocity (velocity where e=e_max) vp = plastic velocity (smallest v where e=e_min) &amp;gt; vc Assume real COR=1 when v=0. e_min = given minimum COR, at v &amp;gt;= vp (a.k.a. plastic_coef_restitution) d = slope = (1-e_min)/vp OR, e_min = 1 - d*vp e_max = maximum COR = 1-d*vc, reached at v=vc e = 0, v &amp;lt;= vc = 1 - d*v, vc &amp;lt; v &amp;lt; vp = e_min, v &amp;gt;= vp dissipation factor = d*min(v,vp) [compliant] cor = e [rigid] Combining rule e = 0, e1==e2==0 = 2*e1*e2/(e1+e2), otherwise Parameters ---------- stiffness: Default contact material stiffness (force/dist or torque/radian). dissipation: dissipation coefficient to be used in compliant contact; if not given it is (1-min_cor)/plastic_impact_velocity plastic_coef_restitution: this is the COR to be used at high velocities for rigid impacts; if not given it is 1 - dissipation*plastic_impact_velocity plastic_impact_velocity: smallest impact velocity at which min COR is reached; set to zero if you want the min COR always to be used static_friction: static friction (mu_s) as described by this plot: http://gazebosim.org/wiki/File:Stribeck_friction.png dynamic_friction: dynamic friction (mu_d) as described by this plot: http://gazebosim.org/wiki/File:Stribeck_friction.png viscous_friction: viscous friction (mu_v) with units of (1/velocity) as described by this plot: http://gazebosim.org/wiki/File:Stribeck_friction.png override_impact_capture_velocity: for rigid impacts only, impact velocity at which COR is set to zero; normally inherited from global default but can be overridden here. Combining rule: use larger velocity override_stiction_transition_velocity: This is the largest slip velocity at which we'll consider a transition to stiction. Normally inherited from a global default setting. For a continuous friction model this is the velocity at which the max static friction force is reached. Combining rule: use larger velocity """ stiffness: float = field( default=100000000.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) dissipation: float = field( default=100.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) plastic_coef_restitution: float = field( default=0.5, metadata={ "type": "Element", "namespace": "", "required": True, }, ) plastic_impact_velocity: float = field( default=0.5, metadata={ "type": "Element", "namespace": "", "required": True, }, ) static_friction: float = field( default=0.9, metadata={ "type": "Element", "namespace": "", "required": True, }, ) dynamic_friction: float = field( default=0.9, metadata={ "type": "Element", "namespace": "", "required": True, }, ) viscous_friction: float = field( default=0.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) override_impact_capture_velocity: float = field( default=0.001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) override_stiction_transition_velocity: float = field( default=0.001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) @dataclass class Bullet: """ Bullet specific physics properties. Parameters ---------- solver: constraints: Bullet constraint parameters. """ solver: Optional["Physics.Bullet.Solver"] = field( default=None, metadata={ "type": "Element", "namespace": "", "required": True, }, ) constraints: Optional["Physics.Bullet.Constraints"] = field( default=None, metadata={ "type": "Element", "namespace": "", "required": True, }, ) @dataclass class Solver: """ Parameters ---------- type: One of the following types: sequential_impulse only. min_step_size: The time duration which advances with each iteration of the dynamics engine, this has to be no bigger than max_step_size under physics block. If left unspecified, min_step_size defaults to max_step_size. iters: Number of iterations for each step. A higher number produces greater accuracy at a performance cost. sor: Set the successive over-relaxation parameter. """ type: str = field( default="sequential_impulse", metadata={ "type": "Element", "namespace": "", "required": True, }, ) min_step_size: float = field( default=0.0001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) iters: int = field( default=50, metadata={ "type": "Element", "namespace": "", "required": True, }, ) sor: float = field( default=1.3, metadata={ "type": "Element", "namespace": "", "required": True, }, ) @dataclass class Constraints: """ Bullet constraint parameters. Parameters ---------- cfm: Constraint force mixing parameter. See the ODE page for more information. erp: Error reduction parameter. See the ODE page for more information. contact_surface_layer: The depth of the surface layer around all geometry objects. Contacts are allowed to sink into the surface layer up to the given depth before coming to rest. The default value is zero. Increasing this to some small value (e.g. 0.001) can help prevent jittering problems due to contacts being repeatedly made and broken. split_impulse: Similar to ODE's max_vel implementation. See http://web.archive.org/web/20120430155635/http://bulletphysics.org/mediawiki-1.5.8/index.php/BtContactSolverInfo#Split_Impulse for more information. split_impulse_penetration_threshold: Similar to ODE's max_vel implementation. See http://web.archive.org/web/20120430155635/http://bulletphysics.org/mediawiki-1.5.8/index.php/BtContactSolverInfo#Split_Impulse for more information. """ cfm: float = field( default=0.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) erp: float = field( default=0.2, metadata={ "type": "Element", "namespace": "", "required": True, }, ) contact_surface_layer: float = field( default=0.001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) split_impulse: bool = field( default=True, metadata={ "type": "Element", "namespace": "", "required": True, }, ) split_impulse_penetration_threshold: float = field( default=-0.01, metadata={ "type": "Element", "namespace": "", "required": True, }, ) @dataclass class Ode: """ ODE specific physics properties. Parameters ---------- solver: constraints: ODE constraint parameters. """ solver: Optional["Physics.Ode.Solver"] = field( default=None, metadata={ "type": "Element", "namespace": "", "required": True, }, ) constraints: Optional["Physics.Ode.Constraints"] = field( default=None, metadata={ "type": "Element", "namespace": "", "required": True, }, ) @dataclass class Solver: """ Parameters ---------- type: One of the following types: world, quick min_step_size: The time duration which advances with each iteration of the dynamics engine, this has to be no bigger than max_step_size under physics block. If left unspecified, min_step_size defaults to max_step_size. iters: Number of iterations for each step. A higher number produces greater accuracy at a performance cost. precon_iters: Experimental parameter. sor: Set the successive over-relaxation parameter. use_dynamic_moi_rescaling: Flag to enable dynamic rescaling of moment of inertia in constrained directions. See gazebo pull request 1114 for the implementation of this feature. https://osrf- migration.github.io/gazebo-gh-pages/#!/osrf/gazebo/pull- request/1114 """ type: str = field( default="quick", metadata={ "type": "Element", "namespace": "", "required": True, }, ) min_step_size: float = field( default=0.0001, metadata={ "type": "Element", "namespace": "", "required": True, }, ) iters: int = field( default=50, metadata={ "type": "Element", "namespace": "", "required": True, }, ) precon_iters: int = field( default=0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) sor: float = field( default=1.3, metadata={ "type": "Element", "namespace": "", "required": True, }, ) use_dynamic_moi_rescaling: bool = field( default=False, metadata={ "type": "Element", "namespace": "", "required": True, }, ) @dataclass class Constraints: """ ODE constraint parameters. Parameters ---------- cfm: Constraint force mixing parameter. See the ODE page for more information. erp: Error reduction parameter. See the ODE page for more information. contact_max_correcting_vel: The maximum correcting velocities allowed when resolving contacts. contact_surface_layer: The depth of the surface layer around all geometry objects. Contacts are allowed to sink into the surface layer up to the given depth before coming to rest. The default value is zero. Increasing this to some small value (e.g. 0.001) can help prevent jittering problems due to contacts being repeatedly made and broken. """ cfm: float = field( default=0.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) erp: float = field( default=0.2, metadata={ "type": "Element", "namespace": "", "required": True, }, ) contact_max_correcting_vel: float = field( default=100.0, metadata={ "type": "Element", "namespace": "", "required": True, }, ) contact_surface_layer: float = field( default=0.001, metadata={ "type": "Element", "namespace": "", "required": True, }, )
21,889
5,498
# -*- coding: utf-8 -*- # # Copyright 2013 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A calliope command that prints help for another calliope command.""" from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import argparse from googlecloudsdk.calliope import base from googlecloudsdk.command_lib.help_search import search from googlecloudsdk.command_lib.help_search import search_util from googlecloudsdk.core import log _DEFAULT_LIMIT = 5 class Help(base.ListCommand): """Search gcloud help text. If a full gcloud command is specified after the ``help'' operand, {command} prints a detailed help message for that command. Otherwise, {command} runs a search for all commands with help text matching the given argument or arguments. It prints the command name and a summary of the help text for any command that it finds as a result. To run a search directly, you can use remainder arguments, following a `--`. By default, command results are displayed in a table that shows the name of the command and snippets of the help text that relate to your search terms. By default, search results are sorted from most to least relevant by default, using a localized rating based on several heuristics. These heuristics may change in future runs of this command. ## EXAMPLES To get the help for the command `gcloud projects describe`, run: $ {command} projects describe To search for all commands whose help text contains the word `project`, run: $ {command} -- project To search for commands whose help text contains the word `project` and the string `--foo`, run: $ {command} -- project --foo To search and receive more than the default limit of 5 search results, run: $ {command} --limit=20 -- project To search for a term and sort the results by a different characteristic, such as command name, run: $ {command} --sort-by=name -- project """ category = base.GCLOUD_SDK_TOOLS_CATEGORY @staticmethod def Args(parser): parser.display_info.AddTransforms(search_util.GetTransforms()) parser.display_info.AddFormat(""" table[all-box,pager]( commandpath():label='COMMAND', summary():wrap) """) base.URI_FLAG.RemoveFromParser(parser) base.LIMIT_FLAG.SetDefault(parser, _DEFAULT_LIMIT) base.SORT_BY_FLAG.SetDefault(parser, '~relevance') parser.add_argument( 'command', nargs='*', help="""\ Sequence of names representing a gcloud group or command name. If the arguments provide the name of a gcloud command, the full help text of that command will be displayed. Otherwise, all arguments will be considered search terms and used to search through all of gcloud's help text. """) parser.add_argument( 'search_terms', nargs=argparse.REMAINDER, help="""\ Search terms. The command will return a list of gcloud commands that are relevant to the searched term. If this argument is provided, the command will always return a list of search results rather than displaying help text of a single command. For example, to search for commands that relate to the term `project` or `folder`, run: $ {command} -- project folder """) def Run(self, args): if not args.search_terms: try: # --document=style=help to signal the metrics.Help() 'help' label in # actions.RenderDocumentAction().Action(). self.ExecuteCommandDoNotUse(args.command + ['--document=style=help']) return None except Exception: # pylint: disable=broad-except # In this case, we will treat the arguments as search terms. pass results = search.RunSearch( args.command + (args.search_terms or []), self._cli_power_users_only) self._resources_found = len(results) self._resources_displayed = min(len(results), args.limit) return results def Epilog(self, resources_were_displayed): if not self._resources_found: return if resources_were_displayed: log.status.Print( 'Listed {} of {} items.'.format(self._resources_displayed, self._resources_found)) else: log.status.Print('Listed 0 items.')
4,826
1,348
from .main import main as run # noqa: F401 __version__ = '0.0.3'
66
30
# Copyright (c) OpenMMLab. All rights reserved. from .darts_backbone import DartsBackbone from .searchable_mobilenet import SearchableMobileNet from .searchable_shufflenet_v2 import SearchableShuffleNetV2 __all__ = ['DartsBackbone', 'SearchableShuffleNetV2', 'SearchableMobileNet']
283
93
#!/usr/bin/env python # Copyright 2018-present Facebook, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from __future__ import absolute_import, division, print_function, unicode_literals import os import shutil import subprocess # The location of the generate grammar kit script DIR = os.path.dirname(__file__) # The location of the plugin directory PLUGIN_PATH = os.path.abspath(os.path.join(DIR, "..")) # The location of the grammar-kit directory GRAMMAR_KIT = os.path.abspath( os.path.join(DIR, "../../../third-party/java/grammar-kit/") ) OUT_DIR = os.path.join(PLUGIN_PATH, "gen") FLEX_OUT_DIR = os.path.join(OUT_DIR, "com/facebook/buck/intellij/ideabuck/lang") GRAMMAR_KIT_JAR = os.path.join(GRAMMAR_KIT, "grammar-kit.jar") GRAMMAR_KIT_JFLEX_JAR = os.path.join(GRAMMAR_KIT, "JFlex.jar") JFLEX_SKELETON = os.path.join(PLUGIN_PATH, "resources/idea-flex.skeleton") FLEX_FILE = os.path.join( PLUGIN_PATH, "src/com/facebook/buck/intellij/ideabuck/lang/Buck.flex" ) BNF_FILE = os.path.join( PLUGIN_PATH, "src/com/facebook/buck/intellij/ideabuck/lang/Buck.bnf" ) def subprocess_call(cmd): print("Running: %s" % (" ".join(cmd))) subprocess.call(cmd) shutil.rmtree(OUT_DIR, ignore_errors=True) subprocess_call(["java", "-jar", GRAMMAR_KIT_JAR, OUT_DIR, BNF_FILE]) subprocess_call( [ "java", "-jar", GRAMMAR_KIT_JFLEX_JAR, "-sliceandcharat", "-skel", JFLEX_SKELETON, "-d", FLEX_OUT_DIR, FLEX_FILE, ] )
2,013
762
from lxml import html from d_parser.d_spider_common import DSpiderCommon from d_parser.helpers.re_set import Ree from helpers.url_generator import UrlGenerator from d_parser.helpers.stat_counter import StatCounter as SC VERSION = 29 # Warn: Don't remove task argument even if not use it (it's break grab and spider crashed) # Warn: noinspection PyUnusedLocal class DSpider(DSpiderCommon): def __init__(self, thread_number, try_limit=0): super().__init__(thread_number, try_limit) # parse categories def task_initial(self, grab, task): try: if self.check_body_errors(grab, task): yield self.check_errors(task) return links = grab.doc.select('//div[@id="main-subitems"]//a') for link in links: url = UrlGenerator.get_page_params(self.domain, link.attr('href'), {'onpage': 99999}) yield self.do_task('parse_page', url, DSpider.get_next_task_priority(task)) except Exception as e: self.process_error(grab, task, e) finally: self.process_finally(task) # parse page def task_parse_page(self, grab, task): try: if self.check_body_errors(grab, task): yield self.check_errors(task) # parse items links items_links = grab.doc.select('//div[@id="catalog-list"]//div[@class="catalog-items"]//a[@property="name"]') for row in items_links: link = row.attr('href') link = UrlGenerator.get_page_params(self.domain, link, {}) yield self.do_task('parse_item', link, DSpider.get_next_task_priority(task)) except Exception as e: self.process_error(grab, task, e) finally: self.process_finally(task) # parse single item def task_parse_item(self, grab, task): try: if self.check_body_errors(grab, task): yield self.check_errors(task) # common block with info product_info = grab.doc.select('//div[@id="product-info"]') # parse fields # A = name product_name = product_info.select('.//h1').text() # B = [const] # C = [const] # D = [const] product_count_string = product_info.select('.//div[@class="product-data-storehouse"]').text(default='[not found]') product_count = '-1' product_status = '0' product_unit = 'ед.' if product_count_string != 'в наличии': self.log.warning(task, 'Skip item, cuz wrong count {}'.format(product_count_string)) return # E = price # if E = "запросить цену и наличие" => zapros # else => float product_price = product_info.select('.//span[@itemprop="price"]').text().replace(' ', '') if product_price == 'Уточняйте': product_price = '-1' else: # E = price (float) # check if correct price if not Ree.float.match(product_price): self.log_warn(SC.MSG_UNKNOWN_PRICE, f'Skip item, cuz wrong price {product_price}', task) return # F = vendor code product_vendor_code = product_info.select('.//div[@class="product-data-articul"]').text() # G = vendor product_vendor = product_info.select('.//div[@class="product-data-producer"]').text() # H = photo url product_photo_url_raw = product_info.select('.//div[@id="product-images-list"]/div[1]/img[@itemprop="contentUrl"]').attr('src') product_photo_url = UrlGenerator.get_page_params(self.domain, product_photo_url_raw, {}) # pre I product_description_part_raw = product_info.select('.//div[@class="product-description description"]/following-sibling::node()[2]')\ .text(default='')\ .replace('$(".description").html(\'', '')\ .replace('\');', '') # I = description # this part insert pure html with js, so we need clear all html tags and &-symbols product_description_part_list = html.fromstring(f'<div>{product_description_part_raw}</div>').xpath('string()') product_description_part = '' for row in product_description_part_list: product_description_part += row product_description = {'Описание': product_description_part} table = product_info.select('.//div[@class="product-description table"]/div') for row in table: key = row.select('./text()').text() value = row.select('./span').text() if key: product_description[key] = value # ID product_id = product_info.select('.//div[@class="product-add-but"]').attr('data-id', '') # save self.result.add({ 'name': product_name, 'quantity': product_count, 'delivery': product_status, 'measure': product_unit, 'price': product_price, 'sku': product_vendor_code, 'manufacture': product_vendor, 'photo': product_photo_url, 'id': product_id, 'properties': product_description }) except Exception as e: self.process_error(grab, task, e) finally: self.process_finally(task)
5,639
1,593
from tkinter import Label, GROOVE from willsmith.gui_display_controller import GUIDisplayController class MCTSDisplay(GUIDisplayController): """ The display controller for MCTSAgent. Creates a Tkinter GUI that displays some stats about the agent's latest moves. """ WINDOW_TITLE = "MCTS Agent" LABEL_FONT = ("Courier New", 14) LABEL_WIDTH = 25 LABEL_BORDER_WIDTH = 1 LABEL_RELIEF = GROOVE def __init__(self): super().__init__() self.playouts_label = None self.action_label = None self.win_pct_label = None self.depth_label = None def _initialize_widgets(self): self.playouts_label = self._create_label() self.action_label = self._create_label() self.win_pct_label = self._create_label() self.depth_label = self._create_label() def _create_label(self): return Label(self.root, font = self.LABEL_FONT, width = self.LABEL_WIDTH, bd = self.LABEL_BORDER_WIDTH, relief = GROOVE) def _place_widgets(self): self.playouts_label.grid(row = 0, column = 0) self.action_label.grid(row = 1, column = 0) self.win_pct_label.grid(row = 2, column = 0) self.depth_label.grid(row = 3, column = 0) def _update_display(self, agent, action): self._update_labels_from_agent(agent, action) def _reset_display(self, agent): self._update_labels_from_agent(agent, None) def _update_labels_from_agent(self, agent, action): self.action_label["text"] = "Latest action:\n{}".format(action) self.playouts_label["text"] = "Latest playout count:\n{}".format(agent.playout_total) win_pct = 0 if agent.action_node is not None: win_pct = agent.action_node.value_estimate() self.win_pct_label["text"] = "Node sim win rate:\n{:.2%}".format(win_pct) self.depth_label["text"] = "Node tree depth:\n{}".format(agent.root.depth())
2,007
691
#!/usr/bin/python3 import logging import argparse from time import time import toml from data.io.knowledge_graph import KnowledgeGraph from data.io.tarball import Tarball from data.io.tsv import TSV from data.utils import is_readable, is_writable from embeddings import graph_structure from tasks.node_classification import build_dataset, build_model, evaluate_model from tasks.utils import mksplits, init_fold, mkfolds, sample_mask, set_seed, strip_graph def single_run(A, X, Y, X_node_map, tsv_writer, config): tsv_writer.writerow(["epoch", "training_loss", "training_accurary", "validation_loss", "validation_accuracy", "test_loss", "test_accuracy"]) # create splits dataset = mksplits(X, Y, X_node_map, config['task']['dataset_ratio']) # compile model computation graph model = build_model(X, Y, A, config) # train model nepoch = config['model']['epoch'] batch_size = X.shape[0] # number of nodes sample_weights = sample_mask(dataset['train']['X_idx'], Y.shape[0]) for epoch in train_model(A, model, dataset, sample_weights, batch_size, nepoch): # log metrics tsv_writer.writerow([str(epoch[0]), str(epoch[1]), str(epoch[2]), str(epoch[3]), str(epoch[4]), "-1", "-1"]) # test model test_loss, test_acc = test_model(A, model, dataset, batch_size) # log metrics tsv_writer.writerow(["-1", "-1", "-1", "-1", "-1", str(test_loss[0]), str(test_acc[0])]) return (test_loss[0], test_acc[0]) def kfold_crossvalidation(A, X, Y, X_node_map, k, tsv_writer, config): tsv_writer.writerow(["fold", "epoch", "training_loss", "training_accurary", "validation_loss", "validation_accuracy", "test_loss", "test_accuracy"]) # generate fold indices folds_idx = mkfolds(X_node_map.shape[0], k) results = [] logger.info("Starting {}-fold cross validation".format(k)) for fold in range(1, k+1): logger.info("Fold {} / {}".format(fold, k)) # compile model computation graph model = build_model(X, Y, A, config) # initialize fold dataset = init_fold(X, Y, X_node_map, folds_idx[fold-1], config['task']['dataset_ratio']) # train model nepoch = config['model']['epoch'] batch_size = X.shape[0] # number of nodes sample_weights = sample_mask(dataset['train']['X_idx'], Y.shape[0]) for epoch in train_model(A, model, dataset, sample_weights, batch_size, nepoch): # log metrics tsv_writer.writerow([str(fold), str(epoch[0]), str(epoch[1]), str(epoch[2]), str(epoch[3]), str(epoch[4]), "-1", "-1"]) # test model test_loss, test_acc = test_model(A, model, dataset, batch_size) results.append((test_loss[0], test_acc[0])) # log metrics tsv_writer.writerow([str(fold), "-1", "-1", "-1", "-1", "-1", str(test_loss[0]), str(test_acc[0])]) mean_loss, mean_acc = tuple(sum(e)/len(e) for e in zip(*results)) tsv_writer.writerow(["-1", "-1", "-1", "-1", "-1", "-1", str(mean_loss), str(mean_acc)]) return (mean_loss, mean_acc) def train_model(A, model, dataset, sample_weights, batch_size, nepoch): logging.info("Training for {} epoch".format(nepoch)) # Log wall-clock time t0 = time() for epoch in range(1, nepoch+1): # Single training iteration model.fit(x=[dataset['train']['X']] + A, y=dataset['train']['Y'], batch_size=batch_size, epochs=1, shuffle=False, sample_weight=sample_weights, validation_data=([dataset['val']['X']] + A, dataset['val']['Y']), callbacks=[], verbose=0) # Predict on full dataset Y_hat = model.predict(x=[dataset['train']['X']] + A, batch_size=batch_size, verbose=0) # Train / validation scores train_val_loss, train_val_acc = evaluate_model(Y_hat, [dataset['train']['Y'], dataset['val']['Y']], [dataset['train']['X_idx'], dataset['val']['X_idx']]) logging.info("{:04d} ".format(epoch) \ + "| train loss {:.4f} / acc {:.4f} ".format(train_val_loss[0], train_val_acc[0]) + "| val loss {:.4f} / acc {:.4f}".format(train_val_loss[1], train_val_acc[1])) yield (epoch, train_val_loss[0], train_val_acc[0], train_val_loss[1], train_val_acc[1]) logging.info("training time: {:.2f}s".format(time()-t0)) def test_model(A, model, dataset, batch_size): # Predict on full dataset Y_hat = model.predict(x=[dataset['train']['X']] + A, batch_size=batch_size, verbose=0) test_loss, test_acc = evaluate_model(Y_hat, [dataset['test']['Y']], [dataset['test']['X_idx']]) logging.info("Performance on test set: loss {:.4f} / accuracy {:.4f}".format( test_loss[0], test_acc[0])) return (test_loss, test_acc) def run(args, tsv_writer, config): set_seed(config['task']['seed']) # prep data if args.input is None: logging.debug("No tarball supplied - building task prequisites") with KnowledgeGraph(path=config['graph']['file']) as kg: targets = strip_graph(kg, config) A = graph_structure.generate(kg, config) X, Y, X_node_map = build_dataset(kg, targets, config) else: assert is_readable(args.input) logging.debug("Importing prepared tarball") with Tarball(args.input, 'r') as tb: A = tb.get('A') X = tb.get('X') Y = tb.get('Y') X_node_map = tb.get('X_node_map') if config['task']['kfolds'] < 0: loss, accuracy = single_run(A, X, Y, X_node_map, tsv_writer, config) else: loss, accuracy = kfold_crossvalidation(A, X, Y, X_node_map, config['task']['kfolds'], tsv_writer, config) logging.info("Mean performance: loss {:.4f} / accuracy {:.4f}".format( loss, accuracy)) if args.verbose < 1: print("Mean performance: loss {:.4f} / accuracy {:.4f}".format( loss, accuracy)) def init_logger(filename, verbose=0): logging.basicConfig(filename=filename, format='[%(asctime)s] %(module)s/%(funcName)s | %(levelname)s: %(message)s', level=logging.DEBUG) if verbose > 0: stream_handler = logging.StreamHandler() level = logging.INFO if verbose >= 2: level = logging.DEBUG stream_handler.setLevel(level) logging.getLogger().addHandler(stream_handler) if __name__ == "__main__": timestamp = int(time()) parser = argparse.ArgumentParser() parser.add_argument("-c", "--config", help="Configuration file (toml)", required=True, default=None) parser.add_argument("-i", "--input", help="Optional prepared input file (tar)", default=None) parser.add_argument("-o", "--output", help="Output directory", default="/tmp/") parser.add_argument("-v", "--verbose", help="Increase output verbosity", action='count', default=0) args = parser.parse_args() # load configuration assert is_readable(args.config) config = toml.load(args.config) # set output base filename baseFilename = "{}{}{}".format(args.output, config['name'], timestamp) if args.output.endswith("/") \ else "{}/{}{}".format(args.output, config['name'], timestamp) assert is_writable(baseFilename) init_logger(baseFilename+'.log', args.verbose) logger = logging.getLogger(__name__) tsv_writer = TSV(baseFilename+'.tsv', 'w') # log parameters logger.debug("Arguments:\n{}".format( "\n".join(["\t{}: {}".format(arg, getattr(args, arg)) for arg in vars(args)]))) logger.debug("Configuration:\n{}".format( "\n".join(["\t{}: {}".format(k,v) for k,v in config.items()]))) # run training run(args, tsv_writer, config) logging.shutdown()
9,461
2,844
import argparse, time, os, cv2, shutil, datetime, math, subprocess, pickle, multiprocessing from actn import * ap = argparse.ArgumentParser() # for help -> python alpha.py --help ap.add_argument("-f", "--file", required=True, help="name of the file") ap.add_argument("-o", "--output", required=True, help="specifiy the folder path of output") ap.add_argument("-b", "--before", required=True, help="seconds to cut before", type=int) ap.add_argument("-a", "--after", required=True, help="seconds to cut after", type=int) args = vars(ap.parse_args()) class core_overwatch(): def __init__(self, file_name, output_folder, before, after): self.file_name = file_name self.output_folder = output_folder self.before = before self.after = after if not os.path.exists(str(self.output_folder)): print("The File Path Doesn't Exist!") print("[++++++]Creating the Folder in Path {0}".format(output_folder)) os.makedirs("{0}".format(self.output_folder)) print("[++++++]Finished Making The Folder in Path {0}".format(output_folder)) try: fh = open('{0}'.format(self.file_name), 'r') except FileNotFoundError: print("[+++++++]The Video File Not Found In Path.Please Try Again") cmd1 = "ffmpeg -i {0} 2>&1 | sed -n \"s/.*, \(.*\) fp.*/\\1/p\"".format(self.file_name) os.system(cmd1 + ' > tmp1') self.fps = int(open('tmp1', 'r').read()) os.system( """ ffprobe -v error -select_streams v:0 -show_entries stream=nb_frames -of default=nokey=1:noprint_wrappers=1 {0} > tmp2 """.format( self.file_name )) self.frame_count = int(open('tmp2', 'r').read()) print('[++++++]fps', self.fps) print('[++++++]frame count', self.frame_count) # get imp vid inf def build_folder(self): folder_names = ['./raw_calc', './raw_calc/frame_db_temp'] for directory in folder_names: if not os.path.exists(str(directory)): os.makedirs(str(directory)) # if exists then delete all the files in that dir tree def which_frame_formula(self): second_length = 1 chunk_size = round(self.fps * second_length) # fps*second_length assert type(chunk_size) is int, "Chunk Size must have to be Integer Type" # upto which frame the ops will execute(for loop to extract one frame from chunk size ) n = round(round(self.frame_count) / chunk_size) start_frame = round(self.fps / 2) common_diff = round(self.fps * second_length) # * second length,taking 1F/60 return start_frame, n, common_diff def select_frame(self, a, n, d): # arithmetic series y=a+(p-1)*d which_frame_list = [a + (p - 1) * d for p in range(1, n + 1)] return which_frame_list def read_save_frame(self): os.system("ffmpeg -hide_banner -loglevel panic -i {video_Fname} -vf fps=1 {f_name}/%d.png".format( f_name='./raw_calc/frame_db_temp', video_Fname=str(self.file_name) )) def get_action_process_multithreaded_cmd_run_commands(self): img_list = ['./raw_calc/frame_db_temp/{0}'.format(x) for x in os.listdir('./raw_calc/frame_db_temp')] img_list.sort(key=lambda fx: int(''.join(filter(str.isdigit, fx)))) az = return_text(img_list) return az # utils function start here -3 def _dbl(self, time): if time < 10: return '0' + str(time) else: return str(time) def time_cut(self, input_in_sec): times = [] hours = 0 minutes = 0 seconds = 0 hours = input_in_sec // 3600 minutes = (input_in_sec % 3600) // 60 seconds = (input_in_sec % 3600) % 60 return "{}:{}:{}".format(core_overwatch._dbl(self, hours), core_overwatch._dbl(self, minutes), core_overwatch._dbl(self, seconds)) def findIndices(self, sequence, _str, extra=0): # 0011 assert len(sequence) < len(_str), "Sequence is Greater Than the Main String" indices = [] for i in range(len(_str) - len(sequence) + 1): temp = _str[i:i + len(sequence)] if (sequence == temp): indices.append(i + 2 - extra) return indices # utils fx ends here def action_index_find(self, raw_list, which_frame): raw_str_hashed = '' for j in raw_list: raw_str_hashed += str(j) assert type(raw_str_hashed) is str, " The parameter to find Indices Type must have to be a String" result_list = core_overwatch.findIndices(self, '01', raw_str_hashed, extra=1) final_result = [] for yx in result_list: final_result.append(int(which_frame[yx])) return final_result def build_frame_range_to_cut(self, action_result): # print(action_result) # input will be taken ->cp from raw code frames = round(self.frame_count) fps = round(self.fps) bef = int(self.before) * fps # count frm aft = int(self.after) * fps # frame range (tuple ds) contained list frame_range = [] # build condition for after and before trimming for ucv in action_result: if int(ucv) < bef and aft < frames: frame_range.append((0, ucv + aft)) elif int(ucv) < bef and aft > frames: frame_range.append((0, frames)) elif int(ucv) > bef and aft < frames: frame_range.append((ucv - bef, ucv + aft)) elif int(ucv) > bef and aft < frames: frame_range.append((ucv - bef, frames)) # (temp) test return frame_range def build_output(self, start, end, video_name, file_name, end1): os.system( 'ffmpeg -hide_banner -loglevel panic -ss {st} -i {ivfname} -to {ed} -c copy {ovfname}'.format(st=start, ed=end1, ivfname=self.file_name, ovfname=video_name)) file_ = open('{}'.format(file_name), 'w') file_.write('Start at : {sec} \n End at : {sec1} '.format(sec=start, sec1=end)) file_.close() def send_frame_signal(self, frame_range): # frame range is like [(0,21),(4,198)] assert type(frame_range) is list, "Frame range must have to be a list" fps = round(self.fps) # build video file path name ax = str(datetime.datetime.now()) tm = ax[0:10] + '_' + ax[11:] file_n_ = str(self.output_folder + '/' + str(tm)) os.makedirs(file_n_) video_type = os.path.splitext(os.path.basename(str(self.file_name)))[1] # output e.g as .mp4 for ux in range(len(frame_range)): start = core_overwatch.time_cut(self, input_in_sec=math.ceil(frame_range[ux][0] / fps)) end = core_overwatch.time_cut(self, input_in_sec=math.ceil(frame_range[ux][1] / fps)) end1 = core_overwatch.time_cut(self, input_in_sec=math.ceil( (frame_range[ux][1] / fps) - (frame_range[ux][0] / fps))) print('[++++++]Start at {0} End at {1}'.format(start, end)) core_overwatch.build_output(self, start=str(start), end=str(end), video_name=file_n_ + '/output{vid_number}{type_v}'.format(vid_number=ux, type_v=video_type), file_name=file_n_ + '/output{0}.txt'.format(ux), end1=end1 ) print("Total {0} Videos have been cut from Main Video".format(len(os.listdir(file_n_))/2)) if __name__ == "__main__": a = core_overwatch(file_name=str(args['file']), output_folder=str(args['output']), before=int(args['before']), after=int(args['after'])) a.build_folder() start_frame, n, common_diff = a.which_frame_formula() # returns a,n,d c = a.select_frame(start_frame, n, common_diff) # returns which_frame_list st = time.time() print("[+++++]Reading Frames....") a.read_save_frame() print("[+++++++]Finished Reading Frames") print("[+++++++]Image Processing Rolling....") d = a.get_action_process_multithreaded_cmd_run_commands() print("[++++++++]Finished Processing Images") f = a.action_index_find(raw_list=d, which_frame=c) # return list to start aft and bef(action first observed) g = a.build_frame_range_to_cut(f) a.send_frame_signal(frame_range=g) print('[++++++]Time req to run The Engine is {0}m'.format((time.time() - st) / 60)) print('Deleting temp folders..') shutil.rmtree('./raw_calc/frame_db_temp') os.remove('./tmp1') os.remove('./tmp2')
9,343
3,068
from dcim.choices import DeviceStatusChoices from dcim.models import Device from extras.reports import Report class DeviceIPReport(Report): description = ( "Check that every device has either an IPv4 or IPv6 primary address assigned" ) def test_primary_ip4(self): for device in Device.objects.filter(status=DeviceStatusChoices.STATUS_ACTIVE): intcount = 0 for interface in device.interfaces.all(): if not interface.mgmt_only: intcount += 1 # There may be dumb devices with no interfaces so no IP addresses, that's OK if intcount == 0: if device.primary_ip4_id is not None: if device.primary_ip6_id is not None: self.log_failure( device, "Device has primary IPv4 and IPv6 address but no interfaces", ) else: self.log_warning( device, "Device has missing primary IPv4 addresses but no interfaces", ) else: self.log_success(device) elif device.primary_ip4_id is None: if device.device_type.is_child_device is True: self.log_success(device) else: if device.primary_ip6_id is None: self.log_failure( device, "Device is missing primary IPv4 and IPv6 address" ) else: self.log_warning( device, "Device is missing primary IPv4 addresses" ) else: if device.device_type.is_child_device is True: self.log_success(device) else: if device.primary_ip6_id is None: self.log_info(device, "Device is missing primary IPv6 address") else: self.log_success(device)
2,168
514
import socket import random import os import requests import re import github import minecraft import string import sys HOST = "xeroxirc.net" PORT = 6667 NICK = "ak_sus" #PASSWORD = os.getenv("PASSWORD") CHANNEL = "#BlockySurvival" SERVER = "" readbuffer = "" def send(message): s.send(message) print(message) s = socket.socket() s.connect((HOST, PORT)) send(bytes("NICK %s\r\n" % NICK, "UTF-8")) send(bytes("USER %s %s %s :%s\r\n" % (NICK, NICK, NICK, NICK), "UTF-8")) #s.send(bytes("PRIVMSG NickServ regain {} {}\r\n".format(NICK, PASSWORD), "UTF-8")) #s.send(bytes("PRIVMSG NickServ identify {} {}\r\n".format(NICK, PASSWORD), "UTF-8")) send(bytes("JOIN {}\r\n".format(CHANNEL), "UTF-8")) #s.send(bytes("PRIVMSG NickServ :identify {}\r\n".format(PASSWORD), "UTF-8")) readbuffer = readbuffer + s.recv(1024).decode("UTF-8") temp = str.split(readbuffer, "\n") readbuffer = temp.pop() for line in temp: SERVER = str.rstrip(line)[1:].split()[0] print(str.rstrip(line)) while 1: readbuffer = readbuffer + s.recv(1024).decode("UTF-8") temp = str.split(readbuffer, "\n") readbuffer = temp.pop() for line in temp: print(str.rstrip(line)) message = str.rstrip(line).split(" PRIVMSG {} :".format(CHANNEL)) if "PING" in line: send("PONG :{}\r\n".format(SERVER).encode("utf-8")) msg = message[-1] tokens = msg.split() if msg == "$hello": send("PRIVMSG {} :Hello!\r\n".format(CHANNEL).encode("utf-8")) if msg == "$ping": send("PRIVMSG {} :Pong!\r\n".format(CHANNEL).encode("utf-8")) if msg == "$random": send("PRIVMSG {} :{}\r\n".format(CHANNEL, random.randint(0, 100)).encode("utf-8")) if msg.startswith("$youtube "): html = requests.get("https://www.youtube.com/results?search_query=" + " ".join(msg.split()[1:])).content video_ids = re.findall(r"watch\?v=(\S{11})", html.decode()) send("PRIVMSG {} :https://www.youtube.com/watch?v={}\r\n".format(CHANNEL, video_ids[0]).encode("utf-8")) #if msg.startswith("$google "): send("PRIVMSG {} :{}\r\n".format(CHANNEL, googlesearch.search(" ".join(msg.split()[1:]))[0]).encode("utf-8")) #if msg.startswith("$wolfram "): send("PRIVMSG {} :{}\r\n".format(CHANNEL, wolfram.get(" ".join(msg.split()[1:]))).encode("utf-8")) if msg.startswith("$github "): if tokens[1] == "url": send("PRIVMSG {} :https://github.com/{}/{}\r\n".format(CHANNEL, tokens[2], tokens[3]).encode("utf-8")) if tokens[1] == "issues": send("PRIVMSG {} :#{}: {}\r\n".format(CHANNEL, tokens[4], github.get_issue_title(tokens[2], tokens[3], tokens[4])).encode("utf-8")) if msg == "$server": send("PRIVMSG {} :{}\r\n".format(CHANNEL, minecraft.get()).encode("utf-8")) if msg == "$help": send("PRIVMSG {} :Avalible commands: $hello, $ping, $youtube, $google, $github, $wolfram.\r\n".format(CHANNEL).encode("utf-8")) if msg.startswith("$help "): if tokens[1] == "hello": send("PRIVMSG {} :Syntax: $hello Action: Says \"Hello!\".\r\n".format(CHANNEL).encode("utf-8")) if tokens[1] == "ping":send("PRIVMSG {} :Syntax: $ping Action: Says \"Ping!\".\r\n".format(CHANNEL).encode("utf-8")) if tokens[1] == "youtube": send("PRIVMSG {} :Syntax: $youtube <keyword> Action: Sends the URL of a YouTube video matching the keyword given.\r\n".format(CHANNEL).encode("utf-8")) #if tokens[1] == "google": send("PRIVMSG {} :Syntax: $google <keyword> Action: Sends the URL of a google search with the keyword given\r\n".format(CHANNEL).encode("utf-8")) if tokens[1] == "github": send("PRIVMSG {} :Syntax: $github <topic> <user> <repo> <number> Action: Returns data about a github repo.\r\n".format(CHANNEL).encode("utf-8")) #if tokens[1] == "wolfram": send("PRIVMSG {} :Syntax: $wolfram <query> Action: Asks Wolfram|Alpha the query given.\r\n".format(CHANNEL).encode("utf-8"))
3,940
1,495
""" The file defines the evaluate process on target dataset. @Author: Yang Lu @Github: https://github.com/luyanger1799 @Project: https://github.com/luyanger1799/amazing-semantic-segmentation """ from sklearn.metrics import multilabel_confusion_matrix from amazingutils.helpers import * from amazingutils.utils import load_image import numpy as np import argparse import sys import cv2 import os def str2bool(v): if v.lower() in ('yes', 'true', 't', 'y', '1'): return True elif v.lower() in ('no', 'false', 'f', 'n', '0'): return False else: raise argparse.ArgumentTypeError('Boolean value expected.') parser = argparse.ArgumentParser() parser.add_argument('--dataset', help='The path of the dataset.', type=str, default='CamVid') parser.add_argument('--crop_height', help='The height to crop the image.', type=int, default=256) parser.add_argument('--crop_width', help='The width to crop the image.', type=int, default=256) parser.add_argument('--predictions', help='The path of predicted image.', type=str, required=True) args = parser.parse_args() # check related paths paths = check_related_path(os.getcwd()) # get image and label file names for training and validation _, _, _, _, _, test_label_names = get_dataset_info(args.dataset) # get color info csv_file = os.path.join(args.dataset, 'class_dict.csv') class_names, _ = get_colored_info(csv_file) # get the prediction file name list if not os.path.exists(args.predictions): raise ValueError('the path of predictions does not exit.') prediction_names = [] for file in sorted(os.listdir(args.predictions)): prediction_names.append(os.path.join(args.predictions, file)) # evaluated classes evaluated_classes = get_evaluated_classes(os.path.join(args.dataset, 'evaluated_classes.txt')) num_classes = len(class_names) class_iou = dict() for name in evaluated_classes: class_iou[name] = list() class_idx = dict(zip(class_names, range(num_classes))) # begin evaluate assert len(test_label_names) == len(prediction_names) for i, (name1, name2) in enumerate(zip(test_label_names, prediction_names)): sys.stdout.write('\rRunning test image %d / %d' % (i + 1, len(test_label_names))) sys.stdout.flush() label = np.array(cv2.resize(load_image(name1), dsize=(args.crop_width, args.crop_height), interpolation=cv2.INTER_NEAREST)) pred = np.array(cv2.resize(load_image(name2), dsize=(args.crop_width, args.crop_height), interpolation=cv2.INTER_NEAREST)) confusion_matrix = multilabel_confusion_matrix(label.flatten(), pred.flatten(), labels=list(class_idx.values())) for eval_cls in evaluated_classes: eval_idx = class_idx[eval_cls] (tn, fp), (fn, tp) = confusion_matrix[eval_idx] if tp + fn > 0: class_iou[eval_cls].append(tp / (tp + fp + fn)) print('\n****************************************') print('* The IoU of each class is as follows: *') print('****************************************') for eval_cls in evaluated_classes: class_iou[eval_cls] = np.mean(class_iou[eval_cls]) print('{cls:}: {iou:.4f}'.format(cls=eval_cls, iou=class_iou[eval_cls])) print('\n**********************************************') print('* The Mean IoU of all classes is as follows: *') print('**********************************************') print('Mean IoU: {mean_iou:.4f}'.format(mean_iou=np.mean(list(class_iou.values()))))
3,456
1,133
from pso.GPSO import GPSO import numpy as np import time import pandas as pd np.random.seed(42) # f1 完成 def Sphere(p): # Sphere函数 out_put = 0 for i in p: out_put += i ** 2 return out_put # f2 完成 def Sch222(x): out_put = 0 out_put01 = 1 for i in x: out_put += abs(i) out_put01 = abs(i)*out_put01 out_put = out_put01+out_put return out_put # f3 完成 def Quadric(x): output = 0 # print(x.shape[0]) for i in range(x.shape[0]): output += (np.sum(x[0:i+1])) ** 2 # print(np.square(np.sum(x[0:i+1]))) return output # f4 完成 def Schl(x): # print(np.max(np.abs(x))) return np.max(np.abs(x)) # f5 完成 def Step(x): output = 0 for i in x: output += (np.floor(i+0.5))**2 return output # f6 完成 def Noise(x): output = 0 cnt = 1 for i in x: output = cnt * (i**4) + output cnt += 1 output += np.random.rand() return output # f7 完成 def Rosenbrock(p): ''' -2.048<=xi<=2.048 函数全局最优点在一个平滑、狭长的抛物线山谷内,使算法很难辨别搜索方向,查找最优也变得十分困难 在(1,...,1)处可以找到极小值0 :param p: :return: ''' n_dim = len(p) res = 0 for i in range(n_dim - 1): res += 100 * np.square(np.square(p[i]) - p[i + 1]) + np.square(p[i] - 1) return res # f8 有问题,忽略,这个是APSO的f8 def Schewel(x): out_put = 0 for i in x: out_put += -i*np.sin(np.sqrt(abs(i))) return out_put # f9 完成 def Rastrigin(p): ''' 多峰值函数,也是典型的非线性多模态函数 -5.12<=xi<=5.12 在范围内有10n个局部最小值,峰形高低起伏不定跳跃。很难找到全局最优 has a global minimum at x = 0 where f(x) = 0 ''' return np.sum([np.square(x) - 10 * np.cos(2 * np.pi * x) + 10 for x in p]) # f10 def Ackley(x): part1 = 0 part2 = 0 for i in x: part1 += (i**2) part2 += np.cos(2 * np.pi * i) left = 20 * np.exp(-0.2 * ((part1 / x.shape[0]) ** .5)) right = np.exp(part2 / x.shape[0]) return -left - right + 20 + np.e # f11 ok def Griewank(p): ''' 存在多个局部最小值点,数目与问题的维度有关。 此函数是典型的非线性多模态函数,具有广泛的搜索空间,是优化算法很难处理的复杂多模态问题。 在(0,...,0)处取的全局最小值0 -600<=xi<=600 ''' part1 = [np.square(x) / 4000 for x in p] part2 = [np.cos(x / np.sqrt(i + 1)) for i, x in enumerate(p)] return np.sum(part1) - np.prod(part2) + 1 g = 10000 times = 30 table = np.zeros((2, 10)) gBest = np.zeros((10, 30)) # 1010个函数的30次的最优值 for i in range(times): optimizer = GPSO(func=Sphere, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-100), ub=np.ones(30) * 100, w=0.9, c1=2, c2=2, acceptance=0.01) start = time.time() optimizer.run() end = time.time() print('Sphere:', optimizer.gbest_y) table[0, 0] += optimizer.gbest_y table[1, 0] += end - start gBest[0, i] = optimizer.gbest_y optimizer = GPSO(func=Sch222, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-10), ub=np.ones(30) * 10, w=0.9, c1=2, c2=2, acceptance=0.01) start = time.time() optimizer.run() end = time.time() print('Sch222:', optimizer.gbest_y) table[0, 1] += optimizer.gbest_y table[1, 1] += end - start gBest[1, i] = optimizer.gbest_y optimizer = GPSO(func=Quadric, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-100), ub=np.ones(30) * 100, w=0.9, c1=2, c2=2, acceptance=100) start = time.time() optimizer.run() end = time.time() print('Quadric:', optimizer.gbest_y) table[0, 2] += optimizer.gbest_y table[1, 2] += end - start gBest[2, i] = optimizer.gbest_y optimizer = GPSO(func=Rosenbrock, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-10), ub=np.ones(30) * 10, w=0.9, c1=2, c2=2, acceptance=100) start = time.time() optimizer.run() end = time.time() print('Rosenbrock:', optimizer.gbest_y) table[0, 3] += optimizer.gbest_y table[1, 3] += end - start gBest[3, i] = optimizer.gbest_y optimizer = GPSO(func=Step, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-100), ub=np.ones(30) * 100, w=0.9, c1=2, c2=2, acceptance=0) start = time.time() optimizer.run() end = time.time() print('Step:', optimizer.gbest_y) table[0, 4] += optimizer.gbest_y table[1, 4] += end - start gBest[4, i] = optimizer.gbest_y optimizer = GPSO(func=Noise, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-1.28), ub=np.ones(30) * 1.28, w=0.9, c1=2, c2=2, acceptance=0.01) start = time.time() optimizer.run() end = time.time() print('Noise:', optimizer.gbest_y) table[0, 5] += optimizer.gbest_y table[1, 5] += end - start gBest[5, i] = optimizer.gbest_y optimizer = GPSO(func=Schewel, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-500), ub=np.ones(30) * 500, w=0.9, c1=2, c2=2, acceptance=-10000) start = time.time() optimizer.run() end = time.time() print('Schewel:', optimizer.gbest_y) table[0, 6] += optimizer.gbest_y table[1, 6] += end - start gBest[6, i] = optimizer.gbest_y optimizer = GPSO(func=Rastrigin, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-5.12), ub=np.ones(30) * 5.12, w=0.9, c1=2, c2=2, acceptance=50) start = time.time() optimizer.run() end = time.time() print('Rastrigin:', optimizer.gbest_y) table[0, 7] += optimizer.gbest_y table[1, 7] += end - start gBest[7, i] = optimizer.gbest_y optimizer = GPSO(func=Ackley, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-32), ub=np.ones(30) * 32, w=0.9, c1=2, c2=2, acceptance=0.01) start = time.time() optimizer.run() end = time.time() print('Ackley:', optimizer.gbest_y) table[0, 8] += optimizer.gbest_y table[1, 8] += end - start gBest[8, i] = optimizer.gbest_y optimizer = GPSO(func=Griewank, dim=30, pop=20, max_iter=g, lb=np.ones(30) * (-600), ub=np.ones(30) * 600, w=0.9, c1=2, c2=2, acceptance=0.01) start = time.time() optimizer.run() end = time.time() print('Griewank:', optimizer.gbest_y) table[0, 9] += optimizer.gbest_y table[1, 9] += end - start gBest[9, i] = optimizer.gbest_y table = table / times table = pd.DataFrame(table) table.columns = ['Sphere', 'Schwefel_P222', 'Quadric', 'Rosenbrock', 'Step', 'Quadric_Noise', 'Schwefel', 'Rastrigin', 'Ackley', 'Griewank'] table.index = ['mean score', 'mean time'] print(table) print('10个测试函数的30次std:', np.std(gBest, axis=1)) print('10个测试函数的30次best:', np.min(gBest, axis=1))
6,524
3,202
# # Copyright (c) 2013 Juniper Networks, Inc. All rights reserved. # from setuptools import setup import setuptools setup( name='contrail-server-manager', version='0.1dev', packages=setuptools.find_packages(exclude=["*.pyc"]), zip_safe=False, long_description="Server Manager package", )
310
106
# Test restoring a scene with multiple lists with different number # of fiducials # first fiducial list displayNode1 = slicer.vtkMRMLMarkupsDisplayNode() slicer.mrmlScene.AddNode(displayNode1) fidNode1 = slicer.vtkMRMLMarkupsFiducialNode() fidNode1.SetName("FidNode1") slicer.mrmlScene.AddNode(fidNode1) fidNode1.SetAndObserveDisplayNodeID(displayNode1.GetID()) coords = [0.0, 0.0, 0.0] numFidsInList1 = 5 for i in range(numFidsInList1): fidNode1.AddFiducialFromArray(coords) coords[0] += 1.0 coords[1] += 2.0 coords[2] += 1.0 # second fiducial list displayNode2 = slicer.vtkMRMLMarkupsDisplayNode() slicer.mrmlScene.AddNode(displayNode2) fidNode2 = slicer.vtkMRMLMarkupsFiducialNode() fidNode2.SetName("FidNode2") slicer.mrmlScene.AddNode(fidNode2) fidNode2.SetAndObserveDisplayNodeID(displayNode2.GetID()) numFidsInList2 = 10 for i in range(numFidsInList2): fidNode2.AddFiducialFromArray(coords) coords[0] += 1.0 coords[1] += 1.0 coords[2] += 3.0 sv = slicer.mrmlScene.AddNode(slicer.vtkMRMLSceneViewNode()) numFidNodesBeforeStore = slicer.mrmlScene.GetNumberOfNodesByClass('vtkMRMLMarkupsFiducialNode') sv.StoreScene() # add a third list that will get removed on restore # second fiducial list displayNode3 = slicer.vtkMRMLMarkupsDisplayNode() slicer.mrmlScene.AddNode(displayNode3) fidNode3 = slicer.vtkMRMLMarkupsFiducialNode() fidNode3.SetName("FidNode3") slicer.mrmlScene.AddNode(fidNode3) fidNode3.SetAndObserveDisplayNodeID(displayNode3.GetID()) numFidsInList3 = 2 for i in range(numFidsInList3): fidNode3.AddFiducialFromArray(coords) coords[0] += 1.0 coords[1] += 2.0 coords[2] += 3.0 sv.RestoreScene() numFidNodesAfterRestore = slicer.mrmlScene.GetNumberOfNodesByClass('vtkMRMLMarkupsFiducialNode') if numFidNodesAfterRestore != numFidNodesBeforeStore: print "After restoring the scene, expected ", numFidNodesBeforeStore, " fiducial nodes, but have ", numFidNodesAfterRestore exceptionMessage = "After restoring the scene, expected " + str(numFidNodesBeforeStore) + " fiducial nodes, but have " + str(numFidNodesAfterRestore) raise Exception(exceptionMessage) #fid1AfterRestore = slicer.mrmlScene.GetNodeByID("vtkMRMLMarkupsFiducialNode1") fid1AfterRestore = slicer.mrmlScene.GetFirstNodeByName("FidNode1") numFidsInList1AfterRestore = fid1AfterRestore.GetNumberOfMarkups() print "After restore, list with name FidNode1 has id ", fid1AfterRestore.GetID(), " and num fids = ", numFidsInList1AfterRestore if numFidsInList1AfterRestore != numFidsInList1: exceptionMessage = "After restoring list 1, id = " + fid1AfterRestore.GetID() exceptionMessage += ", expected " + str(numFidsInList1) + " but got " exceptionMessage += str(numFidsInList1AfterRestore) raise Exception(exceptionMessage) # fid2AfterRestore = slicer.mrmlScene.GetNodeByID("vtkMRMLMarkupsFiducialNode2") fid2AfterRestore = slicer.mrmlScene.GetFirstNodeByName("FidNode2") numFidsInList2AfterRestore = fid2AfterRestore.GetNumberOfMarkups() print "After restore, list with name FidNode2 has id ", fid2AfterRestore.GetID(), " and num fids = ", numFidsInList2AfterRestore if numFidsInList2AfterRestore != numFidsInList2: exceptionMessage = "After restoring list 2, id = " + fid2AfterRestore.GetID() exceptionMessage += ", expected " + str(numFidsInList2) + " but got " exceptionMessage += str(numFidsInList2AfterRestore) raise Exception(exceptionMessage) # check the displayable manager for the right number of widgets/seeds lm = slicer.app.layoutManager() td = lm.threeDWidget(0) ms = vtk.vtkCollection() td.getDisplayableManagers(ms) fidManagerIndex = -1 for i in range(ms.GetNumberOfItems()): m = ms.GetItemAsObject(i) if m.GetClassName() == "vtkMRMLMarkupsFiducialDisplayableManager3D": fidManagerIndex = i print m.GetClassName(), fidManagerIndex if fidManagerIndex == -1: exceptionMessage = "Failed to find markups fiducial displayable manager 3d!" raise Exception(exceptionMessage) mfm = ms.GetItemAsObject(fidManagerIndex) h = mfm.GetHelper() print 'Helper = ',h seedWidget1 = h.GetWidget(fid1AfterRestore) rep1 = seedWidget1.GetRepresentation() print "Seed widget 1 has number of seeds = ",rep1.GetNumberOfSeeds() if rep1.GetNumberOfSeeds() != numFidsInList1AfterRestore: exceptionMessage = "After restoring list 1, expected seed widget to have " exceptionMessage += str(numFidsInList1AfterRestore) + " seeds, but it has " exceptionMessage += str(rep1.GetNumberOfSeeds()) raise Exception(exceptionMessage) # check positions for s in range(numFidsInList1AfterRestore): seed = seedWidget1.GetSeed(s) handleRep = seed.GetHandleRepresentation() worldPos = handleRep.GetWorldPosition() print "seed ",s," world position = ",worldPos fidPos = [0.0,0.0,0.0] fid1AfterRestore.GetNthFiducialPosition(s,fidPos) xdiff = fidPos[0] - worldPos[0] ydiff = fidPos[1] - worldPos[1] zdiff = fidPos[2] - worldPos[2] diffTotal = xdiff + ydiff + zdiff if diffTotal > 0.1: exceptionMessage = "List1: Difference between seed position " + str(s) exceptionMessage += " and fiducial position totals = " + str(diffTotal) raise Exception(exceptionMessage) seedWidget2 = h.GetWidget(fid2AfterRestore) rep2 = seedWidget2.GetRepresentation() print "Seed widget 2 has number of seeds = ",rep2.GetNumberOfSeeds() if rep2.GetNumberOfSeeds() != numFidsInList2AfterRestore: exceptionMessage = "After restoring fid list 2, expected seed widget to have " exceptionMessage += str(numFidsInList2AfterRestore) + " seeds, but it has " exceptionMessage += str(rep2.GetNumberOfSeeds()) raise Exception(exceptionMessage) # check positions for s in range(numFidsInList2AfterRestore): seed = seedWidget2.GetSeed(s) handleRep = seed.GetHandleRepresentation() worldPos = handleRep.GetWorldPosition() print "seed ",s," world position = ",worldPos fidPos = [0.0,0.0,0.0] fid2AfterRestore.GetNthFiducialPosition(s,fidPos) xdiff = fidPos[0] - worldPos[0] ydiff = fidPos[1] - worldPos[1] zdiff = fidPos[2] - worldPos[2] diffTotal = xdiff + ydiff + zdiff if diffTotal > 0.1: exceptionMessage = "List2: Difference between seed position " + str(s) exceptionMessage += " and fiducial position totals = " + str(diffTotal) raise Exception(exceptionMessage) ms.RemoveAllItems()
6,272
2,299
# -*- coding:utf-8 -*- # @Time : 2019/7/21 12:35 PM # @Author : __wutonghe__ # docs https://channels.readthedocs.io/en/latest/tutorial/part_3.html#rewrite-the-consumer-to-be-asynchronous from channels.generic.websocket import AsyncWebsocketConsumer import json class MessageConsumer(AsyncWebsocketConsumer): """ 私信websocket,采用异步通信来增加并发 """ async def connect(self): """当 websocket 一链接上以后触发该函数""" if self.scope['user'].is_anonymous: await self.close() else: await self.channel_layer.group_add(self.scope['user'].username + '-message',self.channel_name) # 创建聊天室 await self.accept() async def receive(self, text_data=None, bytes_data=None): """将答复交回给websocket""" await self.send(text_data=json.dumps(text_data)) # 将消息发送给前端 async def disconnect(self, code): """断开链接时触发该函数""" await self.channel_layer.group_discard(self.scope['user'].username + '-message',self.channel_name) # 将该链接移出聊天室
1,013
413
import eagerpy as ep import numpy as np from scipy.interpolate import RegularGridInterpolator import hypney tl = ep.numpy def test_regular_grid_interpolator(): """Adapted from https://github.com/sbarratt/torch_interpolations/blob/master/tests/test_grid_interpolator.py """ points = [tl.arange(-0.5, 2.5, 0.1) * 1.0, tl.arange(-0.5, 2.5, 0.2) * 1.0] values = ( hypney.utils.eagerpy.sin(points[0])[:, None] + 2 * hypney.utils.eagerpy.cos(points[1])[None, :] + hypney.utils.eagerpy.sin(5 * points[0][:, None] @ points[1][None, :]) ) X, Y = ep.meshgrid(tl.arange(-0.5, 2, 0.1), tl.arange(-0.5, 2, 0.1)) points_to_interp = ep.stack([X.flatten(), Y.flatten()]).T gi = hypney.utils.interpolation.RegularGridInterpolator(points, values) fx = gi(points_to_interp) rgi = RegularGridInterpolator( [p.numpy() for p in points], [x.numpy() for x in values], bounds_error=False ) rfx = rgi(points_to_interp.numpy()) np.testing.assert_allclose(rfx, fx.numpy(), atol=1e-6) # TODO: port derivative test to eagerpy # note that points_to_interp has to be transposed # # def test_regular_grid_interpolator_derivative(): # points = [torch.arange(-.5, 2.5, .5) * 1., torch.arange(-.5, 2.5, .5) * 1.] # values = torch.sin(points[0])[:, None] + 2 * torch.cos(points[1])[None, :] + torch.sin(5 * points[0][:, None] @ points[1][None, :]) # values.requires_grad_(True) # # X, Y = np.meshgrid(np.arange(-.5, 2, .19), np.arange(-.5, 2, .19)) # points_to_interp = [torch.from_numpy( # X.flatten()).float(), torch.from_numpy(Y.flatten()).float()] # # def f(values): # return torch_interpolations.RegularGridInterpolator( # points, values)(points_to_interp) # # torch.autograd.gradcheck(f, (values,), eps=1e-5, atol=1e-1, rtol=1e-1) def test_interpolator_builder(): itp = hypney.utils.interpolation.InterpolatorBuilder([(-1, 0, 1)]) def scalar_f(z): return z[0] z = ep.astensor(np.array([1, 0, -1, 0, 1, 1, -1])) scalar_itp = itp.make_interpolator(scalar_f) np.testing.assert_array_equal(scalar_itp(z).numpy(), z.numpy()) def matrix_f(z): return ep.astensor(np.ones((2, 2)) * z[0]) matrix_itp = itp.make_interpolator(matrix_f) np.testing.assert_array_equal( matrix_itp(z).numpy(), z[:, None, None].numpy() * np.ones((1, 2, 2)) ) # What happened here? Does the test not make sense or did the API change? # np.testing.assert_array_equal( # matrix_itp(ep.numpy.array([0, 0, 0])).numpy(), # np.ones((2, 2)))
2,618
1,100
#!/usr/bin/python3 ########################################## # Routino database access from Python. # # Part of the Routino routing software. ########################################## # This file Copyright 2018 Andrew M. Bishop # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. ########################################## import routino.database # Database, access all attributes database = routino.database.LoadDatabase("../../src/test/fat", "turns") if database is None: database = routino.database.LoadDatabase("../src/test/fat", "turns") if database is None: print("Failed to load database") exit(1) print(database) database_attrs = ['nnodes', 'nsegments', 'nways', 'nrelations'] for attr in database_attrs: print(" Attribute: " + attr + " =", getattr(database, attr)) print("") # A single node, access all attributes and all functions node=database.GetNode(0) print("1st node =", node) node_attrs = ['id', 'firstsegment', 'latitude', 'longitude', 'allow', 'flags'] node_infos = ['', '', 'degrees', 'degrees', '[note 1]', '[note 2]'] for attr,info in zip(node_attrs,node_infos): print(" Attribute: " + attr + " =", getattr(node, attr), info) segments = node.Segments() print(" Function: " + "Segments()" + " = [" + ", ".join([str(segments[x]) for x in range(len(segments))]) + "]") print("") # A single segment, access all attributes and all functions segment=database.GetSegment(0) print("1st segment =", segment) segment_attrs = ['id', 'node1', 'node2', 'next2', 'way', 'distance', 'flags'] segment_infos = ['', '', '', '', '', 'km', '[note 3]'] for attr,info in zip(segment_attrs,segment_infos): print(" Attribute: " + attr + " =", getattr(segment, attr), info) print(" Function: " + "Node1()" + " = " + str(segment.Node1())) print(" Function: " + "Node2()" + " = " + str(segment.Node2())) print(" Function: " + "Way()" + " = " + str(segment.Way())) print("") # A single way, access all attributes and all functions way=database.GetWay(0) print("1st way =", way) way_attrs = ['id', 'name', 'allow', 'type', 'props', 'speed', 'weight', 'height', 'width', 'length'] way_infos = ['', '', '[note 1]', '[note 4]', '[note 5]', 'km/hr [note 6]', 'tonnes [note 6]', 'metres [note 6]', 'metres [note 6]', 'metres [note 6]'] for attr,info in zip(way_attrs,way_infos): print(" Attribute: " + attr + " =", getattr(way, attr), info) print("") # A single relation, access all attributes and all functions relation=database.GetRelation(0) print("1st relation =", relation) relation_attrs = ['id', 'from_seg', 'via_node', 'to_seg', 'from_way', 'to_way', 'from_node', 'to_node', 'except_transport'] relation_infos = ['', '', '', '', '', '', '', '', '[note 7]'] for attr,info in zip(relation_attrs,relation_infos): print(" Attribute: " + attr + " =", getattr(relation, attr), info) print(" Function: " + "FromSegment()" + " = " + str(relation.FromSegment())) print(" Function: " + "ViaNode()" + " = " + str(relation.ViaNode())) print(" Function: " + "ToSegment()" + " = " + str(relation.ToSegment())) print(" Function: " + "FromWay()" + " = " + str(relation.FromWay())) print(" Function: " + "ToWay()" + " = " + str(relation.ToWay())) print(" Function: " + "FromNode()" + " = " + str(relation.FromNode())) print(" Function: " + "ToNode()" + " = " + str(relation.ToNode())) print("") # The list of nodes as a list and an iterable (just the first 4) nodes=database.Nodes() print("len(database.Nodes()) = " + str(len(nodes))) print("database.Nodes() = [" + ", ".join([str(nodes[x]) for x in range(4)]) + ", ...]") for node in nodes: if node.id == 4: break print(node) print("") # The list of segments as a list and an iterable (just the first 4) segments=database.Segments() print("len(database.Segments()) = " + str(len(segments))) print("database.Segments() = [" + ", ".join([str(segments[x]) for x in range(4)]) + ", ...]") for segment in segments: if segment.id == 4: break print(segment) print("") # The list of ways as a list and an iterable (just the first 4) ways=database.Ways() print("len(database.Ways()) = " + str(len(ways))) print("database.Ways() = [" + ", ".join([str(ways[x]) for x in range(4)]) + ", ...]") for way in ways: if way.id == 4: break print(way) print("") # The list of relations as a list and an iterable (just the first 4) relations=database.Relations() print("len(database.Relations()) = " + str(len(relations))) print("database.Relations() = [" + ", ".join([str(relations[x]) for x in range(4)]) + ", ...]") for relation in relations: if relation.id == 4: break print(relation) print("") # Enumerated lists transports_enum = ["Transports_None", "Transports_Foot", "Transports_Horse", "Transports_Wheelchair", "Transports_Bicycle", "Transports_Moped", "Transports_Motorcycle", "Transports_Motorcar", "Transports_Goods", "Transports_HGV", "Transports_PSV", "Transports_ALL"] nodeflags_enum = ["Nodeflag_Super", "Nodeflag_U_Turn", "Nodeflag_Mini_Roundabout", "Nodeflag_Turn_Restrict", "Nodeflag_Turn_Restrict2"] segmentflags_enum = ["Segmentflag_Area", "Segmentflag_Oneway_1to2", "Segmentflag_Oneway_2to1", "Segmentflag_Super", "Segmentflag_Normal"] properties_enum = ["Properties_None", "Properties_Paved", "Properties_Multilane", "Properties_Bridge", "Properties_Tunnel", "Properties_FootRoute", "Properties_BicycleRoute", "Properties_ALL"] highway_enum = ["Highway_Motorway", "Highway_Trunk", "Highway_Primary", "Highway_Secondary", "Highway_Tertiary", "Highway_Unclassified", "Highway_Residential", "Highway_Service", "Highway_Track", "Highway_Cycleway", "Highway_Path", "Highway_Steps", "Highway_Ferry", "Highway_Count", "Highway_CycleBothWays", "Highway_OneWay", "Highway_Roundabout", "Highway_Area"] def print_enum(list): for item in list: print(" routino.database."+item) print("Note 1: The Node's and Way's 'allow' parameter can be the combination of these enumerated values:") print_enum(transports_enum) print("") print("Note 2: The Node's 'flags' parameter can be the combination of these enumerated values:") print_enum(nodeflags_enum) print("") print("Note 3: The Segment's 'flags' parameter can be the combination of these enumerated values:") print_enum(segmentflags_enum) print("") print("Note 4: The Way's 'type' parameter can be one the combination of these enumerated values:") print_enum(highway_enum) print("") print("Note 5: The Way's 'props' parameter can be the combination of these enumerated values:") print_enum(properties_enum) print("") print("Note 6: A value of zero for a Way's speed, weight, height, width or length means that there is no limit.") print("") print("Note 7: The Relation's 'except_transport' parameter can be the combination of these enumerated values:") print_enum(transports_enum) print("") import gc gc.collect()
8,451
2,696
import io from .. import util def test_parsing_pkg_info_file(mocker): open_mock = mocker.patch('vcsver.util.open') open_mock.return_value = io.StringIO( 'Name: name\n' 'Version: 1.0\n' ) pkg_info_data = util.parse_pkg_info_file(mocker.sentinel.path) open_mock = open_mock.assert_called_once_with( mocker.sentinel.path, 'rt', ) assert { 'Name': 'name', 'Version': '1.0', } == pkg_info_data
475
186
def determinant(matA): dimA = [] # find dimensions of arrA a = matA while type(a) == list: dimA.append(len(a)) a = a[0] #is it square if dimA[0] != dimA[1]: raise Exception("Matrix is not square") #find determinant total = 0 if dimA[0] == 2: total = matA[0][0] * matA[1][1] - matA[1][0] * matA[0][1] return total else: sign = 1 for i in range(dimA[0]): temp = matA[1:] #remove the current column from the temp stuff for j in range(dimA[0]-1): temp[j] = temp[j][0:i] + temp[j][i+1:] sub = determinant(temp) total = total + sign * matA[0][i] * sub sign *= -1 return total matA = [[1,2,3],[4,5,6],[7,8,15]] print(determinant(matA))
820
306
# # PySNMP MIB module BAY-STACK-MIB (http://snmplabs.com/pysmi) # ASN.1 source file:///Users/davwang4/Dev/mibs.snmplabs.com/asn1/BAY-STACK-MIB # Produced by pysmi-0.3.4 at Mon Apr 29 17:19:06 2019 # On host DAVWANG4-M-1475 platform Darwin version 18.5.0 by user davwang4 # Using Python version 3.7.3 (default, Mar 27 2019, 09:23:15) # OctetString, ObjectIdentifier, Integer = mibBuilder.importSymbols("ASN1", "OctetString", "ObjectIdentifier", "Integer") NamedValues, = mibBuilder.importSymbols("ASN1-ENUMERATION", "NamedValues") SingleValueConstraint, ConstraintsUnion, ConstraintsIntersection, ValueSizeConstraint, ValueRangeConstraint = mibBuilder.importSymbols("ASN1-REFINEMENT", "SingleValueConstraint", "ConstraintsUnion", "ConstraintsIntersection", "ValueSizeConstraint", "ValueRangeConstraint") NotificationGroup, ModuleCompliance = mibBuilder.importSymbols("SNMPv2-CONF", "NotificationGroup", "ModuleCompliance") TimeTicks, MibScalar, MibTable, MibTableRow, MibTableColumn, Gauge32, Counter64, Bits, Counter32, ModuleIdentity, ObjectIdentity, IpAddress, iso, Integer32, NotificationType, MibIdentifier, Unsigned32 = mibBuilder.importSymbols("SNMPv2-SMI", "TimeTicks", "MibScalar", "MibTable", "MibTableRow", "MibTableColumn", "Gauge32", "Counter64", "Bits", "Counter32", "ModuleIdentity", "ObjectIdentity", "IpAddress", "iso", "Integer32", "NotificationType", "MibIdentifier", "Unsigned32") TruthValue, TextualConvention, DisplayString = mibBuilder.importSymbols("SNMPv2-TC", "TruthValue", "TextualConvention", "DisplayString") bayStackMibs, = mibBuilder.importSymbols("SYNOPTICS-ROOT-MIB", "bayStackMibs") bayStackMib = ModuleIdentity((1, 3, 6, 1, 4, 1, 45, 5, 13)) bayStackMib.setRevisions(('2013-10-11 00:00', '2012-10-02 00:00', '2009-09-28 00:00', '2007-09-04 00:00', '2005-08-22 00:00',)) if mibBuilder.loadTexts: bayStackMib.setLastUpdated('201310110000Z') if mibBuilder.loadTexts: bayStackMib.setOrganization('Nortel Networks') bayStackObjects = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 5, 13, 1)) bayStackConfig = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 1)) bayStackConfigExpectedStackSize = MibScalar((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 8))).setMaxAccess("readwrite") if mibBuilder.loadTexts: bayStackConfigExpectedStackSize.setStatus('current') bayStackConfigStackErrorNotificationInterval = MibScalar((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 1, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 65535)).clone(60)).setUnits('Seconds').setMaxAccess("readwrite") if mibBuilder.loadTexts: bayStackConfigStackErrorNotificationInterval.setStatus('current') bayStackConfigStackErrorNotificationEnabled = MibScalar((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 1, 3), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: bayStackConfigStackErrorNotificationEnabled.setStatus('current') bayStackConfigStackRebootUnitOnFailure = MibScalar((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 1, 4), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: bayStackConfigStackRebootUnitOnFailure.setStatus('current') bayStackConfigStackRetryCount = MibScalar((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 1, 5), Unsigned32().subtype(subtypeSpec=ValueRangeConstraint(0, 4294967295))).setMaxAccess("readwrite") if mibBuilder.loadTexts: bayStackConfigStackRetryCount.setStatus('current') bayStackUnitConfigTable = MibTable((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 2), ) if mibBuilder.loadTexts: bayStackUnitConfigTable.setStatus('current') bayStackUnitConfigEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 2, 1), ).setIndexNames((0, "BAY-STACK-MIB", "bayStackUnitConfigIndex")) if mibBuilder.loadTexts: bayStackUnitConfigEntry.setStatus('current') bayStackUnitConfigIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 2, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 8))).setMaxAccess("readonly") if mibBuilder.loadTexts: bayStackUnitConfigIndex.setStatus('current') bayStackUnitConfigRearPortAdminMode = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 2, 1, 2), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3))).clone(namedValues=NamedValues(("standalone", 1), ("stacking", 2), ("spb", 3)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: bayStackUnitConfigRearPortAdminMode.setStatus('current') bayStackUnitConfigRearPortOperMode = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 5, 13, 1, 2, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3))).clone(namedValues=NamedValues(("standalone", 1), ("stacking", 2), ("spb", 3)))).setMaxAccess("readonly") if mibBuilder.loadTexts: bayStackUnitConfigRearPortOperMode.setStatus('current') mibBuilder.exportSymbols("BAY-STACK-MIB", bayStackMib=bayStackMib, bayStackUnitConfigIndex=bayStackUnitConfigIndex, bayStackConfigStackErrorNotificationEnabled=bayStackConfigStackErrorNotificationEnabled, PYSNMP_MODULE_ID=bayStackMib, bayStackConfigStackRetryCount=bayStackConfigStackRetryCount, bayStackConfigStackErrorNotificationInterval=bayStackConfigStackErrorNotificationInterval, bayStackUnitConfigRearPortOperMode=bayStackUnitConfigRearPortOperMode, bayStackUnitConfigEntry=bayStackUnitConfigEntry, bayStackConfigStackRebootUnitOnFailure=bayStackConfigStackRebootUnitOnFailure, bayStackObjects=bayStackObjects, bayStackUnitConfigRearPortAdminMode=bayStackUnitConfigRearPortAdminMode, bayStackConfig=bayStackConfig, bayStackUnitConfigTable=bayStackUnitConfigTable, bayStackConfigExpectedStackSize=bayStackConfigExpectedStackSize)
5,498
2,149
#!/usr/bin/env python # This "flattens" a LaTeX document by replacing all # \input{X} lines w/ the text actually contained in X. See # associated README.md for details. # Use as a python module in a python script by saying import flatex then flatex.main(in file, out file) import os import re import sys def is_input(line): """ Determines whether or not a read in line contains an uncommented out \input{} statement. Allows only spaces between start of line and '\input{}'. """ #tex_input_re = r"""^\s*\\input{[^}]*}""" # input only tex_input_re = r"""(^[^\%]*\\input{[^}]*})|(^[^\%]*\\include{[^}]*})""" # input or include return re.search(tex_input_re, line) def get_input(line): """ Gets the file name from a line containing an input statement. """ tex_input_filename_re = r"""{[^}]*""" m = re.search(tex_input_filename_re, line) return m.group()[1:] def combine_path(base_path, relative_ref): """ Combines the base path of the tex document being worked on with the the relate reference found in that document. """ #if (base_path != ""): #print "os.getcwd()", os.getcwd() #os.chdir(base_path) filePath = os.path.abspath(relative_ref) filePath = filePath + ".tex" return filePath def expand_file(base_file): """ Recursively-defined function that takes as input a file and returns it with all the inputs replaced with the contents of the referenced file. """ output_lines = [] f = open(base_file, "r") for line in f: if is_input(line): new_base_file = combine_path(current_path, get_input(line)) output_lines += expand_file(new_base_file) output_lines.append('\n') # add a new line after each file input else: output_lines.append(line) f.close() return output_lines def main(base_file, output_file): g = open(output_file, "w") g.write(''.join(expand_file(base_file))) g.close() return None if __name__ == '__main__': base_file, output_file = sys.argv[1:] current_path = os.path.split(base_file)[0] main(base_file, output_file)
2,184
738
from .encrypt import register
30
8
import datetime import os import pandas from polygon.rest.client import RESTClient def ts_to_datetime(ts) -> str: return datetime.datetime.fromtimestamp(ts / 1000.0).strftime('%Y-%m-%d %H:%M') def pull_day(Symbol, from_): POLYGON_API_KEY = os.environ.get('POLYGON_API_KEY') enddate = datetime.datetime.fromisoformat(from_) enddate += datetime.timedelta(days=1) enddate = str(enddate)[0:10] with RESTClient(POLYGON_API_KEY) as client: resp = client.stocks_equities_aggregates(Symbol, 1, "minute", from_, enddate, unadjusted=False) #print(f"Minute aggregates for {resp.ticker} between {from_} and {enddate}.") out = {} df = pandas.DataFrame(out, columns = ['Datetime', 'Open', 'High', 'Low','Close','Adj Close','Volume']) for result in resp.results: #dt = ts_to_datetime(result["t"]) #print(f"{dt}\n\tO: {result['o']}\n\tH: {result['h']}\n\tL: {result['l']}\n\tC: {result['c']} ") date = {"Datetime": result['t']} open = {"Open": result['o']} high = {"High": result['h']} low = {"Low": result['l']} close = {"Close": result['c']} volume = {"Volume": result['v']} bar = {**date, **open, **high, **low, **close, **volume} df = df.append(bar,ignore_index=True) return(df) # ---------------------------- daystr = "2021-09-10" df = pull_day("LC", daystr) fname = r"M:\data\out.csv" print("Writing: ", fname) df.to_csv (fname, index = False, header=True)
1,600
597
from random import choice, randint from re import sub, split, findall from string import ascii_letters from subprocess import PIPE, Popen from sys import argv, executable, stderr from .responses import pronouns, reactions, remarks def owoify(text): if type(text) == bytes: text = str(text)[2:-1].replace("\\n", "\n") text = sub("[rlv]", "w", text) text = sub("[RLV]", "W", text) text = sub("ee", "wee", text) # This is to convert the string into a array whilst maintaining whitespace words = split(r"\s+", text) whitespace = findall(r"\s+", text) text = [None] * (len(words) + len(whitespace)) text[::2], text[1::2] = words, whitespace # Random stutter for idx, word in enumerate(text): if len(word) > 0: if word[0] in ascii_letters and word[0].lower() not in "aeiouw": if randint(1, 10) == 1: text[idx] = f"{word[0]}-{word}" text = "".join(text) return text def main(): process = Popen([executable] + argv[1:], stderr=PIPE) while process.poll() is None: for line in iter(process.stderr.readline, b""): if line == b"Traceback (most recent call last):\n": # Easter egg :) if randint(1, 10) == 1: stderr.write(f"{choice(pronouns)}, {choice(remarks)}, you sussy baka {choice(reactions)}\n") else: stderr.write(f"{choice(pronouns)}, {choice(remarks)} {choice(reactions)}\n") stderr.write(owoify(line)) if __name__ == "__main__": main()
1,595
545
import os from fabric.api import env, run, sudo, execute, local, settings, \ hide, open_shell, parallel, serial, put from fabric.decorators import hosts from fabric.contrib.console import confirm import fabric.colors as fab_col import paramiko import getpass from tabulate import tabulate file_hosts = "hosts.txt" paramiko.util.log_to_file("paramiko.log") env.colorize_errors = True # The selected hosts are the hosts in env (at the beginning) selected_hosts = env.hosts running_hosts = {} env.connection_attempts = 2 # env.skip_bad_hosts = True def load_hosts(): """ Load hosts from hosts.txt. A host can either be in form username@host[:port] password or username@host[:port] If no port is specified, port 22 is selected. """ with open(file_hosts, "r") as f: data = f.readlines() for line in data: try: host, password = line.strip().split() except ValueError: host = line.strip() password = None if len(host.split(':')) == 1: host = host + ":22" env.hosts.append(host) if password is not None: env.passwords[host] = password.strip() env.hosts = list(set(env.hosts)) # Remove duplicates def add_host(): """ Add a new host to the running hosts. The user can decide whether to add the host also to the external hosts.txt file. """ name = raw_input("Username: ") host = raw_input("Host: ") port = input("Port: ") new_host = name + "@" + host + ":" + str(port) selected_hosts.append(new_host) password = None if confirm("Authenticate using a password? "): password = getpass.getpass("Password: ").strip() env.passwords[new_host] = password # Append the new host to the hosts file if confirm("Add the new host to the hosts file? "): if password is not None: line = new_host + " " + password + "\n" else: line = new_host + "\n" with open(file_hosts, 'a') as f: f.write(line) def print_hosts(): """ Print selected hosts. If hosts haven't been hand-selected yet, all hosts are selected. """ hosts = map(lambda x: [x, env.passwords.get(x, None)], selected_hosts) print(fab_col.green(tabulate(hosts, ["Host", "Password"]))) def check_hosts(): """ Check if hosts are active or not and print the result. """ global running_hosts running_hosts = dict() for host in selected_hosts: print(fab_col.magenta("\nPing host %d of %d" % (selected_hosts.index(host) + 1, len(selected_hosts)))) response = os.system("ping -c 1 " + host.split("@")[1].split(":")[0]) if response == 0: running_hosts[host] = True else: running_hosts[host] = False # Convert running_hosts in order to print it as table mylist = map(lambda index: [index[0], str(index[1])], running_hosts.items()) print(fab_col.green(tabulate(mylist, ["Host", "Running"]))) def select_running_hosts(): """ Select all active hosts. """ global selected_hosts with hide('stdout'): check_hosts() host_up = filter(lambda x: running_hosts.get(x, False), running_hosts.keys()) selected_hosts = host_up def choose_hosts(): """ Select the hosts to be used. """ global selected_hosts mylist = map(lambda (num, h): [num, h], enumerate(env.hosts)) print(fab_col.blue("Select Hosts (space-separated):")) print(fab_col.blue(tabulate(mylist, ["Number", "Host"]))) choices = raw_input("> ").split() # Avoid letters in string index choices = filter(lambda x: x.isdigit(), choices) # Convert to int list choices = map(int, choices) # Avoid IndexError choices = filter(lambda x: x < len(env.hosts), choices) # Remove duplicates choices = list(set(choices)) # If no hosts are selected, keep the current hosts if len(choices) == 0: return # Get only selected hosts selected_hosts = map(lambda i: env.hosts[i], choices) def run_locally(cmd=None): """ Execute a command locally. """ if cmd is None: cmd = raw_input("Insert command: ") with settings(warn_only=True): local(cmd) # This function cannot have the parallel decorator since # a sudo command must receive the user password @serial def _execute_sudo(command): """ Execute a sudo command on a host. Returns: The results of the execution. """ with settings(warn_only=True): return sudo(command[4:].strip(), shell=True) @parallel def _execute_command(command): """ Execute a command on a host. Returns: The results of the execution. """ with settings(warn_only=True): try: return run(command) except: print(fab_col.red("Error execution in host %s" % env.host)) return None @parallel def run_command(cmd=None): """ Execute a command on hosts. """ if cmd is None: cmd = raw_input("Insert command: ") if cmd.strip()[:4] == "sudo": execute(_execute_sudo, cmd, hosts=selected_hosts) else: execute(_execute_command, cmd, hosts=selected_hosts) @hosts(selected_hosts) def execute_script(): """ Execute a script file. """ # Attention to script name. # Add security checks script_file = raw_input("Name of the script: ") remote_path = "~/" if len(script_file) < 4 or ".." in script_file: # Invalid script print(fab_col.red("Error. Invalid script name.")) return for h in selected_hosts: with settings(host_string=h): with hide('running'): put(script_file, remote_path, mode=777) # Remove the path from the name of the script script_file = script_file.split("/")[-1] # Execution extension = script_file.split(".")[-1] if extension == script_file: print(fab_col.red("Invalid script")) return if extension == 'py': run_command("python " + remote_path + script_file) elif extension == "sh" or extension == "bash": run_command("bash " + remote_path + script_file) else: print(fab_col.red("Extension not supported")) # Delete the script with hide('running', 'stdout'): run_command("rm -f " + remote_path + script_file) def open_sh(): """ Open a shell on a host. """ mylist = map(lambda (num, h): [num, h], enumerate(selected_hosts)) print(fab_col.blue(tabulate(mylist, ["Number", "Host"]))) try: n = input("Open shell in host number: ") h = selected_hosts[n] execute(open_shell, host=h) except (NameError, IndexError): print(fab_col.red("Error: invalid host selection.")) print(fab_col.red("Shell not opened."))
6,966
2,104
objects = {} def instantiate(): # This function is called once during server startup. Modify the global 'objects' dict with of instantiated # shared objects that you wish to store in the parent process and have access to from child request handler # processes. Each object must support being shared via the multiproccessing module or else the object will # just be copied into the children. See http://docs.python.org/library/multiprocessing.html # # For example, in this function you might put: # # import multiprocessing # objects['num_requests'] = multiprocessing.Value('i',0) # # And in your request handler, put: # # from magnum.shared import objects # objects['num_requests'].value += 1 return
781
214
# Copyright 2017 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. import json from google.appengine.ext import ndb from google.protobuf import json_format from google.protobuf import struct_pb2 from protorpc import messages from components import utils from proto import common_pb2 import bbutil import config import logging import model # Names of well-known parameters. BUILDER_PARAMETER = 'builder_name' PROPERTIES_PARAMETER = 'properties' def format_luci_bucket(bucket_id): """Returns V1 luci bucket name, e.g. "luci.chromium.try".""" return 'luci.%s.%s' % config.parse_bucket_id(bucket_id) def parse_luci_bucket(bucket): """Converts V1 LUCI bucket to a bucket ID string. Returns '' if bucket is not a LUCI bucket. """ parts = bucket.split('.', 2) if len(parts) == 3 and parts[0] == 'luci': return config.format_bucket_id(parts[1], parts[2]) return '' @ndb.tasklet def to_bucket_id_async(bucket): """Converts a bucket string to a bucket id. A bucket string is either a bucket id (e.g. "chromium/try") or a legacy bucket name (e.g. "master.tryserver.x", "luci.chromium.try"). Does not check access. Returns: bucket id string or None if such bucket does not exist. Raises: errors.InvalidInputError if bucket is invalid or ambiguous. """ is_legacy = config.is_legacy_bucket_id(bucket) if not is_legacy: config.validate_bucket_id(bucket) raise ndb.Return(bucket) config.validate_bucket_name(bucket) bucket_id = parse_luci_bucket(bucket) if bucket_id: raise ndb.Return(bucket_id) # The slowest code path. # Does not apply to LUCI. bucket_id = config.resolve_bucket_name_async(bucket).get_result() if bucket_id: logging.info('resolved bucket id %r => %r', bucket, bucket_id) raise ndb.Return(bucket_id) class CanaryPreference(messages.Enum): # The build system will decide whether to use canary or not AUTO = 1 # Use the production build infrastructure PROD = 2 # Use the canary build infrastructure CANARY = 3 CANARY_PREFERENCE_TO_TRINARY = { CanaryPreference.AUTO: common_pb2.UNSET, CanaryPreference.PROD: common_pb2.NO, CanaryPreference.CANARY: common_pb2.YES, } TRINARY_TO_CANARY_PREFERENCE = { v: k for k, v in CANARY_PREFERENCE_TO_TRINARY.iteritems() } class BuildMessage(messages.Message): """Describes model.Build, see its docstring.""" id = messages.IntegerField(1, required=True) bucket = messages.StringField(2, required=True) tags = messages.StringField(3, repeated=True) parameters_json = messages.StringField(4) status = messages.EnumField(model.BuildStatus, 5) result = messages.EnumField(model.BuildResult, 6) result_details_json = messages.StringField(7) failure_reason = messages.EnumField(model.FailureReason, 8) cancelation_reason = messages.EnumField(model.CancelationReason, 9) lease_expiration_ts = messages.IntegerField(10) lease_key = messages.IntegerField(11) url = messages.StringField(12) created_ts = messages.IntegerField(13) started_ts = messages.IntegerField(20) updated_ts = messages.IntegerField(14) completed_ts = messages.IntegerField(15) created_by = messages.StringField(16) status_changed_ts = messages.IntegerField(17) utcnow_ts = messages.IntegerField(18, required=True) retry_of = messages.IntegerField(19) canary_preference = messages.EnumField(CanaryPreference, 21) canary = messages.BooleanField(22) project = messages.StringField(23) experimental = messages.BooleanField(24) service_account = messages.StringField(25) def proto_to_timestamp(ts): if not ts.seconds: return None return utils.datetime_to_timestamp(ts.ToDatetime()) def legacy_bucket_name(bucket_id, is_luci): if is_luci: # In V1, LUCI builds use a "long" bucket name, e.g. "luci.chromium.try" # as opposed to just "try". This is because in the past bucket names # were globally unique, as opposed to unique per project. return format_luci_bucket(bucket_id) _, bucket_name = config.parse_bucket_id(bucket_id) return bucket_name # List of deprecated properties that are converted from float to int for # backward compatibility. # TODO(crbug.com/877161): remove this list. INTEGER_PROPERTIES = [ 'buildnumber', 'issue', 'patchset', 'patch_issue', 'patch_set', ] def get_build_url(build): """Returns view URL of the build.""" if build.url: return build.url settings = config.get_settings_async().get_result() return 'https://%s/b/%d' % (settings.swarming.milo_hostname, build.proto.id) def properties_to_json(properties): """Converts properties to JSON. properties should be struct_pb2.Struct, but for convenience in tests a dict is also accepted. CAUTION: in general converts all numbers to floats, because JSON format does not distinguish floats and ints. For backward compatibility, temporarily (crbug.com/877161) renders widely used, deprecated properties as integers, see INTEGER_PROPERTIES. """ return json.dumps(_properties_to_dict(properties), sort_keys=True) def _properties_to_dict(properties): """Implements properties_to_json.""" assert isinstance(properties, (dict, struct_pb2.Struct)), properties if isinstance(properties, dict): # pragma: no branch properties = bbutil.dict_to_struct(properties) # Note: this dict does not necessarily equal the original one. # In particular, an int may turn into a float. as_dict = json_format.MessageToDict(properties) for p in INTEGER_PROPERTIES: if isinstance(as_dict.get(p), float): as_dict[p] = int(as_dict[p]) return as_dict def build_to_message(build_bundle, include_lease_key=False): """Converts a model.BuildBundle to BuildMessage.""" build = build_bundle.build assert build assert build.key assert build.key.id() bp = build.proto infra = build_bundle.infra.parse() sw = infra.swarming logdog = infra.logdog recipe = infra.recipe result_details = (build.result_details or {}).copy() result_details['properties'] = {} if build_bundle.output_properties: # pragma: no branch result_details['properties'] = _properties_to_dict( build_bundle.output_properties.parse() ) if bp.summary_markdown: result_details['ui'] = {'info': bp.summary_markdown} parameters = (build.parameters or {}).copy() parameters[BUILDER_PARAMETER] = bp.builder.builder parameters[PROPERTIES_PARAMETER] = _properties_to_dict( infra.buildbucket.requested_properties ) recipe_name = recipe.name if build_bundle.input_properties: # pragma: no cover input_props = build_bundle.input_properties.parse() if 'recipe' in input_props.fields: recipe_name = input_props['recipe'] if bp.status != common_pb2.SUCCESS and bp.summary_markdown: result_details['error'] = { 'message': bp.summary_markdown, } if sw.bot_dimensions: by_key = {} for d in sw.bot_dimensions: by_key.setdefault(d.key, []).append(d.value) result_details.setdefault('swarming', {})['bot_dimensions'] = by_key tags = set(build.tags) if build.is_luci: tags.add('swarming_hostname:%s' % sw.hostname) tags.add('swarming_task_id:%s' % sw.task_id) # Milo uses swarming tags. tags.add('swarming_tag:recipe_name:%s' % recipe_name) tags.add( 'swarming_tag:recipe_package:%s' % (bp.exe.cipd_package or recipe.cipd_package) ) tags.add( 'swarming_tag:log_location:logdog://%s/%s/%s/+/annotations' % (logdog.hostname, logdog.project, logdog.prefix) ) tags.add('swarming_tag:luci_project:%s' % bp.builder.project) # Try to find OS for d in sw.bot_dimensions: if d.key == 'os': tags.add('swarming_tag:os:%s' % d.value) break msg = BuildMessage( id=build.key.id(), project=bp.builder.project, bucket=legacy_bucket_name(build.bucket_id, build.is_luci), tags=sorted(tags), parameters_json=json.dumps(parameters, sort_keys=True), status=build.status_legacy, result=build.result, result_details_json=json.dumps(result_details, sort_keys=True), cancelation_reason=build.cancelation_reason, failure_reason=build.failure_reason, lease_key=build.lease_key if include_lease_key else None, url=get_build_url(build), created_ts=proto_to_timestamp(bp.create_time), started_ts=proto_to_timestamp(bp.start_time), updated_ts=proto_to_timestamp(bp.update_time), completed_ts=proto_to_timestamp(bp.end_time), created_by=build.created_by.to_bytes() if build.created_by else None, status_changed_ts=utils.datetime_to_timestamp(build.status_changed_time), utcnow_ts=utils.datetime_to_timestamp(utils.utcnow()), retry_of=build.retry_of, canary_preference=( # This is not accurate, but it does not matter at this point. # This is deprecated. CanaryPreference.CANARY if build.canary else CanaryPreference.PROD ), canary=build.canary, experimental=build.experimental, service_account=sw.task_service_account, # when changing this function, make sure build_to_dict would still work ) if build.lease_expiration_date is not None: msg.lease_expiration_ts = utils.datetime_to_timestamp( build.lease_expiration_date ) return msg def build_to_dict(build_bundle, include_lease_key=False): """Converts a build to an externally consumable dict. This function returns a dict that a BuildMessage would be encoded to. """ # Implementing this function in a generic way (message_to_dict) requires # knowledge of many protorpc and endpoints implementation details. # Not worth it. msg = build_to_message(build_bundle, include_lease_key=include_lease_key) # Special cases result = { 'tags': msg.tags, # a list } for f in msg.all_fields(): v = msg.get_assigned_value(f.name) if f.name in result or v is None: # None is the default. It is omitted by Cloud Endpoints. continue if isinstance(v, messages.Enum): v = str(v) else: assert isinstance(v, (basestring, int, long, bool)), v if (isinstance(f, messages.IntegerField) and f.variant == messages.Variant.INT64): v = str(v) result[f.name] = v return result
10,398
3,539
""" This package implements random variables. Random variables are the primary in- and outputs of probabilistic numerical methods. A generic signature of such methods looks like this: .. highlight:: python .. code-block:: python randvar_out, info = probnum_method(problem, randvar_in, **kwargs) """ from ._dirac import Dirac from ._normal import Normal from ._random_variable import ( ContinuousRandomVariable, DiscreteRandomVariable, RandomVariable, ) from ._scipy_stats import ( WrappedSciPyContinuousRandomVariable, WrappedSciPyDiscreteRandomVariable, WrappedSciPyRandomVariable, ) from ._utils import asrandvar # Public classes and functions. Order is reflected in documentation. __all__ = [ "asrandvar", "RandomVariable", "DiscreteRandomVariable", "ContinuousRandomVariable", "Dirac", "Normal", "WrappedSciPyRandomVariable", "WrappedSciPyDiscreteRandomVariable", "WrappedSciPyContinuousRandomVariable", ] # Set correct module paths. Corrects links and module paths in documentation. RandomVariable.__module__ = "probnum.random_variables" DiscreteRandomVariable.__module__ = "probnum.random_variables" ContinuousRandomVariable.__module__ = "probnum.random_variables" WrappedSciPyRandomVariable.__module__ = "probnum.random_variables" WrappedSciPyDiscreteRandomVariable.__module__ = "probnum.random_variables" WrappedSciPyContinuousRandomVariable.__module__ = "probnum.random_variables" Dirac.__module__ = "probnum.random_variables" Normal.__module__ = "probnum.random_variables" asrandvar.__module__ = "probnum.random_variables"
1,606
474
from app import App import requests import json import polling2 from behave import step from openshift import Openshift from util import substitute_scenario_id from string import Template class GenericTestApp(App): deployment_name_pattern = "{name}" def __init__(self, name, namespace, app_image="ghcr.io/multi-arch/sbo-generic-test-app:latest"): App.__init__(self, name, namespace, app_image, "8080") def get_env_var_value(self, name): resp = polling2.poll(lambda: requests.get(url=f"http://{self.route_url}/env/{name}"), check_success=lambda r: r.status_code in [200, 404], step=5, timeout=400, ignore_exceptions=(requests.exceptions.ConnectionError,)) print(f'env endpoint response: {resp.text} code: {resp.status_code}') if resp.status_code == 200: return json.loads(resp.text) else: return None def format_pattern(self, pattern): return pattern.format(name=self.name) def get_file_value(self, file_path): resp = polling2.poll(lambda: requests.get(url=f"http://{self.route_url}{file_path}"), check_success=lambda r: r.status_code == 200, step=5, timeout=400, ignore_exceptions=(requests.exceptions.ConnectionError,)) print(f'file endpoint response: {resp.text} code: {resp.status_code}') return resp.text def assert_file_not_exist(self, file_path): polling2.poll(lambda: requests.get(url=f"http://{self.route_url}{file_path}"), check_success=lambda r: r.status_code == 404, step=5, timeout=400, ignore_exceptions=(requests.exceptions.ConnectionError,)) def set_label(self, label): self.openshift.set_label(self.name, label, self.namespace) @step(u'Generic test application "{application_name}" is running') @step(u'Generic test application "{application_name}" is running with binding root as "{bindingRoot}"') @step(u'Generic test application is running') @step(u'Generic test application is running with binding root as "{bindingRoot}"') def is_running(context, application_name=None, bindingRoot=None, asDeploymentConfig=False): if application_name is None: application_name = substitute_scenario_id(context) application = GenericTestApp(application_name, context.namespace.name) if asDeploymentConfig: application.resource = "deploymentconfig" if not application.is_running(): print("application is not running, trying to import it") application.install(bindingRoot=bindingRoot) context.application = application # save the generation number context.original_application_generation = application.get_generation() context.latest_application_generation = application.get_generation() @step(u'Generic test application is running as deployment config') def is_running_deployment_config(context): is_running(context, asDeploymentConfig=True) @step(u'The application env var "{name}" has value "{value}"') def check_env_var_value(context, name, value): value = substitute_scenario_id(context, value) found = polling2.poll(lambda: context.application.get_env_var_value(name) == value, step=5, timeout=400) assert found, f'Env var "{name}" should contain value "{value}"' @step(u'The env var "{name}" is not available to the application') def check_env_var_existence(context, name): output = polling2.poll(lambda: context.application.get_env_var_value(name) is None, step=5, timeout=400) assert output, f'Env var "{name}" should not exist' @step(u'Content of file "{file_path}" in application pod is') def check_file_value(context, file_path): value = Template(context.text.strip()).substitute(NAMESPACE=context.namespace.name) resource = substitute_scenario_id(context, file_path) polling2.poll(lambda: context.application.get_file_value(resource) == value, step=5, timeout=400) @step(u'File "{file_path}" exists in application pod') def check_file_exists(context, file_path): resource = substitute_scenario_id(context, file_path) polling2.poll(lambda: context.application.get_file_value(resource) != "", step=5, timeout=400) @step(u'File "{file_path}" is unavailable in application pod') def check_file_unavailable(context, file_path): context.application.assert_file_not_exist(file_path) @step(u'Test applications "{first_app_name}" and "{second_app_name}" is running') def are_two_apps_running(context, first_app_name, second_app_name, bindingRoot=None): application1 = GenericTestApp(first_app_name, context.namespace.name) if not application1.is_running(): print("application1 is not running, trying to import it") application1.install(bindingRoot=bindingRoot) context.application1 = application1 application2 = GenericTestApp(second_app_name, context.namespace.name) if not application2.is_running(): print("application2 is not running, trying to import it") application2.install(bindingRoot=bindingRoot) context.application2 = application2 @step(u'The common label "{label}" is set for both apps') def set_common_label(context, label): context.application1.set_label(f"{label}") context.application2.set_label(f"{label}") @step(u'The application env var "{name}" has value "{value}" in both apps') def check_env_var_value_in_both_apps(context, name, value): polling2.poll(lambda: context.application1.get_env_var_value(name) == value, step=5, timeout=400) polling2.poll(lambda: context.application2.get_env_var_value(name) == value, step=5, timeout=400) @step(u'The container declared in application resource contains env "{envVar}" set only once') @step(u'The container declared in application "{app_name}" resource contains env "{envVar}" set only once') def check_env_var_count_set_on_container(context, envVar, app_name=None): openshift = Openshift() if app_name is None: app_name = context.application.name app_name = substitute_scenario_id(context, app_name) env = openshift.get_deployment_env_info(app_name, context.namespace.name) assert str(env).count(envVar) == 1
6,147
1,878
import json import logging from math import sqrt, floor, ceil, log10 from copy import deepcopy logger = logging.getLogger(__name__) def limit_pairs(pairs): # Hack to limit aspect ratios when there are a lot of choices if len(pairs) > 12: new_pairs = [] log10_aspect_ratios = [-0.3, 0, 0.3] for l in log10_aspect_ratios: best_pair = min((abs(log10(newy) - log10(newx) - l), (newx, newy)) for newx, newy in pairs)[1] new_pairs.append(best_pair) return new_pairs else: return pairs def add_primitive(primitives, block_name, block_args): if block_name in primitives: if not primitives[block_name] == block_args: logger.warning(f"Distinct devices mapped to the same primitive {block_name}: \ existing: {primitives[block_name]}\ new: {block_args}") else: logger.debug(f"Found primitive {block_name} with {block_args}") if 'x_cells' in block_args and 'y_cells' in block_args: x, y = block_args['x_cells'], block_args['y_cells'] pairs = set() m = x*y y_sqrt = floor(sqrt(x*y)) for y in range(y_sqrt, 0, -1): if m % y == 0: pairs.add((y, m//y)) pairs.add((m//y, y)) if y == 1: break pairs = limit_pairs((pairs)) for newx, newy in pairs: concrete_name = f'{block_name}_X{newx}_Y{newy}' if concrete_name not in primitives: primitives[concrete_name] = deepcopy(block_args) primitives[concrete_name]['x_cells'] = newx primitives[concrete_name]['y_cells'] = newy primitives[concrete_name]['abstract_template_name'] = block_name primitives[concrete_name]['concrete_template_name'] = concrete_name else: primitives[block_name] = block_args primitives[block_name]['abstract_template_name'] = block_name primitives[block_name]['concrete_template_name'] = block_name def gen_param(subckt, primitives, pdk_dir): block_name = subckt.name vt = subckt.elements[0].model values = subckt.elements[0].parameters generator_name = subckt.generator["name"] block_name = subckt.name generator_name = subckt.generator["name"] layers_json = pdk_dir / "layers.json" with open(layers_json, "rt") as fp: pdk_data = json.load(fp) design_config = pdk_data["design_info"] if len(subckt.elements) == 1: values = subckt.elements[0].parameters else: mvalues = {} for ele in subckt.elements: mvalues[ele.name] = ele.parameters if generator_name == 'CAP': assert float(values["VALUE"]) or float(values["C"]), f"unidentified size {values} for {block_name}" if "C" in values: size = round(float(values["C"]) * 1E15, 4) elif 'VALUE' in values: size = round(float(values["VALUE"]) * 1E15, 4) assert size <= design_config["max_size_cap"], f"caps larger that {design_config['max_size_cap']}fF are not supported" # TODO: use float in name logger.debug(f"Generating capacitor for:{block_name}, {size}") block_args = { 'primitive': generator_name, 'value': float(size) } add_primitive(primitives, block_name, block_args) elif generator_name == 'RES': assert float(values["VALUE"]) or float(values["R"]), f"unidentified size {values['VALUE']} for {name}" if "R" in values: size = round(float(values["R"]), 2) elif 'VALUE' in values: size = round(float(values["VALUE"]), 2) # TODO: use float in name if size.is_integer(): size = int(size) height = ceil(sqrt(float(size) / design_config["unit_height_res"])) logger.debug(f'Generating resistor for: {block_name} {size}') block_args = { 'primitive': generator_name, 'value': (height, float(size)) } add_primitive(primitives, block_name, block_args) else: assert 'MOS' == generator_name, f'{generator_name} is not recognized' if "vt_type" in design_config: vt = [vt.upper() for vt in design_config["vt_type"] if vt.upper() in subckt.elements[0].model] mvalues = {} for ele in subckt.elements: mvalues[ele.name] = ele.parameters device_name_all = [*mvalues.keys()] device_name = next(iter(mvalues)) for key in mvalues: assert mvalues[key]["W"] != str, f"unrecognized size of device {key}:{mvalues[key]['W']} in {block_name}" assert int( float(mvalues[key]["W"])*1E+9) % design_config["Fin_pitch"] == 0, \ f"Width of device {key} in {block_name} should be multiple of fin pitch:{design_config['Fin_pitch']}" size = int(float(mvalues[key]["W"])*1E+9/design_config["Fin_pitch"]) mvalues[key]["NFIN"] = size name_arg = 'NFIN'+str(size) if 'NF' in mvalues[device_name].keys(): for key in mvalues: assert int(mvalues[key]["NF"]), f"unrecognized NF of device {key}:{mvalues[key]['NF']} in {name}" assert int(mvalues[key]["NF"]) % 2 == 0, f"NF must be even for device {key}:{mvalues[key]['NF']} in {name}" name_arg = name_arg+'_NF'+str(int(mvalues[device_name]["NF"])) if 'M' in mvalues[device_name].keys(): for key in mvalues: assert int(mvalues[key]["M"]), f"unrecognized M of device {key}:{mvalues[key]['M']} in {name}" if "PARALLEL" in mvalues[key].keys() and int(mvalues[key]['PARALLEL']) > 1: mvalues[key]["PARALLEL"] = int(mvalues[key]['PARALLEL']) mvalues[key]['M'] = int(mvalues[key]['M'])*int(mvalues[key]['PARALLEL']) name_arg = name_arg+'_M'+str(int(mvalues[device_name]["M"])) size = 0 logger.debug(f"Generating lef for {block_name}") if isinstance(size, int): for key in mvalues: assert int(mvalues[device_name]["NFIN"]) == int(mvalues[key]["NFIN"]), f"W should be same for all devices in {name} {mvalues}" size_device = int(mvalues[key]["NF"])*int(mvalues[key]["M"]) size = size + size_device no_units = ceil(size / (2*len(mvalues))) # Factor 2 is due to NF=2 in each unit cell; needs to be generalized if any(x in block_name for x in ['DP', '_S']) and floor(sqrt(no_units/3)) >= 1: square_y = floor(sqrt(no_units/3)) else: square_y = floor(sqrt(no_units)) while no_units % square_y != 0: square_y -= 1 yval = square_y xval = int(no_units / square_y) if 'SCM' in block_name: if int(mvalues[device_name_all[0]]["NFIN"])*int(mvalues[device_name_all[0]]["NF"])*int(mvalues[device_name_all[0]]["M"]) != \ int(mvalues[device_name_all[1]]["NFIN"])*int(mvalues[device_name_all[1]]["NF"])*int(mvalues[device_name_all[1]]["M"]): square_y = 1 yval = square_y xval = int(no_units / square_y) block_args = { 'primitive': generator_name, 'value': mvalues[device_name]["NFIN"], 'x_cells': xval, 'y_cells': yval, 'parameters': mvalues } if 'STACK' in mvalues[device_name].keys() and int(mvalues[device_name]["STACK"]) > 1: block_args['stack'] = int(mvalues[device_name]["STACK"]) if vt: block_args['vt_type'] = vt[0] add_primitive(primitives, block_name, block_args) return True
8,033
2,595
from rest_framework.throttling import SimpleRateThrottle class ForcedPhotRateThrottle(SimpleRateThrottle): """ Limits the rate of API calls by different amounts for various parts of the API. Any view that has the `throttle_scope` property set will be throttled. The unique cache key will be generated by concatenating the user id of the request, and the scope of the view being accessed. """ scope_attr = 'throttle_scope' def __init__(self): # Override the usual SimpleRateThrottle, because we can't determine # the rate until called by the view. pass def allow_request(self, request, view): if request.method == 'GET': return True # We can only determine the scope once we're called by the view. self.scope = getattr(view, self.scope_attr, None) # If a view does not have a `throttle_scope` always allow the request if not self.scope: return True # Determine the allowed request rate as we normally would during # the `__init__` call. self.rate = self.get_rate() self.num_requests, self.duration = self.parse_rate(self.rate) # We can now proceed as normal. return super().allow_request(request, view) def get_cache_key(self, request, view): """ If `view.throttle_scope` is not set, don't apply this throttle. Otherwise generate the unique cache key by concatenating the user id with the '.throttle_scope` property of the view. """ if request.user.is_authenticated: ident = request.user.pk else: ident = self.get_ident(request) return self.cache_format % { 'scope': self.scope, 'ident': ident }
1,804
497
# encoding: utf-8 from test.vim_test_case import VimTestCase as _VimTest from test.constant import * from test.util import running_on_windows class SnippetOptions_OnlyExpandWhenWSInFront_Expand(_VimTest): snippets = ("test", "Expand me!", "", "b") keys = "test" + EX wanted = "Expand me!" class SnippetOptions_OnlyExpandWhenWSInFront_Expand2(_VimTest): snippets = ("test", "Expand me!", "", "b") keys = " test" + EX wanted = " Expand me!" class SnippetOptions_OnlyExpandWhenWSInFront_DontExpand(_VimTest): snippets = ("test", "Expand me!", "", "b") keys = "a test" + EX wanted = "a test" + EX class SnippetOptions_OnlyExpandWhenWSInFront_OneWithOneWO(_VimTest): snippets = (("test", "Expand me!", "", "b"), ("test", "not at beginning", "", "")) keys = "a test" + EX wanted = "a not at beginning" class SnippetOptions_OnlyExpandWhenWSInFront_OneWithOneWOChoose(_VimTest): snippets = (("test", "Expand me!", "", "b"), ("test", "not at beginning", "", "")) keys = " test" + EX + "1\n" wanted = " Expand me!" class SnippetOptions_ExpandInwordSnippets_SimpleExpand(_VimTest): snippets = (("test", "Expand me!", "", "i"),) keys = "atest" + EX wanted = "aExpand me!" class SnippetOptions_ExpandInwordSnippets_ExpandSingle(_VimTest): snippets = (("test", "Expand me!", "", "i"),) keys = "test" + EX wanted = "Expand me!" class SnippetOptions_ExpandInwordSnippetsWithOtherChars_Expand(_VimTest): snippets = (("test", "Expand me!", "", "i"),) keys = "$test" + EX wanted = "$Expand me!" class SnippetOptions_ExpandInwordSnippetsWithOtherChars_Expand2(_VimTest): snippets = (("test", "Expand me!", "", "i"),) keys = "-test" + EX wanted = "-Expand me!" class SnippetOptions_ExpandInwordSnippetsWithOtherChars_Expand3(_VimTest): skip_if = lambda self: running_on_windows() snippets = (("test", "Expand me!", "", "i"),) keys = "ßßtest" + EX wanted = "ßßExpand me!" class _SnippetOptions_ExpandWordSnippets(_VimTest): snippets = (("test", "Expand me!", "", "w"),) class SnippetOptions_ExpandWordSnippets_NormalExpand( _SnippetOptions_ExpandWordSnippets ): keys = "test" + EX wanted = "Expand me!" class SnippetOptions_ExpandWordSnippets_NoExpand(_SnippetOptions_ExpandWordSnippets): keys = "atest" + EX wanted = "atest" + EX class SnippetOptions_ExpandWordSnippets_ExpandSuffix( _SnippetOptions_ExpandWordSnippets ): keys = "a-test" + EX wanted = "a-Expand me!" class SnippetOptions_ExpandWordSnippets_ExpandSuffix2( _SnippetOptions_ExpandWordSnippets ): keys = "a(test" + EX wanted = "a(Expand me!" class SnippetOptions_ExpandWordSnippets_ExpandSuffix3( _SnippetOptions_ExpandWordSnippets ): keys = "[[test" + EX wanted = "[[Expand me!" class _No_Tab_Expand(_VimTest): snippets = ("test", "\t\tExpand\tme!\t", "", "t") class No_Tab_Expand_Simple(_No_Tab_Expand): keys = "test" + EX wanted = "\t\tExpand\tme!\t" class No_Tab_Expand_Leading_Spaces(_No_Tab_Expand): keys = " test" + EX wanted = " \t\tExpand\tme!\t" class No_Tab_Expand_Leading_Tabs(_No_Tab_Expand): keys = "\ttest" + EX wanted = "\t\t\tExpand\tme!\t" class No_Tab_Expand_No_TS(_No_Tab_Expand): def _extra_vim_config(self, vim_config): vim_config.append("set sw=3") vim_config.append("set sts=3") keys = "test" + EX wanted = "\t\tExpand\tme!\t" class No_Tab_Expand_ET(_No_Tab_Expand): def _extra_vim_config(self, vim_config): vim_config.append("set sw=3") vim_config.append("set expandtab") keys = "test" + EX wanted = "\t\tExpand\tme!\t" class No_Tab_Expand_ET_Leading_Spaces(_No_Tab_Expand): def _extra_vim_config(self, vim_config): vim_config.append("set sw=3") vim_config.append("set expandtab") keys = " test" + EX wanted = " \t\tExpand\tme!\t" class No_Tab_Expand_ET_SW(_No_Tab_Expand): def _extra_vim_config(self, vim_config): vim_config.append("set sw=8") vim_config.append("set expandtab") keys = "test" + EX wanted = "\t\tExpand\tme!\t" class No_Tab_Expand_ET_SW_TS(_No_Tab_Expand): def _extra_vim_config(self, vim_config): vim_config.append("set sw=3") vim_config.append("set sts=3") vim_config.append("set ts=3") vim_config.append("set expandtab") keys = "test" + EX wanted = "\t\tExpand\tme!\t" class _TabExpand_RealWorld: snippets = ( "hi", r"""hi `!p snip.rv="i1\n" snip.rv += snip.mkline("i1\n") snip.shift(1) snip.rv += snip.mkline("i2\n") snip.unshift(2) snip.rv += snip.mkline("i0\n") snip.shift(3) snip.rv += snip.mkline("i3")` snip.rv = repr(snip.rv) End""", ) class No_Tab_Expand_RealWorld(_TabExpand_RealWorld, _VimTest): def _extra_vim_config(self, vim_config): vim_config.append("set noexpandtab") keys = "\t\thi" + EX wanted = """\t\thi \t\ti1 \t\ti1 \t\t\ti2 \ti0 \t\t\t\ti3 \t\tsnip.rv = repr(snip.rv) \t\tEnd""" class SnippetOptions_Regex_Expand(_VimTest): snippets = ("(test)", "Expand me!", "", "r") keys = "test" + EX wanted = "Expand me!" class SnippetOptions_Regex_WithSpace(_VimTest): snippets = ("test ", "Expand me!", "", "r") keys = "test " + EX wanted = "Expand me!" class SnippetOptions_Regex_Multiple(_VimTest): snippets = ("(test *)+", "Expand me!", "", "r") keys = "test test test" + EX wanted = "Expand me!" class _Regex_Self(_VimTest): snippets = ("((?<=\W)|^)(\.)", "self.", "", "r") class SnippetOptions_Regex_Self_Start(_Regex_Self): keys = "." + EX wanted = "self." class SnippetOptions_Regex_Self_Space(_Regex_Self): keys = " ." + EX wanted = " self." class SnippetOptions_Regex_Self_TextAfter(_Regex_Self): keys = " .a" + EX wanted = " .a" + EX class SnippetOptions_Regex_Self_TextBefore(_Regex_Self): keys = "a." + EX wanted = "a." + EX class SnippetOptions_Regex_PythonBlockMatch(_VimTest): snippets = ( r"([abc]+)([def]+)", r"""`!p m = match snip.rv += m.group(2) snip.rv += m.group(1) `""", "", "r", ) keys = "test cabfed" + EX wanted = "test fedcab" class SnippetOptions_Regex_PythonBlockNoMatch(_VimTest): snippets = (r"cabfed", r"""`!p snip.rv = match or "No match"`""") keys = "test cabfed" + EX wanted = "test No match" # Tests for Bug #691575 class SnippetOptions_Regex_SameLine_Long_End(_VimTest): snippets = ("(test.*)", "Expand me!", "", "r") keys = "test test abc" + EX wanted = "Expand me!" class SnippetOptions_Regex_SameLine_Long_Start(_VimTest): snippets = ("(.*test)", "Expand me!", "", "r") keys = "abc test test" + EX wanted = "Expand me!" class SnippetOptions_Regex_SameLine_Simple(_VimTest): snippets = ("(test)", "Expand me!", "", "r") keys = "abc test test" + EX wanted = "abc test Expand me!" class MultiWordSnippet_Simple(_VimTest): snippets = ("test me", "Expand me!") keys = "test me" + EX wanted = "Expand me!" class MultiWord_SnippetOptions_OnlyExpandWhenWSInFront_Expand(_VimTest): snippets = ("test it", "Expand me!", "", "b") keys = "test it" + EX wanted = "Expand me!" class MultiWord_SnippetOptions_OnlyExpandWhenWSInFront_Expand2(_VimTest): snippets = ("test it", "Expand me!", "", "b") keys = " test it" + EX wanted = " Expand me!" class MultiWord_SnippetOptions_OnlyExpandWhenWSInFront_DontExpand(_VimTest): snippets = ("test it", "Expand me!", "", "b") keys = "a test it" + EX wanted = "a test it" + EX class MultiWord_SnippetOptions_OnlyExpandWhenWSInFront_OneWithOneWO(_VimTest): snippets = ( ("test it", "Expand me!", "", "b"), ("test it", "not at beginning", "", ""), ) keys = "a test it" + EX wanted = "a not at beginning" class MultiWord_SnippetOptions_OnlyExpandWhenWSInFront_OneWithOneWOChoose(_VimTest): snippets = ( ("test it", "Expand me!", "", "b"), ("test it", "not at beginning", "", ""), ) keys = " test it" + EX + "1\n" wanted = " Expand me!" class MultiWord_SnippetOptions_ExpandInwordSnippets_SimpleExpand(_VimTest): snippets = (("test it", "Expand me!", "", "i"),) keys = "atest it" + EX wanted = "aExpand me!" class MultiWord_SnippetOptions_ExpandInwordSnippets_ExpandSingle(_VimTest): snippets = (("test it", "Expand me!", "", "i"),) keys = "test it" + EX wanted = "Expand me!" class _MultiWord_SnippetOptions_ExpandWordSnippets(_VimTest): snippets = (("test it", "Expand me!", "", "w"),) class MultiWord_SnippetOptions_ExpandWordSnippets_NormalExpand( _MultiWord_SnippetOptions_ExpandWordSnippets ): keys = "test it" + EX wanted = "Expand me!" class MultiWord_SnippetOptions_ExpandWordSnippets_NoExpand( _MultiWord_SnippetOptions_ExpandWordSnippets ): keys = "atest it" + EX wanted = "atest it" + EX class MultiWord_SnippetOptions_ExpandWordSnippets_ExpandSuffix( _MultiWord_SnippetOptions_ExpandWordSnippets ): keys = "a-test it" + EX wanted = "a-Expand me!"
9,206
3,486
from scipy.optimize import minimize class Solver(object): # initialize the object def __init__(self, servers, alpha, teta, lambda_s): self.c_prime = 0 self.teta = teta self.alpha = alpha self.lambda_s = lambda_s # fill servers specification in arrays for solver use self.k = [] self.a = [] self.mu = [] self.lambs = [] self.bound = [] self.active = [] for server in servers: self.a.append((server.e_p - server.e_i) / server.mu * server.k) self.k.append(server.k) self.mu.append(server.mu) self.lambs.append(server.lamb) self.active.append(server.active) # solve sub-problem def solve(self, linear): if linear: return self.linear_solver() else: return self.non_linear_solver() # implementation of linear solver # this method uses cvxopt solver # visit http://cvxopt.org for more information def linear_solver(self): print("Initiating linear solver") # implementation of non-linear solver # this method uses scipy solver # visit http://docs.scipy.org for more information def non_linear_solver(self): print("Initiating non-linear solver") for active in self.active: print(active) cons = [{'type': 'eq', 'fun': self.workload_preserving_constraint}] cons += self.inactive_server_constraint(self.lambs) cons += self.positive_variables(self.lambs) res = minimize(self.objective_function, self.lambs, method='SLSQP', bounds=None, constraints=tuple(cons)) print(res.x) self.c_prime = res.fun return res.x # definition of objective function def objective_function(self, x): objective = 0 for index, lamb in enumerate(x): if lamb == 0: continue objective += self.teta * (lamb / (self.mu[index] - lamb / self.k[index])) + self.alpha * ( self.a[index] * lamb) return objective # definition of workload preserving constraint def workload_preserving_constraint(self, x): constraint = 0 for index, lamb in enumerate(x): constraint += lamb constraint -= self.lambda_s return constraint # definition of inactive server constraint using "Big M" method def inactive_server_constraint(self, x): constraints = [] for index, lamb in enumerate(x): if self.active[index] == 0: continue constraints.append({'type': 'ineq', 'fun': lambda x: self.lambda_s - x[index]}) return constraints # all variables must be positive def positive_variables(self, x): constraints = [] for index, lamb in enumerate(x): if self.active[index] == 0: constraints.append({'type': 'eq', 'fun': lambda x: x[index]}) else: constraints.append({'type': 'ineq', 'fun': lambda x: x[index]}) return constraints
3,119
910
# This code is licensed under the MIT License (see LICENSE file for details) import threading import time from ..util import smart_serial from ..util import property_device from ..util import state_stack from ..config import scope_configuration from . import iotool def _make_dac_bytes(IIC_Addr, bit): dac_bytes = bytearray(b'\x53\x00\x03\x00\x00\x00\x50') dac_bytes[1] = IIC_Addr dac_bytes[3] = 1 << bit return dac_bytes LAMP_DAC_COMMANDS = { 'uv': _make_dac_bytes(0x18, 0), 'blue': _make_dac_bytes(0x1A, 0), 'cyan': _make_dac_bytes(0x18, 1), 'teal': _make_dac_bytes(0x1A, 1), 'green_yellow': _make_dac_bytes(0x18, 2), 'red': _make_dac_bytes(0x18, 3) } LAMP_SPECS = { 'uv': (396, 16), 'blue': (434, 22), 'cyan': (481, 22), 'teal': (508, 29), 'green_yellow': (545, 70), 'red': (633, 19) } LAMP_NAMES = set(LAMP_DAC_COMMANDS.keys()) class Lamp(state_stack.StateStackDevice): def __init__(self, name, spectra): super().__init__() self._name = name self._spectra = spectra def set_intensity(self, value): """Set lamp intensity in the range [0, 255]""" self._spectra._lamp_intensity(self._name, value) def get_intensity(self): return self._spectra._lamp_intensities[self._name] def set_enabled(self, enabled): self._spectra._lamp_enable(self._name, enabled) def get_enabled(self): return self._spectra._lamp_enableds[self._name] class SpectraX(property_device.PropertyDevice): _DESCRIPTION = 'Lumencor Spectra X' _EXPECTED_INIT_ERRORS = (smart_serial.SerialException,) def __init__(self, iotool: iotool.IOTool, property_server=None, property_prefix=''): super().__init__(property_server, property_prefix) self._spconfig = scope_configuration.get_config().spectra self._serial_port = smart_serial.Serial(self._spconfig.SERIAL_PORT, timeout=1, **self._spconfig.SERIAL_ARGS) # RS232 Lumencor docs state: "The [following] two commands MUST be issued after every power cycle to properly configure controls for further commands." # "Set GPIO0-3 as open drain output" self._serial_port.write(b'\x57\x02\xFF\x50') # "Set GPI05-7 push-pull out, GPIO4 open drain out" self._serial_port.write(b'\x57\x03\xAB\x50') # test if we can connect: try: self.get_temperature() except smart_serial.SerialTimeout: # explicitly clobber traceback from SerialTimeout exception raise smart_serial.SerialException('Could not read data from Spectra -- is it turned on?') self._iotool = iotool if property_server: self._update_property('temperature', self.get_temperature()) self._sleep_time = 10 self._timer_running = True self._timer_thread = threading.Thread(target=self._timer_update_temp, daemon=True) self._timer_thread.start() self._available_lamps = set(self._spconfig.IOTOOL_LAMP_PINS.keys()) for name in self._available_lamps: setattr(self, name, Lamp(name, self)) self._lamp_intensities = {} self._lamp_enableds = {} self.lamps(**{lamp+'_enabled': False for lamp in self._available_lamps}) self.lamps(**{lamp+'_intensity': 255 for lamp in self._available_lamps}) def _timer_update_temp(self): while self._timer_running: self._update_property('temperature', self.get_temperature()) time.sleep(self._sleep_time) def _lamp_intensity(self, lamp, value): assert 0 <= value <= 255 inverted = 255 - value # Put the intensity value, which is from 0xFF (off) to 0 (on), into the middle 8 bits of a 16-bit integer, # with the high 4 bits as 0xF and the low 4 as 0. Bizarre, but that's the wire protocol. intensity_bytes = 0xF000 | (inverted << 4) dac_bytes = LAMP_DAC_COMMANDS[lamp] dac_bytes[4] = intensity_bytes >> 8 dac_bytes[5] = intensity_bytes & 0x00FF self._serial_port.write(bytes(dac_bytes)) self._lamp_intensities[lamp] = value self._update_property(lamp+'.intensity', value) def _lamp_enable(self, lamp, enabled): self._iotool.execute(*self._iotool_lamp_commands(**{lamp: enabled})) self._lamp_enableds[lamp] = enabled self._update_property(lamp+'.enabled', enabled) def _iotool_lamp_commands(self, **lamps): """Produce a sequence of IOTool commands to enable and disable given Spectra lamps. Keyword arguments must be lamp names, as specified in the scope configuration. The values specified must be True to enable that lamp, False to disable, or None to do nothing (unspecified lamps are also not altered).""" commands = [] for lamp, enabled in lamps.items(): if enabled is None: continue pin = self._spconfig.IOTOOL_LAMP_PINS[lamp] if enabled: commands.append(self._iotool.commands.set_high(pin)) else: commands.append(self._iotool.commands.set_low(pin)) return commands def get_lamp_specs(self): """Return a dict mapping lamp names to tuples of (peak_wavelength, bandwidth), in nm, where bandwidth is the minimum width required to contain 75% of the spectral intensity of the lamp output.""" return {lamp: LAMP_SPECS[lamp] for lamp in self._available_lamps} def get_temperature(self): self._serial_port.write(b'\x53\x91\x02\x50') r = self._serial_port.read(2) return ((r[0] << 3) | (r[1] >> 5)) * 0.125 def lamps(self, **lamp_parameters): """Set a number of lamp parameters at once using keyword arguments, e.g. spectra.lamps(red_enabled=True, red_intensity=255, blue_enabled=False) Intensity values must be in the range [0, 255]. Valid lamp names can be retrieved with get_lamp_specs(). """ self._set_state(lamp_parameters.items()) def _get_getter_setter(self, prop): """Return a property setter/getter pair, either from a "real" property get/set pair, or a "virtual" property like "red_enabled" or "cyan_intensity".""" if hasattr(self, 'set_'+prop): return getattr(self, 'get_'+prop), getattr(self, 'set_'+prop) else: lamp_name, lamp_prop = prop.rsplit('_', 1) if lamp_name not in self._available_lamps: raise ValueError('Invalid lamp name') lamp = getattr(self, lamp_name) if hasattr(lamp, 'set_'+ lamp_prop): return getattr(lamp, 'get_'+lamp_prop), getattr(lamp, 'set_'+lamp_prop) else: raise ValueError('Invalid lamp parameter "{}"'.format(lamp_prop)) def _set_state(self, properties_and_values): for prop, value in properties_and_values: setter = self._get_getter_setter(prop)[1] setter(value) def push_state(self, **lamp_parameters): """Set a number of parameters at once using keyword arguments, while saving the old values of those parameters. (See lamps() for a description of valid parameters.) pop_state() will restore those previous values. push_state/pop_state pairs can be nested arbitrarily.""" # Also note that we do not filter out identical states from being pushed. # Since the enabled state can be fiddled with IOTool, there is good reason # for pushing an enabled state identical to the current one, so that it # will be restored after any such fiddling. old_state = {} for prop, value in lamp_parameters.items(): getter, setter = self._get_getter_setter(prop) old_state[prop] = getter() setter(value) self._state_stack.append(old_state) class Spectra(SpectraX): _DESCRIPTION = 'Lumencor Spectra' def __init__(self, iotool: iotool.IOTool, property_server=None, property_prefix=''): super().__init__(iotool, property_server, property_prefix) self.set_green_yellow_filter('green') def _iotool_enable_green_command(self): """Produce a command that switches the green/yellow paddle to the green filter position.""" return self._iotool.commands.set_high(self._spconfig.IOTOOL_GREEN_YELLOW_SWITCH_PIN) def _iotool_enable_yellow_command(self): """Produce a command that switches the green/yellow paddle to the yellow filter position.""" return self._iotool.commands.set_low(self._spconfig.IOTOOL_GREEN_YELLOW_SWITCH_PIN) def set_green_yellow_filter(self, position): """'position' should be either 'green' or 'yellow' to insert the corresponding excitation filter into the green/yellow beam.""" if position not in {'green', 'yellow'}: raise ValueError('"position" parameter must be either "green" or "yellow"') if position == 'green': self._iotool.execute(self._iotool_enable_green_command()) else: self._iotool.execute(self._iotool_enable_yellow_command()) time.sleep(self._spconfig.FILTER_SWITCH_DELAY) self._green_yellow_pos = position self._update_property('green_yellow_filter', position) def get_green_yellow_filter(self): return self._green_yellow_pos
9,444
3,066
"""Performance testing utility.""" from ._format_perf_counter import format_perf_counter from ._multi_case_timer import MultiCaseTimer from ._util import plot_run, to_dataframe from ._wrappers import run_multivariate_test __all__ = [ "MultiCaseTimer", "run_multivariate_test", "format_perf_counter", "plot_run", "to_dataframe", ]
352
121
import numpy as np import torch from sklearn.metrics import roc_auc_score from sklearn.metrics import average_precision_score def test(model, features, adj, sparse, adj_sparse, test_edges, test_edges_false): eps = 1e-4 embeds, _,_, S= model.embed(features, adj, sparse, None, 100) embs = embeds[0, :] embs = embs / (embs.norm(dim=1)[:, None] + eps) sc_roc, sc_ap = get_roc_score(test_edges, test_edges_false, embs.cpu().detach().numpy(), adj_sparse) return sc_roc, sc_ap def get_roc_score(edges_pos, edges_neg, embeddings, adj_sparse): "from https://github.com/tkipf/gae" score_matrix = np.dot(embeddings, embeddings.T) def sigmoid(x): return 1 / (1 + np.exp(-x)) # Store positive edge predictions, actual values preds_pos = [] pos = [] for edge in edges_pos: preds_pos.append(sigmoid(score_matrix[edge[0], edge[1]])) # predicted score pos.append(adj_sparse[edge[0], edge[1]]) # actual value (1 for positive) # Store negative edge predictions, actual values preds_neg = [] neg = [] for edge in edges_neg: preds_neg.append(sigmoid(score_matrix[edge[0], edge[1]])) # predicted score neg.append(adj_sparse[edge[0], edge[1]]) # actual value (0 for negative) # Calculate scores preds_all = np.hstack([preds_pos, preds_neg]) labels_all = np.hstack([np.ones(len(preds_pos)), np.zeros(len(preds_neg))]) #print(preds_all, labels_all ) roc_score = roc_auc_score(labels_all, preds_all) ap_score = average_precision_score(labels_all, preds_all) return roc_score, ap_score def torch_normalize_adj(adj): adj = adj + torch.eye(adj.shape[0]).cuda() rowsum = adj.sum(1) d_inv_sqrt = torch.pow(rowsum, -0.5).flatten() d_inv_sqrt[torch.isinf(d_inv_sqrt)] = 0.0 d_mat_inv_sqrt = torch.diag(d_inv_sqrt).cuda() return adj.mm(d_mat_inv_sqrt).t().mm(d_mat_inv_sqrt)
1,951
755
import email import logging import re from imapclient import IMAPClient class IMAPError(IOError): pass class ImapWrapper: """A wrapper around imaplib, since that's a bit lower-level than I'd prefer to work with.""" #This regex is: # list of flags in parens # quoted delimiter # possible-quoted folder name list_matcher = re.compile(r'^\(([^()]*)\) "([^"]*)" (([^" ]+)|"([^"]*)")$') def __init__(self, host, user, pw, **kwargs): """kwargs: Paassed through to IMAPClient""" self.M = IMAPClient(host, **kwargs) self.M.login(user, pw) self._selected_folder = None self._update_folders() def logout(self): self.M.logout() def _update_folders(self): listing = self.M.list_folders() self.folder_list = [name for (flags, delim, name) in listing] def ensure_folder(self, name): """Return True if the folder was created, False if it already existed.""" l = logging.getLogger(__name__) search_name = name[:-1] if name.endswith('/') else name if not any(n == search_name for n in self.folder_list): rslt = self.M.create_folder(name) l.info(f"Folder create result: {rslt}") self.folder_list.append(search_name) return True else: return False def fetch_messages(self, folder, *search_args): l = logging.getLogger(__name__) ret = [] self.select_folder(folder) message_ids = self.M.search(search_args) message_dict = self.M.fetch(message_ids, 'RFC822') for msg in message_dict.values(): l.debug("Got message: %s", msg) msg = email.message_from_string(msg[b'RFC822'].decode('UTF-8')) ret.append(msg) return ret def check_folder_for_message_ids(self, folder, msgids): self.select_folder(folder) search_ids = [] for msgid in msgids: if len(search_ids) > 0: search_ids.insert(0, 'OR') search_ids.append(['HEADER', 'Message-Id', msgid]) message_numbers = self.M.search(['NOT', 'DELETED', search_ids]) message_envelopes = self.M.fetch(message_numbers, 'ENVELOPE') have_ids = [] for msgdata in message_envelopes.values(): envelope = msgdata[b'ENVELOPE'] have_ids.append(envelope.message_id) return have_ids def append(self, folder_name, email): response = self.M.append(folder_name, str(email).encode('utf-8')) logging.getLogger(__name__).debug("Append response: %s", response) # FIXME sets the context folder def select_folder(self, name): if self._selected_folder == name: return dtl = self.M.select_folder(name) logging.getLogger(__name__).debug("select_folder = %s", dtl) self._selected_folder = name def create_subscribe_folder(self, name): created = self.ensure_folder(name) if created: res = self.M.subscribe_folder(name) logging.getLogger(__name__).debug("Subscribe result: %s", res)
3,143
980
VERSION = '2.11.2'
19
14
import logging import os import pandas as pd from .example_filetype_format import FileTypeFormat from . import process_functions logger = logging.getLogger(__name__) class seg(FileTypeFormat): _fileType = "seg" _process_kwargs = ["newPath", "databaseSynId"] def _validateFilename(self, filePath): assert os.path.basename(filePath[0]) == "genie_data_cna_hg19_%s.%s" % (self.center, self._fileType) def _process(self, seg): seg.columns = [col.upper() for col in seg.columns] newsamples = [process_functions.checkGenieId(i, self.center) for i in seg['ID']] seg['ID'] = newsamples seg = seg.drop_duplicates() seg = seg.rename(columns= {'LOC.START':'LOCSTART','LOC.END':'LOCEND','SEG.MEAN':'SEGMEAN','NUM.MARK':'NUMMARK'}) seg['CHROM'] = [str(chrom).replace("chr","") for chrom in seg['CHROM']] seg['CENTER'] = self.center seg['LOCSTART'] = seg['LOCSTART'].astype(int) seg['LOCEND'] = seg['LOCEND'].astype(int) seg['NUMMARK'] = seg['NUMMARK'].astype(int) return(seg) def process_steps(self, seg, newPath, databaseSynId): seg = self._process(seg) process_functions.updateData(self.syn, databaseSynId, seg, self.center, toDelete=True) seg.to_csv(newPath,sep="\t",index=False) return(newPath) def _validate(self, segDF): total_error = "" warning = "" segDF.columns = [col.upper() for col in segDF.columns] REQUIRED_HEADERS = pd.Series(['ID','CHROM','LOC.START','LOC.END','NUM.MARK','SEG.MEAN']) if not all(REQUIRED_HEADERS.isin(segDF.columns)): total_error += "Your seg file is missing these headers: %s.\n" % ", ".join(REQUIRED_HEADERS[~REQUIRED_HEADERS.isin(segDF.columns)]) else: intCols = ['LOC.START','LOC.END','NUM.MARK'] nonInts = [col for col in intCols if segDF[col].dtype != int] if len(nonInts) > 0: total_error += "Seg: Only integars allowed in these column(s): %s.\n" % ", ".join(sorted(nonInts)) if not segDF['SEG.MEAN'].dtype in [float, int]: total_error += "Seg: Only numerical values allowed in SEG.MEAN.\n" checkNA = segDF.isna().apply(sum) nullCols = [ind for ind in checkNA.index if checkNA[ind] > 0] if len(nullCols) > 0: total_error += "Seg: No null or empty values allowed in column(s): %s.\n" % ", ".join(sorted(nullCols)) return(total_error, warning)
2,527
860
import sys from . import app sys.path.append(str(app.config['LIB_PATH'])) from musicautobot.music_transformer import * from musicautobot.config import * from flask import Response, send_from_directory, send_file, request, jsonify from .save import to_s3 import torch import traceback torch.set_num_threads(4) data = load_data(app.config['DATA_PATH'], app.config['DATA_SAVE_NAME'], num_workers=1) learn = music_model_learner(data, pretrained_path=app.config['MUSIC_MODEL_PATH']) if torch.cuda.is_available(): learn.model.cuda() # learn.to_fp16(loss_scale=512) # fp16 not supported for cpu - https://github.com/pytorch/pytorch/issues/17699 @app.route('/predict/midi', methods=['POST']) def predict_midi(): args = request.form.to_dict() midi = request.files['midi'].read() print('THE ARGS PASSED:', args) bpm = float(args['bpm']) # (AS) TODO: get bpm from midi file instead temperatures = (float(args.get('noteTemp', 1.2)), float(args.get('durationTemp', 0.8))) n_words = int(args.get('nSteps', 200)) seed_len = int(args.get('seedLen', 12)) # debugging 1 - send exact midi back # with open('/tmp/test.mid', 'wb') as f: # f.write(midi) # return send_from_directory('/tmp', 'test.mid', mimetype='audio/midi') # debugging 2 - test music21 conversion # stream = file2stream(midi) # 1. # debugging 3 - test npenc conversion # seed_np = midi2npenc(midi) # music21 can handle bytes directly # stream = npenc2stream(seed_np, bpm=bpm) # debugging 4 - midi in, convert, midi out # stream = file2stream(midi) # 1. # midi_in = Path(stream.write("musicxml")) # print('Midi in:', midi_in) # stream_sep = separate_melody_chord(stream) # midi_out = Path(stream_sep.write("midi")) # print('Midi out:', midi_out) # s3_id = to_s3(midi_out, args) # result = { # 'result': s3_id # } # return jsonify(result) # Main logic try: full = predict_from_midi(learn, midi=midi, n_words=n_words, seed_len=seed_len, temperatures=temperatures) stream = separate_melody_chord(full.to_stream(bpm=bpm)) midi_out = Path(stream.write("midi")) print('Wrote to temporary file:', midi_out) except Exception as e: traceback.print_exc() return jsonify({'error': f'Failed to predict: {e}'}) s3_id = to_s3(midi_out, args) result = { 'result': s3_id } return jsonify(result) # return send_from_directory(midi_out.parent, midi_out.name, mimetype='audio/midi') # @app.route('/midi/song/<path:sid>') # def get_song_midi(sid): # return send_from_directory(file_path/data_dir, htlist[sid]['midi'], mimetype='audio/midi') @app.route('/midi/convert', methods=['POST']) def convert_midi(): args = request.form.to_dict() if 'midi' in request.files: midi = request.files['midi'].read() elif 'midi_path'in args: midi = args['midi_path'] stream = file2stream(midi) # 1. # stream = file2stream(midi).chordify() # 1. stream_out = Path(stream.write('musicxml')) return send_from_directory(stream_out.parent, stream_out.name, mimetype='xml')
3,150
1,182
from __future__ import absolute_import, division, print_function from tests.core import mock from trakt import Trakt from httmock import HTTMock import pytest def test_likes(): with HTTMock(mock.fixtures, mock.unknown): with Trakt.configuration.auth('mock', 'mock'): likes = Trakt['users'].likes() assert likes is not None likes = list(likes) assert len(likes) == 3 assert likes[0].keys == [ ('trakt', 1519) ] assert likes[1].keys == [ ('trakt', '1238362'), ('slug', 'star-wars-machete') ] assert likes[2].keys == [ ('trakt', '840781'), ('slug', 'star-wars-timeline') ] def test_likes_invalid_response(): with HTTMock(mock.fixtures, mock.unknown): likes = Trakt['users'].likes() assert likes is None def test_likes_invalid_type(): with HTTMock(mock.fixtures, mock.unknown): with pytest.raises(ValueError): likes = Trakt['users'].likes('invalid') assert likes is not None likes = list(likes)
1,086
372
import sys import PyQt5.QtWidgets as qtw import PyQt5.QtCore as qtc from Image import Image from main_layout import Ui_MainWindow import logging import os logger = logging.getLogger(__name__) logger.setLevel(logging.INFO) formatter = logging.Formatter('%(levelname)s:%(name)s:%(asctime)s - %(message)s') file_handler = logging.FileHandler('log') file_handler.setLevel(logging.INFO) file_handler.setFormatter(formatter) logger.addHandler(file_handler) class MainWindow(qtw.QMainWindow): def __init__(self): super().__init__() self.ui = Ui_MainWindow() self.ui.setupUi(self) self.show() self.images = { '1': { 'original': self.ui.image_1_original, 'filtered': self.ui.image_1_after_filter, 'picker': self.ui.image_1_pick }, '2': { 'original': self.ui.image_2_original, 'filtered': self.ui.image_2_after_filter, 'picker': self.ui.image_2_pick } } self.img = {} self.modes = {'Output 1': '', 'Output 2': ''} self.output_channels = { 'Output 1': self.ui.output_1, 'Output 2': self.ui.output_2 } self.output_channels_controlers = { '': { 'select1': '', 'select2': '', 'slider1': 0, 'slider2': 0, 'type1': '', 'type2': '', 'percentage1': 0, 'percentage2': 0, }, 'Output 1': { 'select1': '', 'select2': '', 'slider1': 0, 'slider2': 0, 'type1': '', 'type2': '', 'percentage1': 0, 'percentage2': 0, }, 'Output 2': { 'select1': '', 'select2': '', 'slider1': 0, 'slider2': 0, 'type1': '', 'type2': '', 'percentage1': 0, 'percentage2': 0, }, } self.output_complementary = { '': ['', 'Magnitude', 'Phase', 'Real', 'Imaginary', 'Uniform Magnitude', 'Uniform Phase'], 'Magnitude': ['Phase', 'Uniform Phase'], 'Phase': ['Magnitude', 'Uniform Magnitude'], 'Real': ['Imaginary'], 'Imaginary': ['Real'], 'Uniform Magnitude': ['Phase', 'Uniform Phase'], 'Uniform Phase': ['Magnitude', 'Uniform Magnitude'], } self.available_images = { '': '' } self.enables = { '': [self.ui.component_1_select, self.ui.component_2_select, self.ui.component_1_percentage, self.ui.component_1_slider, self.ui.component_1_type, self.ui.component_2_percentage, self.ui.component_2_slider, self.ui.component_2_type], 'output-select': [self.ui.component_1_select, self.ui.component_2_select], 'select1': [self.ui.component_1_percentage, self.ui.component_1_type], 'select2': [self.ui.component_2_percentage, self.ui.component_2_type], 'type1': [self.ui.component_1_slider], 'type2': [self.ui.component_2_slider] } self.current_output_channel = None self.ui.action_new.triggered.connect(self.new_instance) self.ui.action_exit.triggered.connect(self.close) self.ui.action_open_image_1.triggered.connect(lambda: self.open_image(self.images['1'], 1)) self.ui.action_open_image_2.triggered.connect(lambda: self.open_image(self.images['2'], 2)) self.ui.image_1_pick.currentIndexChanged.connect(lambda: self.display_component(self.img['Image 1'])) self.ui.image_2_pick.currentIndexChanged.connect(lambda: self.display_component(self.img['Image 2'])) self.ui.output_select.currentIndexChanged.connect(lambda: self.pick_mixer_output()) self.ui.component_1_select.currentIndexChanged.connect(lambda: self.select_enable('select1', self.ui.component_1_select.currentText())) self.ui.component_2_select.currentIndexChanged.connect(lambda: self.select_enable('select2', self.ui.component_2_select.currentText())) self.ui.component_1_slider.sliderReleased.connect(lambda: self.mixer('slider1', str(self.ui.component_1_slider.value()))) self.ui.component_2_slider.sliderReleased.connect(lambda: self.mixer('slider2', str(self.ui.component_2_slider.value()))) self.ui.component_1_percentage.valueChanged.connect(lambda: self.change_image('percentage1', str(self.ui.component_1_percentage.value()))) self.ui.component_2_percentage.valueChanged.connect(lambda: self.change_image('percentage2', str(self.ui.component_2_percentage.value()))) self.ui.component_1_type.currentIndexChanged.connect(lambda: self.component_1_conplementary()) self.ui.component_1_type.currentIndexChanged.connect(lambda: self.select_enable('type1', str(self.ui.component_1_type.currentText()))) self.ui.component_2_type.currentIndexChanged.connect(lambda: self.select_enable('type2', str(self.ui.component_2_type.currentText()))) def new_instance(self) -> None: self.child_window = MainWindow() self.child_window.show() def open_image(self, imageWidget: dict, channel: int) -> None: image = Image() if not image.path: return if len(self.img) == 1: if f'Image {2//channel}' in self.img: if not image.compare(self.img[f'Image {2//channel}']['image']): qtw.QMessageBox.warning(self, 'failed', 'The Two Images Must be of the same size') return else : self.img[f'Image {channel}'] = {'image': image, 'widgets': imageWidget} if f'Image {channel}' not in self.available_images: self.available_images[f'Image {channel}'] = f'Image {channel}' self.append_outputs(isOneChanneled=False) else : self.img[f'Image {channel}'] = {'image': image, 'widgets': imageWidget} elif len(self.img) >= 2: if not image.compare(self.img[f'Image {2//channel}']['image']): qtw.QMessageBox.warning(self, 'failed', 'The Two Images Must be of the same size') return self.img[f'Image {channel}']["image"] = image self.img[f'Image {channel}']["widgets"] = imageWidget else : self.img[f'Image {channel}'] = {'image': image, 'widgets': imageWidget} if f'Image {channel}' not in self.available_images: self.available_images[f'Image {channel}'] = f'Image {channel}' self.append_outputs(channel=self.available_images[f'Image {channel}']) imageWidget['original'].setPixmap(image.get_pixmap().scaled(300,300, aspectRatioMode=qtc.Qt.KeepAspectRatio, transformMode=qtc.Qt.SmoothTransformation)) imageWidget['picker'].setDisabled(False) self.ui.output_select.setDisabled(False) def append_outputs(self, isOneChanneled: bool=True, channel: str='') -> None: if isOneChanneled: self.ui.component_1_select.addItem('') self.ui.component_2_select.addItem('') self.ui.component_1_select.setItemText(0, '') self.ui.component_1_select.setItemText(1, channel) self.ui.component_2_select.setItemText(0, '') self.ui.component_2_select.setItemText(1, channel) else: self.ui.component_1_select.addItem('') self.ui.component_2_select.addItem('') self.ui.component_1_select.setItemText(0, '') self.ui.component_1_select.setItemText(1, 'Image 1') self.ui.component_1_select.setItemText(2, 'Image 2') self.ui.component_2_select.setItemText(0, '') self.ui.component_2_select.setItemText(1, 'Image 1') self.ui.component_2_select.setItemText(2, 'Image 2') def display_component(self, imageWidget: dict) -> None: component = imageWidget['widgets']['picker'].currentText() imageWidget['widgets']['filtered'].setPixmap(imageWidget['image'].get_component_pixmap(component).scaled(300,300, aspectRatioMode=qtc.Qt.KeepAspectRatio, transformMode=qtc.Qt.SmoothTransformation)) try: os.remove('test.png') except: pass def pick_mixer_output(self) -> None: self.current_output_channel = self.ui.output_select.currentText() self.ui.component_1_slider.setValue(int(self.output_channels_controlers[self.ui.output_select.currentText()]['slider1'])) self.ui.component_1_percentage.setValue(int(self.output_channels_controlers[self.ui.output_select.currentText()]['percentage1'])) self.ui.component_1_select.setCurrentText(self.output_channels_controlers[self.ui.output_select.currentText()]['select1']) self.ui.component_1_type.setCurrentText(self.output_channels_controlers[self.ui.output_select.currentText()]['type1']) self.ui.component_2_slider.setValue(int(self.output_channels_controlers[self.ui.output_select.currentText()]['slider2'])) self.ui.component_2_percentage.setValue(int(self.output_channels_controlers[self.ui.output_select.currentText()]['percentage2'])) self.ui.component_2_select.setCurrentText(self.output_channels_controlers[self.ui.output_select.currentText()]['select2']) self.ui.component_2_type.setCurrentText(self.output_channels_controlers[self.ui.output_select.currentText()]['type2']) if self.ui.output_select.currentText() != '': self.set_mixer_components_disabled(self.enables['output-select'] ,False) else: self.set_mixer_components_disabled(self.enables['output-select'], True) def set_mixer_components_disabled(self, components: list, logic: bool) -> None: for component in components: component.setDisabled(logic) def select_enable(self, component: str, value: str): self.change_image(component, value) if value != '': self.set_mixer_components_disabled(self.enables[component], False) else: self.set_mixer_components_disabled(self.enables[component], True) def change_image(self, component: str, value: str) -> None: self.output_channels_controlers[self.current_output_channel][component] = value def component_1_conplementary(self): self.ui.component_2_type.clear() self.ui.component_2_type.addItems(self.output_complementary[self.ui.component_1_type.currentText()]) self.ui.component_2_type.update() self.change_image('type1', self.ui.component_1_type.currentText()) def mixer(self, slider: str, sliderValue: str) -> None: self.change_image(slider, sliderValue) channel_1_ratio = float(self.output_channels_controlers[self.current_output_channel]['slider1']) / 100 channel_2_ratio = float(self.output_channels_controlers[self.current_output_channel]['slider2']) / 100 image_1 = self.output_channels_controlers[self.current_output_channel]['select1'] image_2 = self.output_channels_controlers[self.current_output_channel]['select2'] type1 = self.output_channels_controlers[self.current_output_channel]['type1'] type2 = self.output_channels_controlers[self.current_output_channel]['type2'] if image_1 == "" or image_2 == "" or type1 == "" or type2 == "": return try: if (type1 in ['Magnitude', 'Phase', 'Uniform Magnitude', 'Uniform Phase'] and type2 in ['Magnitude', 'Phase', 'Uniform Magnitude', 'Uniform Phase']): self.modes[self.current_output_channel] = 'mag-phase' elif (type1 in ['Real', 'Imaginary']and type2 in ['Real', 'Imaginary']): self.modes[self.current_output_channel] = 'real-imag' else: print('Error') return self.outImage = self.img[image_1]['image'].mix(self.img[image_2]['image'], self.output_channels_controlers[self.current_output_channel]['type1'], self.output_channels_controlers[self.current_output_channel]['type2'], channel_1_ratio, channel_2_ratio, self.modes[self.current_output_channel]) self.output_channels[self.current_output_channel].setPixmap(self.outImage.scaled(300,300, aspectRatioMode=qtc.Qt.KeepAspectRatio, transformMode=qtc.Qt.SmoothTransformation)) except: pass try: os.remove('test.png') except: pass def main_window(): app = qtw.QApplication(sys.argv) app.setStyle("Fusion") window = MainWindow() sys.exit(app.exec_()) if __name__ == '__main__': main_window()
12,983
4,034
from django import forms # Form for create sales class FormSales(forms.Form): # Atributes for form # NOTE: date_creation and total fields, it will be created dynamically name = forms.CharField(label="Nombre", max_length=50, required=True, help_text="Ingrese su nombre") description = forms.CharField(label="Descripción", max_length=120, help_text="Ingrese una descripción de su producto", widget=forms.Textarea()) count = forms.IntegerField(label="Cantidad",min_value=0,required=True) price = forms.DecimalField(label="Precio",min_value=0,required=True) # Defaul value is True paid_out = forms.BooleanField(label="Pagado", initial=True, help_text="Si es verdadero, está pagado", widget=forms.CheckboxInput())
1,029
264
# -*- coding: utf-8 -*- """Pipeline for VarGenPath""" from typing import Optional from .constants import LINKSET_PATH, FILE_TYPES from .utils import ( get_cytoscape_connection, get_associated_genes, var_genes_network, extend_vargen_network, save_session, save_image, save_network ) def get_vargenpath_network( *, variant_list: list, network_name: Optional[str] = 'VarGenPath network', linkset_path: Optional[str] = LINKSET_PATH, session_path: Optional[str] = None, image_path: Optional[str] = None, extend_network: bool = True, image_type: Optional[str] = 'svg', network_path: Optional[str] = None, network_file_path: Optional[str] = 'cyjs', ) -> dict: """ Pipeline for creating vargenpath network. :param network_file_path: the type of network file to be saved. :param network_path: if input path, the cytoscape network will be saved to this path. :param variant_list: list of variants. :param network_name: the name of the network. :param linkset_path: the path to the linkset to extend network. :param session_path: if input path, the cytoscape session will be saved to this path. :param image_path: if input path, the image of the network will be saved to this path. :param extend_network: if true, the network will be extended. :param image_type: the type of the image to be saved. :return: cytoscape network """ try: cy = get_cytoscape_connection() except Exception: raise Exception('Uh-oh! Make sure that cytoscape is open then try again.') vargen_df = get_associated_genes(variant_list) network = var_genes_network(variants_genes_df=vargen_df, client=cy, network_name=network_name) if extend_network: network = extend_vargen_network(linkset_path, client=cy) if session_path is not None: save_session(session_file=session_path, client=cy) if image_path is not None: save_image(network_image=image_path, image_type=FILE_TYPES[image_type]) if network_path is not None: save_network(network_path=network_path, file_type=FILE_TYPES[network_file_path]) return network
2,159
689
# coding=utf-8 # Copyright 2021 Google LLC. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A model for to embed structured features.""" from typing import Any, Tuple import flax.linen as nn import jax.numpy as jnp class FeaturesEncoder(nn.Module): """Encodes structured features.""" input_dims: Tuple[int] embed_dim: int = 32 @nn.compact def __call__(self, x): result = [] index = 0 for d in self.input_dims: arr = x[..., index:index+d] result.append(arr if d == 1 else nn.Dense(self.embed_dim)(arr)) index += d return jnp.concatenate(result, axis=-1) class AdultModel(nn.Module): """A model to predict if the income is above 50k (adult dataset).""" encoder_cls: Any hidden: Tuple[int] = (64, 64) @nn.compact def __call__(self, x, train: bool = True): x = self.encoder_cls()(x) for h in self.hidden: x = nn.Dense(h)(x) x = nn.relu(x) x = nn.Dense(1)(x) x = nn.sigmoid(x) return x[..., 0]
1,485
538
from models import shoe_box def createHBjsons() -> None: for i in range(4): # increment on shoebox dimensions _width = 4.0 + i*0.5 _height = 3.5 + i*0.1 _depth = 4.0 + i*0.5 # init shoe_box sb = shoe_box.Shoebox(width=_width , height=_height , depth=_depth ) # set grid size and offset sb.gridSize = 0.5 sb.gridOffset = 0.75 # set window to wall ratio sb.wwr = 0.4 # create room sb.createRoom() # create model sb.createModel() # save to hbjson sb.saveToHBJson() # run if __name__ == "__main__": # create models createHBjsons()
699
254
from reikna.core.signature import Type, Annotation, Parameter, Signature from reikna.core.computation import Computation from reikna.core.transformation import Transformation, Indices
184
50
from sqlalchemy import engine_from_config from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import sessionmaker from sqlalchemy.schema import MetaData import zope.sqlalchemy from .node import Node NAMING_CONVENTION = { "ix": 'ix_%(column_0_label)s', "uq": "uq_%(table_name)s_%(column_0_name)s", "ck": "ck_%(table_name)s_%(constraint_name)s", "fk": "fk_%(table_name)s_%(column_0_name)s_%(referred_table_name)s", "pk": "pk_%(table_name)s" } metadata = MetaData(naming_convention=NAMING_CONVENTION) Base = declarative_base(metadata=metadata) def get_session(transaction_manager, dbmaker): dbsession = dbmaker() zope.sqlalchemy.register(dbsession, transaction_manager=transaction_manager) return dbsession def get_engine(settings, prefix='sqlalchemy.'): return engine_from_config(settings, prefix) def get_dbmaker(engine): dbmaker = sessionmaker() dbmaker.configure(bind=engine) return dbmaker def includeme(config): settings = config.get_settings() dbmaker = get_dbmaker(get_engine(settings)) config.add_request_method( lambda r: get_session(r.tm, dbmaker), 'dbsession', reify=True )
1,234
436
from __future__ import absolute_import from __future__ import print_function from __future__ import division def Shape(visitor, node): input = visitor.visit(node.input[0]) shape = input.shape if (input.get_layout() is not None and input.get_onnx_layout() is not None and input.get_layout() != input.get_onnx_layout()): shape = [shape[input.get_layout().index(l)] for l in input.get_onnx_layout()] return tuple(shape)
458
149
import typing from typing import ( Any, Callable, Dict, Iterable, List, Literal, Optional, Set, Tuple, Union, ) from ncadquery import Workplane from OCP.Quantity import Quantity_NameOfColor from OCP.TCollection import TCollection_ExtendedString from OCP.TDataStd import TDataStd_Name from OCP.TDF import TDF_Label, TDF_TagSource from OCP.TNaming import TNaming_Builder, TNaming_NamedShape from OCP.TopoDS import TopoDS_Shape from OCP.TPrsStd import TPrsStd_AISPresentation from PyQt5.QtCore import QPersistentModelIndex, Qt from nales.nales_cq_impl import NALES_TYPES, CQMethodCall, Part from nales.utils import TypeChecker from nales.widgets.msg_boxs import StdErrorMsgBox class NNode: def __init__(self, name=None, parent=None): self._parent = parent self._columns_nb = 1 self._childs = [] if parent: self._row = len(parent._childs) parent._childs.append(self) parent._columns_nb = max(self.column, parent.column) self._label = TDF_TagSource.NewChild_s(parent._label) self._name = name TDataStd_Name.Set_s(self._label, TCollection_ExtendedString(name)) else: self._label = TDF_Label() self._name = "root" self._row = 0 def _create_sublabel(self): """ Create an additionnal OCCT label that is needed if you want to display several shapes (It's one shape per label) """ sublabel = TDF_TagSource.NewChild_s(self._label) TDataStd_Name.Set_s( sublabel, TCollection_ExtendedString(f"{self.name} subshape") ) return sublabel def walk(self, node: "NNode" = None) -> "NNode": """ Walks all the node starting from 'node' If 'node' is None, starts from the called node """ base_node = node if node else self yield base_node for child in base_node.childs: yield from self.walk(child) def find(self, node_name: str, node_type=None) -> "NNode" or None: for node in self.walk(): if node.name == node_name: if node_type: if isinstance(node, node_type): return node else: return node def data(self, column): if column >= 0 and column < len(self._data): return self._data[column] @property def column(self): return self._columns_nb def child_count(self): return len(self._childs) def child(self, row) -> "NNode": if row >= 0 and row < self.child_count(): return self._childs[row] def has_children(self): if len(self._childs) != 0: return True else: return False @property def parent(self): return self._parent @property def childs(self): return self._childs @childs.setter def childs(self, new_childs): self._childs = new_childs @property def name(self): return self._name @name.setter def name(self, value): self._name = value @property def root_node(self): root = self.parent while True: if root.parent: root = root.parent else: return root @property def row(self): return self._row class NPart(NNode): def __init__(self, name: str, parent): super().__init__(name, parent=parent) self.visible = True self._solid = TopoDS_Shape() self._active_shape = None self.display() @property def part(self): return self.childs[-1].part_obj def _update_display_shapes(self): try: solid = self.part._findSolid().wrapped except ValueError: solid = TopoDS_Shape() self._solid = solid if not (active_shape := self.part._val().wrapped) is solid and isinstance( active_shape, TopoDS_Shape ): self._active_shape = active_shape else: self._active_shape = None def hide(self): self.visible = False self.ais_solid.Erase(remove=True) self.ais_active_shape.Erase(remove=True) self.root_node._viewer.Update() def display(self, update=False): """ Builds the display object and attach it to the OCAF tree """ if update: self.ais_solid.Erase(remove=True) if self._active_shape: self.ais_active_shape.Erase(remove=True) self._update_display_shapes() # self.root_node._viewer.Update() solid_bldr = TNaming_Builder(self._label) # _label is TDF_Label solid_bldr.Generated(self._solid) solid_shape_attr = solid_bldr.NamedShape() self.ais_solid = TPrsStd_AISPresentation.Set_s(solid_shape_attr) if self._active_shape: active_shape_bldr = TNaming_Builder(self._create_sublabel()) active_shape_bldr.Generated(self._active_shape) active_shape_attr = active_shape_bldr.NamedShape() self.ais_active_shape = TPrsStd_AISPresentation.Set_s(active_shape_attr) self.ais_active_shape.Display(update=True) self.root_node._viewer.Update() # There is color mixing due to overlapping, maybe this can help to solve the issue : # https://dev.opencascade.org/doc/refman/html/class_a_i_s___interactive_context.html#a1e0f9550cc001adbb52329ac243bb3b2 # It's considered good enough for now self.ais_solid.SetTransparency(0.9) self.ais_solid.Display() self.root_node._viewer.Update() self.visible = True def update(self): """ When called this method rebuild the entire Part, by calling each child Operation """ child_ops = self.childs for pos, child_op in enumerate(child_ops): child_op.update(pos) def remove_operation(self, row: int): """ Remove an operation from the operation tree """ ops: List[NOperation] = self.childs ops.pop(row) ops[row - 1].update_from_node() class NShape(NNode): def __init__(self, name, cq_shape, parent: NNode): self._occt_shape = shape = cq_shape.wrapped self.shape = cq_shape self.visible = True super().__init__(name, parent=parent) self.bldr = TNaming_Builder(self._label) # _label is TDF_Label self.bldr.Generated(shape) named_shape = self.bldr.NamedShape() self._label.FindAttribute(TNaming_NamedShape.GetID_s(), named_shape) self.ais_shape = TPrsStd_AISPresentation.Set_s(named_shape) self.ais_shape.SetTransparency(0.5) self.ais_shape.SetColor(Quantity_NameOfColor.Quantity_NOC_ALICEBLUE) self.ais_shape.Display(update=True) def hide(self): self.visible = False self.ais_shape.Erase() self.root_node._viewer.Update() def display(self, update=False): """ Builds the display object and attach it to the OCAF tree """ if update: self.ais_shape.Erase(remove=True) self.root_node._viewer.Update() self.bldr = TNaming_Builder(self._label) # _label is TDF_Label self.bldr.Generated(self._occt_shape) named_shape = self.bldr.NamedShape() self._label.FindAttribute(TNaming_NamedShape.GetID_s(), named_shape) self.ais_shape = TPrsStd_AISPresentation.Set_s(named_shape) self.ais_shape.SetTransparency(0.5) self.ais_shape.SetColor(Quantity_NameOfColor.Quantity_NOC_ALICEBLUE) self.ais_shape.Display(update=True) self.root_node._viewer.Update() self.visible = True def update(self): """ Update the shape object """ self._occt_shape = self.shape.wrapped self.display(True) class NShapeOperation(NNode): def __init__(self, maker_method: Callable, shape_class, parent=None): super().__init__(maker_method.__name__, parent) self.maker_method = maker_method self.shape_class = shape_class def update(self) -> None: args = [child.value for child in self.childs] self.parent.shape = self.maker_method(self.shape_class, *args) self.parent.update() class NOperation(NNode): def __init__( self, method_name: str, part_obj: Part, parent: NNode, operation: CQMethodCall ): super().__init__(method_name, parent=parent) self.part_obj = part_obj self.operation = operation self.method = getattr(part_obj, method_name).__func__ if method_name == "Workplane": self._root_operation = True else: self._root_operation = False def update_from_node(self): """ Update the Part from this node It recomputes every operation from this node to the end """ ops: List[NOperation] = self.parent.childs[self.row :] for op in ops: op.update() self.parent.display(update=True) def _update_init_part(self): """ This method is called when the user try to update __init__ method arguments There is a special handling because it is a bit different from the regular methods """ args = [ child.value if not child.is_linked("obj") else child.linked_obj for child in self.childs ] try: self.method(self.part_obj, *args, internal_call=True) except Exception as exc: StdErrorMsgBox(repr(exc)) def update(self) -> bool: """ Update the CQ objects stack from param modification in the GUI view """ # Special handling of __init__ method if self.row == 0: self._update_init_part() return True previous_operations: List[NOperation] = self.parent.childs[: self.row] old_part_obj = previous_operations[-1].part_obj args = [ child.value if not child.is_linked("obj") else child.linked_obj for child in self.childs ] try: self.part_obj = self.method(old_part_obj, *args, internal_call=True) return True except ValueError as exc: # we update parent operations until pending wires have reset if exc.args[0] == "No pending wires present": tried_updates = [self] # recursively call parent ops and store all the failed updates to update them again afterwards while (tried_update := previous_operations[-1].update()) is False: tried_updates.append(tried_update) for tried_update in tried_updates: tried_update.update() else: StdErrorMsgBox(repr(exc)) return False except Exception as exc: StdErrorMsgBox(repr(exc)) return False def _restore_pending_wires(self): index = 2 previous_ops = self.parent.childs[: self._row] while len(self.parent.part.ctx.pendingWires) == 0: op = previous_ops[-index] op.update(len(previous_ops) - op._row) index += 1 class NShapeArgument(NNode): def __init__(self, name=None, parent=None): super().__init__(name, parent) class NArgument(NNode): """ The underlying data of an Argument is as follow : name : cq argument name value : value linked_param : the name of the parameter linked to this arg, None if not connected to any type: value type : a voir si je garde ca If the Argument is linked to a Parameter, the Parameter name is displayed """ def __init__(self, arg_name: str, value, arg_type, parent: NNode, kwarg=False): super().__init__(arg_name, parent=parent) self._name = arg_name self._type = arg_type self._value = value self._typechecker = TypeChecker(arg_type) self._kwarg = kwarg # Boolean indicating if the arg is a kwarg or not self._linked_param = None self._linked_nobj_idx: QPersistentModelIndex = None self._param_name_pidx = None self._param_value_pidx = None def link( self, by: Literal["param", "obj"], value: Union[Tuple, QPersistentModelIndex, Any], ): """ Link this parameter to an object in available in the data model """ if by == "param": raw_val = value[1] if not self.is_type_compatible(raw_val): raise TypeError("Couldn't link the param") self._linked_param = value[0] self._value = value[1] self._param_name_pidx = value[2] self._param_value_pidx = value[3] else: self._linked_nobj_idx = value def unlink_param(self): self._linked_param = None self._param_name_pidx = None self._param_value_pidx = None def is_kwarg(self): return self._kwarg def is_linked(self, by: str = None): if by == "obj": return True if self._linked_nobj_idx else False elif by == "param": return True if self._linked_param else False elif by is None: if self._linked_param or self._linked_nobj_idx: return True else: return False else: raise ValueError("Argument 'by' must be either 'obj' or 'param'") def is_optional_type(self) -> bool: """ Indicates if the NArgument is optional, i.e the function signature looks something like : method(nargument:Union[float,None] = None) or method(nargument:Optional[float] = None) """ if self.is_kwarg(): origin = typing.get_origin(self._type) if origin == Optional: return True if origin == Union: for allowed_type in typing.get_args(self._type): if allowed_type == type(None): return True return False else: return False else: return False def is_literal_type(self) -> bool: origin = typing.get_origin(self.type) if self.type == str or origin == Literal: return True if origin == Union: possible_types = typing.get_args(self.type) for possible_type in possible_types: if possible_type == str or possible_type == Literal: return True return False def is_type_compatible(self, value: str) -> bool: return self._typechecker.check(value) def _cast(self, value: Any): if type(value) == self._type: return value return self._typechecker.cast(value) @property def type(self): return self._type @type.setter def type(self, value): self._type = value @property def linked_param(self): if self.is_linked(): return self._linked_param else: raise ValueError("This argument is not linked to a param") @property def linked_node(self): if not self._linked_nobj_idx: raise ValueError("This argument isn't linked to any node") else: return self._linked_nobj_idx.data(Qt.EditRole) @property def linked_obj(self): if self.is_linked(by="obj"): if hasattr(self.linked_node, "part"): return self.linked_node.part elif hasattr(self.linked_node, "shape"): return self.linked_node.shape else: raise NotImplementedError( "This argument is linked to a object that is not supported yet" ) else: raise ValueError("This argument is not linked to an object") @property def columns_nb(self): return 1 @property def name(self): return self._name @name.setter def name(self, value): self._name = value @property def value(self): if self.is_optional_type() and self._value is None: return None if self.is_linked(by="param"): return self._cast(self._param_value_pidx.data()) elif self.is_linked(by="obj"): return self.linked_obj elif not isinstance(self._value, str): # Upon argument creation self._value is already of the right type return self._value else: # If self._value is a string, means the users modified the argument in the GUI return self._cast(self._value) @value.setter def value(self, value): self._value = value @property def linked_param(self): return self._linked_param
17,072
5,083
from flyingObject import FlyingObject from bullet import Bullet import random class Hero(FlyingObject): """英雄机""" index = 2 # 标志位 def __init__(self, screen, images): # self.screen = screen self.images = images # 英雄级图片数组,为Surface实例 # image = pygame.image.load(images[0]) image = images[0] x = screen.get_rect().centerx y = screen.get_rect().bottom super(Hero, self).__init__(screen, x, y, image) self.life = 3 # 生命值为3 self.doubleFire = 100 # 初始火力值为0 self.RED = 1 self.PURPLE = 2 self.Fire_MOD = self.RED self.Fire_speed_RED = 40 self.Fire_speed_PURPLE = 40 self.mod = 5 self.difficulty = 0 def addDifficulty(self): self.difficulty+=1 def shoot_speed(self,x): if x==self.RED: return self.Fire_speed_RED else: return self.Fire_speed_PURPLE def getFire_MOD(self): """返回武器类型,RED为1 PURPLE为2""" return self.Fire_MOD def setFire_MOD(self, x): """设置武器类型""" if x == self.RED: if self.getFire_MOD() == x: self.addFire() if self.shoot_speed(x) > 30: self.Fire_speed_RED -= 10 else: self.clearFire() self.Fire_speed_PURPLE = 20 self.Fire_MOD = self.RED if x == self.PURPLE: if self.getFire_MOD() == x: self.addFire() if self.shoot_speed(x) > 20: self.Fire_speed_PURPLE -= 10 else: self.clearFire() self.Fire_speed_RED = 20 self.Fire_MOD = self.PURPLE def getFire(self): """获取火力值""" return self.doubleFire def setFire(self): """设置火力值""" self.doubleFire = 400 def addFire(self): """增加火力值""" self.doubleFire += 100 def clearFire(self): """清空火力值""" self.doubleFire = 100 def addLife(self): """增命""" self.life += 1 def sublife(self): """减命""" self.life -= 1 def getLife(self): """获取生命值""" return self.life def reLife(self): """重置值""" self.life = 3 self.clearFire() def outOfBounds(self): return False def step(self): """动态显示飞机""" if (len(self.images) > 0): # fclock = pygame.time.Clock() # fps = 10 # 帧数 # fclock.tick(fps) Hero.index += 0.3 Hero.index %= len(self.images) # self.image = pygame.image.load(self.images[int(Hero.index)]) # 切换图片 self.image = self.images[int(Hero.index)] # 切换图片 def move(self, x, y): self.x = x - self.width / 2 self.y = y - self.height / 2 def shoot(self, image): """英雄机射击""" xStep = int(self.width / 4 - 5) yStep = 20 if self.mod > 0: self.mod -= 1 else: self.mod = 5 if self.getFire_MOD() == 1: #print(self.getFire()) if self.doubleFire >= 400: heroBullet = [Bullet(self.screen, image, (self.x - (self.mod - 1) * xStep), self.y - yStep), Bullet(self.screen, image, (self.x + 2 * xStep), self.y - yStep), Bullet(self.screen, image, (self.x + (self.mod + 3) * xStep), self.y - yStep)] return heroBullet elif self.doubleFire >= 200: heroBullet = [Bullet(self.screen, image, (self.x - (5-self.mod - 1) * xStep), self.y - yStep), Bullet(self.screen, image, (self.x + (5-self.mod + 3) * xStep), self.y - yStep)] return heroBullet else: heroBullet = [Bullet(self.screen, image, self.x + 2 * xStep, self.y - yStep)] return heroBullet if self.getFire_MOD() == 2: if self.doubleFire >= 400: heroBullet = [Bullet(self.screen, image, self.x + 1 * xStep, self.y - yStep), Bullet(self.screen, image, self.x + 2 * xStep, self.y - yStep), Bullet(self.screen, image, self.x + 3 * xStep, self.y - yStep)] return heroBullet elif self.doubleFire >=200: heroBullet = [Bullet(self.screen, image, self.x + 1 * xStep, self.y - yStep), Bullet(self.screen, image, self.x + 3 * xStep, self.y - yStep)] return heroBullet else: heroBullet = [Bullet(self.screen, image, self.x + 2 * xStep, self.y - yStep)] return heroBullet def hit(self, other): """英雄机和其他飞机""" x1 = other.x - self.width / 2 x2 = other.x + self.width / 2 + other.width y1 = other.y - self.height / 2 y2 = other.y + self.height / 2 + other.height x = self.x + self.width / 2 y = self.y + self.height return x > x1 and x < x2 and y > y1 and y < y2
5,134
1,832
import sys import pygame as pg import numpy as np import random import time pic = np.zeros(shape=(128,64)) width = 128 height = 64 refresh_rate = 60 interval = 1 / refresh_rate bootrom_file = "bootrom0" rom_file = "rom" # rom_file = "hello_world" debug = False pg.display.init() display = pg.display.set_mode((width*4, height*4), flags=0, depth=8) screen = pg.Surface((width, height), flags=0, depth=8) pg.transform.scale(screen, (width*4, height*4), display) def screen_update(silent=True): pg.transform.scale(screen, (width*4, height*4), display) pg.display.flip() if not silent: print("Screen Update") def screen_clear(): screen.fill((0,0,0)) #screen_update() def screen_draw_line(x, y, pixels): # print("----------DRAW----------") # print("x:",x) # print("y:",y) # print("pix:",bin(pixels)) j = 0b10000000 for i in range(8): x_pos = x + i y_pos = y if x_pos >= 0 and x_pos < width: if y_pos >= 0 and y_pos < height: if pixels & j: pg.draw.rect(screen, 255, pg.Rect(x_pos,y_pos,1,1)) else: pg.draw.rect(screen, 0, pg.Rect(x_pos,y_pos,1,1)) j = j >> 1 #screen_update() screen_clear() # screen_draw_line(0,0,0b10101011) # input() class memByte: def __init__(self): self.value = 0x00000000 def write(self, value): self.value = value & 0xff def readUpper(self): return (self.value & 0b11110000) >> 4 def readLower(self): return self.value & 0b1111 class Flags: def __init__(self): self.z = 0 self.n = 0 self.h = 0 self.c = 0 def setZero(self): self.z = 1 def clearZero(self): self.z = 0 def setNeg(self): self.n = 1 def clearNeg(self): self.n = 0 def setHalf(self): self.h = 1 def clearHalf(self): self.h = 0 def setCarry(self): self.c = 1 def clearCarry(self): self.c = 0 def clearFlags(self): self.z = 0 self.n = 0 self.h = 0 self.c = 0 class reg: def __init__(self): self.value = 0b00000000 self.value = random.randint(0,255) def send(self): sys.stdout.write(chr(self.value)) sys.stdout.flush() class Dreg: def __init__(self, r1, r2): self.r1 = r1 self.r2 = r2 def getvalue(self): self.value = (self.r1.value << 8) + self.r2.value def setvalue(self): self.r1.value = self.value >> 8 self.r2.value = self.value & 0xff class regPC: def __init__(self): self.value = 0x0 def inc(self, length=1): self.value += length self.value = self.value & 0xffff def jump(self, address): self.value = address & 0xffff class regSP: def __init__(self): self.value = 0xfffe def inc(self): self.value += 2 self.value = self.value & 0xffff def dec(self): self.value -= 2 def setvalue(self): #print("SPSET:",hex(self.value)) pass # JUST TO MAKE LDX SIMPLER ONE_REG = reg() ONE_REG.value = 1 FL = Flags() halt = False A = reg() B = reg() C = reg() D = reg() E = reg() H = reg() L = reg() BC = Dreg(B, C) DE = Dreg(D, E) HL = Dreg(H, L) #E.value = 0x1 # Randomness loop PC = regPC() SP = regSP() memory = [] jumped = False print("RESERVING MEMORY...") for i in range(0x10000): memory.append(memByte()) print("MEMORY RESERVED.") print("LOADING MEMORY...") f = open(bootrom_file, "rb") rom_data = f.read() f.close() for i in range(len(rom_data)): memory[i+0x0].value = rom_data[i] f = open(rom_file, "rb") rom_data = f.read() f.close() for i in range(len(rom_data)): memory[i+0x597].value = rom_data[i] print("MEMORY LOADED.") def LDI(R, mem=False): PC.inc() if not mem: R.value = memory[PC.value].value else: R.getvalue() memory[R.value].value = memory[PC.value].value def LDX(R): PC.inc() low = memory[PC.value].value PC.inc() R.value = low + (memory[PC.value].value << 8) R.setvalue() def PUSH_R(R, mem=False): if not mem: memory[SP.value].value = R.value else: R.getvalue() memory[SP.value].value = memory[R.value].value SP.dec() def PUSH_RR(RR): RR.getvalue() memory[SP.value].value = RR.value & 0xff memory[SP.value + 1].value = RR.value >> 8 SP.dec() def POP_R(R, mem=False): SP.inc() if not mem: #print(hex(SP.value)) R.value = memory[SP.value].value else: R.getvalue() memory[R.value].value = memory[SP.value].value def POP_RR(RR): SP.inc() RR.value = memory[SP.value].value + (memory[SP.value + 1].value << 8) RR.setvalue() MOV_REGISTERS = [B, C, D, E, H, L, HL, A] MOVB_OPCODES = [0x09, 0x19, 0x29, 0x39, 0x49, 0x59, 0x69, 0x79] MOVC_OPCODES = [0x99, 0x99, 0xA9, 0xB9, 0xC9, 0xD9, 0xE9, 0xF9] MOVD_OPCODES = [0x0A, 0x1A, 0x2A, 0x3A, 0x4A, 0x5A, 0x6A, 0x7A] MOVE_OPCODES = [0x8A, 0x9A, 0xAA, 0xBA, 0xCA, 0xDA, 0xEA, 0xFA] MOVH_OPCODES = [0x0B, 0x1B, 0x2B, 0x3B, 0x4B, 0x5B, 0x6B, 0x7B] MOVL_OPCODES = [0x8B, 0x9B, 0xAB, 0xBB, 0xCB, 0xDB, 0xEB, 0xFB] MOVMHL_OPCODES = [0x0C, 0x1C, 0x2C, 0x3C, 0x4C, 0x5C, 0x6C, 0x7C] MOVA_OPCODES = [0x8C, 0x9C, 0xAC, 0xBC, 0xCC, 0xDC, 0xEC, 0xFC] def MOV(R1, R2index, mem=False): R2 = MOV_REGISTERS[R2index] if not mem: if R2index == 6: R2.getvalue() R1.value = memory[R2.value].value else: R1.value = R2.value else: memory[R1.value].value = R2.value R1.setvalue() def MOV_RR(RR1, RR2): RR2.getvalue() RR1.value = RR2.value RR1.setvalue() def ADD_8(value1, value2): nib = (value1 & 0xf) + (value2 & 0xf) value = value1 + value2 FL.clearFlags() if value & 0xff == 0: FL.setZero() if value & 0b10000000: FL.setNeg() if nib & 0xf0: FL.setHalf() if value >> 8: FL.setCarry() return value & 0xff def ADD_R(R, mem=False): if not mem: value = ADD_8(A.value, R.value) R.value = value else: R.getvalue() value = ADD_8(A.value, memory[R.value].value) memory[R.value].value = value def ADD_16(value1, value2): nib = (value1 & 0xf) + (value2 & 0xf) value = value1 + value2 FL.clearFlags() if value & 0xffff == 0: FL.setZero() if value & 0b1000000000000000: FL.setNeg() if nib & 0xf0: FL.setHalf() if value >> 16: FL.setCarry() return value & 0xffff def ADDX_RR(RR): RR.getvalue() value = ADD_16(A.value, RR.value) RR.value = value RR.setvalue() def SUB_8(value1, value2): value = value1 - value2 if value < 0: value += 0x100 FL.clearFlags() if value == 0: FL.setZero() if value & 0b10000000: FL.setNeg() if (value1 & 0xf) <= (value2 & 0xf): FL.setHalf() if value1 <= value2: FL.setCarry() return value & 0xff def SUB_R(R, compare_only, mem=False): if not mem: value = SUB_8(R.value, A.value) if not compare_only: R.value = value else: R.getvalue() value = SUB_8(memory[R.value].value, A.value) if not compare_only: memory[R.value].value = value def INC(R, mem=False): if not mem: value = ADD_8(ONE_REG.value, R.value) R.value = value else: R.getvalue() value = ADD_8(ONE_REG.value, memory[R.value].value) memory[R.value].value = value def DEC(R, mem=False): if not mem: value = SUB_8(R.value, ONE_REG.value) R.value = value else: R.getvalue() value = SUB_8(memory[R.value].value, ONE_REG.value) memory[R.value].value = value def AND_8(value1, value2): value = value1 & value2 FL.clearFlags() if value == 0: FL.setZero() if value & 0b10000000: FL.setNeg() return value & 0xff def AND_R(R, mem=False): if not mem: value = AND_8(A.value, R.value) R.value = value else: R.getvalue() value = AND_8(A.value, memory[R.value].value) memory[R.value].value = value def OR_8(value1, value2): value = value1 | value2 FL.clearFlags() if value == 0: FL.setZero() if value & 0b10000000: FL.setNeg() return value & 0xff def OR_R(R, mem=False): if not mem: value = OR_8(A.value, R.value) R.value = value else: R.getvalue() value = OR_8(A.value, memory[R.value].value) memory[R.value].value = value def XOR_8(value1, value2): value = value1 ^ value2 FL.clearFlags() if value == 0: FL.setZero() if value & 0b10000000: FL.setNeg() return value & 0xff def XOR_R(R, mem=False): if not mem: value = XOR_8(A.value, R.value) R.value = value else: R.getvalue() value = XOR_8(A.value, memory[R.value].value) memory[R.value].value = value def CMPS(R, mem=False): if not mem: Rval = R.value if Rval & 0b10000000: Rval = - ((0x100 - Rval) & 0xff) Aval = A.value if Aval & 0b10000000: Aval = - ((0x100 - Aval) & 0xff) FL.clearFlags() if Rval == Aval: FL.setZero() elif Rval < Aval: FL.setNeg() else: R.getvalue() Rval = memory[R.value].value if Rval & 0b10000000: Rval = - ((0x100 - Rval) & 0xff) Aval = A.value if Aval & 0b10000000: Aval = - ((0x100 - Aval) & 0xff) FL.clearFlags() if Rval == Aval: FL.setZero() elif Rval < Aval: FL.setNeg() def JUMP(): PC.inc() low = memory[PC.value].value PC.inc() high = memory[PC.value].value global jumped jumped = True PC.value = (high << 8) + low print("JUMP:",hex((high << 8) + low)) def REL_JUMP(): PC.inc() value = memory[PC.value].value if value & 0b10000000: value = - ((0x100 - value) & 0xff) # ACCORDING TO DOCUMENTATION RELATIVE JUMPS USE THE +2 PC INC PC.value += value screen_update() last_update = time.time() while not halt: b_up = memory[PC.value].readUpper() b_down = memory[PC.value].readLower() b_val = memory[PC.value].value jumped = False if time.time() > last_update + interval: screen_update() last_update = time.time() # Handle pygame events for event in pg.event.get(): # print("EVENT:",event.type) # input() pass if debug: pass#input() if debug or False: print(hex(PC.value), hex(b_val)) # if b_val in [0x86, 0x96, 0xA6, 0xB6, 0xC6, 0xD6, 0xE6, 0xF6]: # print("CMP R") # input() # if b_val == 0xF7: # print("CMPI") # input() # HCF (HALT) if b_val == 0x6C: halt = True # LDI R, xx if b_val == 0x20: LDI(B) elif b_val == 0x30: LDI(C) elif b_val == 0x40: LDI(D) elif b_val == 0x50: LDI(E) elif b_val == 0x60: LDI(H) elif b_val == 0x70: LDI(L) elif b_val == 0x80: LDI(HL, mem=True) elif b_val == 0x90: LDI(A) # LDX RR, xxyy elif b_val == 0x21: LDX(BC) elif b_val == 0x31: LDX(DE) elif b_val == 0x41: LDX(HL) elif b_val == 0x22: LDX(SP) # PUSH R elif b_val == 0x81: PUSH_R(B) elif b_val == 0x91: PUSH_R(C) elif b_val == 0xA1: PUSH_R(D) elif b_val == 0xB1: PUSH_R(E) elif b_val == 0xC1: PUSH_R(H) elif b_val == 0xD1: PUSH_R(L) elif b_val == 0xC0: PUSH_R(HL, mem=True) elif b_val == 0xD0: PUSH_R(A) # PUSH RR elif b_val == 0x51: PUSH_RR(BC) elif b_val == 0x61: PUSH_RR(DE) elif b_val == 0x71: PUSH_RR(HL) # POP R elif b_val == 0x82: POP_R(B) elif b_val == 0x92: POP_R(C) elif b_val == 0xA2: POP_R(D) elif b_val == 0xB2: POP_R(E) elif b_val == 0xC2: POP_R(H) elif b_val == 0xD2: POP_R(L) elif b_val == 0xC3: POP_R(HL, mem=True) elif b_val == 0xD3: POP_R(A) # POP RR elif b_val == 0x52: POP_RR(BC) elif b_val == 0x62: POP_RR(DE) elif b_val == 0x72: POP_RR(HL) # MOV R1, R2 elif b_val in MOVB_OPCODES: MOV(B, MOVB_OPCODES.index(b_val)) elif b_val in MOVC_OPCODES: MOV(C, MOVC_OPCODES.index(b_val)) elif b_val in MOVD_OPCODES: MOV(D, MOVD_OPCODES.index(b_val)) elif b_val in MOVE_OPCODES: MOV(E, MOVE_OPCODES.index(b_val)) elif b_val in MOVH_OPCODES: MOV(H, MOVH_OPCODES.index(b_val)) elif b_val in MOVL_OPCODES: MOV(L, MOVL_OPCODES.index(b_val)) elif b_val in MOVMHL_OPCODES: MOV(HL, MOVMHL_OPCODES.index(b_val), mem=True) elif b_val in MOVA_OPCODES: MOV(A, MOVA_OPCODES.index(b_val)) # MOV RR1, RR2 elif b_val == 0xED: MOV_RR(HL, BC) elif b_val == 0xFD: MOV_RR(HL, DE) # CLRFLAG elif b_val == 0x08: FL.clearFlags() # SETFLAG f, x elif b_val == 0x18: FL.setZero() elif b_val == 0x28: FL.clearZero() elif b_val == 0x38: FL.setNeg() elif b_val == 0x48: FL.clearNeg() elif b_val == 0x58: FL.setHalf() elif b_val == 0x68: FL.clearHalf() elif b_val == 0x78: FL.setCarry() elif b_val == 0x88: FL.clearCarry() # ADD R elif b_val == 0x04: ADD_R(B) elif b_val == 0x14: ADD_R(C) elif b_val == 0x24: ADD_R(D) elif b_val == 0x34: ADD_R(E) elif b_val == 0x44: ADD_R(H) elif b_val == 0x54: ADD_R(L) elif b_val == 0x64: ADD_R(HL, mem=True) elif b_val == 0x74: ADD_R(A) # ADDI xx elif b_val == 0xA7: PC.inc() value = ADD_8(A.value, memory[PC.value].value) A.value = value # ADDX RR elif b_val == 0x83: ADDX_RR(BC) elif b_val == 0x93: ADDX_RR(DE) elif b_val == 0xA3: ADDX_RR(HL) # SUB R | CMP R elif b_val == 0x84 or b_val == 0x86: SUB_R(B, b_val == 0x86) elif b_val == 0x94 or b_val == 0x96: SUB_R(C, b_val == 0x96) elif b_val == 0xA4 or b_val == 0xA6: SUB_R(D, b_val == 0xA6) elif b_val == 0xB4 or b_val == 0xB6: SUB_R(E, b_val == 0xB6) elif b_val == 0xC4 or b_val == 0xC6: SUB_R(H, b_val == 0xC6) elif b_val == 0xD4 or b_val == 0xD6: SUB_R(L, b_val == 0xD6) elif b_val == 0xE4 or b_val == 0xE6: SUB_R(HL, b_val == 0xE6, mem=True) elif b_val == 0xF4 or b_val == 0xF6: SUB_R(A, b_val == 0xF6) # SUBI xx | CMPI xx elif b_val == 0xB7 or b_val == 0xF7: PC.inc() value = SUB_8(A.value, memory[PC.value].value) if b_val == 0xB7: # SUBI xx A.value = value # INC R elif b_val == 0x03: INC(B) elif b_val == 0x13: INC(C) elif b_val == 0x23: INC(D) elif b_val == 0x33: INC(E) elif b_val == 0x43: INC(H) elif b_val == 0x53: INC(L) elif b_val == 0x63: INC(HL, mem=True) elif b_val == 0x73: INC(A) # INX RR elif b_val == 0xA8: BC.getvalue() BC.value += 1 BC.value & 0xffff BC.setvalue() elif b_val == 0xB8: DE.getvalue() DE.value += 1 DE.value & 0xffff DE.setvalue() elif b_val == 0xC8: HL.getvalue() HL.value += 1 HL.value & 0xffff HL.setvalue() # DEC R elif b_val == 0x07: DEC(B) elif b_val == 0x17: DEC(C) elif b_val == 0x27: DEC(D) elif b_val == 0x37: DEC(E) elif b_val == 0x47: DEC(H) elif b_val == 0x57: DEC(L) elif b_val == 0x67: DEC(HL, mem=True) elif b_val == 0x77: DEC(A) # AND R elif b_val == 0x05: AND_R(B) elif b_val == 0x15: AND_R(C) elif b_val == 0x25: AND_R(D) elif b_val == 0x35: AND_R(E) elif b_val == 0x45: AND_R(H) elif b_val == 0x55: AND_R(L) elif b_val == 0x65: AND_R(HL, mem=True) elif b_val == 0x75: AND_R(A) # ANDI xx elif b_val == 0xC7: PC.inc() value = AND_8(memory[PC.value].value, A.value) A.value = value # OR R elif b_val == 0x85: OR_R(B) elif b_val == 0x95: OR_R(C) elif b_val == 0xA5: OR_R(D) elif b_val == 0xB5: OR_R(E) elif b_val == 0xC5: OR_R(H) elif b_val == 0xD5: OR_R(L) elif b_val == 0xE5: OR_R(HL, mem=True) elif b_val == 0xF5: OR_R(A) # ORI xx elif b_val == 0xD7: PC.inc() value = OR_8(memory[PC.value].value, A.value) A.value = value # XOR R elif b_val == 0x06: XOR_R(B) elif b_val == 0x16: XOR_R(C) elif b_val == 0x26: XOR_R(D) elif b_val == 0x36: XOR_R(E) elif b_val == 0x46: XOR_R(H) elif b_val == 0x56: XOR_R(L) elif b_val == 0x66: XOR_R(HL, mem=True) elif b_val == 0x76: XOR_R(A) # XORI xx elif b_val == 0xE7: PC.inc() value = XOR_8(memory[PC.value].value, A.value) A.value = value # CMPS R elif b_val == 0x0D: CMPS(B) elif b_val == 0x1D: CMPS(C) elif b_val == 0x2D: CMPS(D) elif b_val == 0x3D: CMPS(E) elif b_val == 0x4D: CMPS(H) elif b_val == 0x5D: CMPS(L) elif b_val == 0x6D: CMPS(HL, mem=True) elif b_val == 0x7D: CMPS(A) # SIN elif b_val == 0xE0: A.value = ord(sys.stdin.buffer.read(1)) & 0xff pass # SOUT elif b_val == 0xE1: print(chr(A.value),end="",flush=True) if A.value == 7: print("[BELL]") pass # CLRSCR elif b_val == 0xF0: screen_clear() # DRAW elif b_val == 0xF1: x = C.value if x & 0b10000000: x = - ((0x100 - x) & 0xff) y = B.value if y & 0b10000000: y = - ((0x100 - y) & 0xff) screen_draw_line(x, y, A.value & 0xff) # JMP xxyy elif b_val == 0x0F: JUMP() # JMPcc xxyy elif b_val == 0x1F: if FL.z: JUMP() else: PC.inc(2) elif b_val == 0x2F: if not FL.z: JUMP() else: PC.inc(2) elif b_val == 0x3F: if FL.n: JUMP() else: PC.inc(2) elif b_val == 0x4F: if not FL.n: JUMP() else: PC.inc(2) elif b_val == 0x5F: if FL.h: JUMP() elif b_val == 0x6F: if not FL.h: JUMP() else: PC.inc(2) elif b_val == 0x7F: if FL.c: JUMP() else: PC.inc(2) elif b_val == 0x8F: if not FL.c: JUMP() else: PC.inc(2) # JMP xx elif b_val == 0x9F: REL_JUMP() # JMPcc xx elif b_val == 0xAF: if FL.z: REL_JUMP() else: PC.inc() elif b_val == 0xBF: if not FL.z: REL_JUMP() else: PC.inc() elif b_val == 0xCF: if FL.n: REL_JUMP() else: PC.inc() elif b_val == 0xDF: if not FL.n: REL_JUMP() else: PC.inc() elif b_val == 0xEF: if FL.h: REL_JUMP() else: PC.inc() elif b_val == 0xFF: if not FL.h: REL_JUMP() else: PC.inc() elif b_val == 0xEE: if FL.c: REL_JUMP() else: PC.inc() elif b_val == 0xFE: if not FL.c: REL_JUMP() else: PC.inc() # CALL xxyy elif b_val == 0x1E: memory[SP.value].value = (PC.value+3) & 0xff memory[SP.value + 1].value = (PC.value+3) >> 8 SP.dec() JUMP() # RET elif b_val == 0x0E: SP.inc() PC.value = memory[SP.value].value + (memory[SP.value + 1].value << 8) jumped = True # NOP elif b_val == 0x00: pass else: pass print("UNKNOWN:",hex(b_val),"@",hex(PC.value)) if debug: BC.getvalue() DE.getvalue() HL.getvalue() print("A:",hex(A.value),"B:",hex(B.value),"C:",hex(C.value),"D:",hex(D.value),"E:",hex(E.value),"H:",hex(H.value), "L:",hex(L.value),"BC:",hex(BC.value),"DE:",hex(DE.value),"HL:",hex(HL.value),"PC:",hex(PC.value),"SP:",hex(SP.value)) if not jumped: PC.inc() else: pass #print("JUMPED")
21,118
9,462
from __future__ import print_function import json import os class Data(): shared_state = {} def __init__(self): self.__dict__ = self.shared_state self.set_checklist(None) self.set_issues() def set_checklist(self, file_name): is_empty = file_name is None if is_empty: file_name = os.getcwd() + os.sep + "conf" + os.sep + "checklist.json" try: with open(file_name) as data_file: data = json.load(data_file) self.checklist = data["checklist"] except Exception as e: print(e) def get_checklist(self): return self.checklist def set_issues(self): file_name = os.getcwd() + os.sep + "conf" + os.sep + "issues.json" try: with open(file_name) as data_file: self.issues = json.load(data_file) except Exception as e: print(e) def get_issues(self): return self.issues def set_bugs(self, functionality_name, test_name, request, response): bug = { "request": request, "response": response } self.checklist["Functionality"][functionality_name]["tests"][test_name]["bugs"].append(bug) def set_notes(self, functionality_name, test_name, notes): self.checklist["Functionality"][functionality_name]["tests"][test_name]["notes"] = notes
1,423
438
# -*- test-case-name: twisted.test.test_stdio.StandardInputOutputTests.test_write -*- # Copyright (c) Twisted Matrix Laboratories. # See LICENSE for details. """ Main program for the child process run by L{twisted.test.test_stdio.StandardInputOutputTests.test_write} to test that ITransport.write() works for process transports. """ __import__('_preamble') import sys from twisted.internet import stdio, protocol from twisted.python import reflect class WriteChild(protocol.Protocol): def connectionMade(self): for ch in 'ok!': self.transport.write(ch) self.transport.loseConnection() def connectionLost(self, reason): reactor.stop() if __name__ == '__main__': reflect.namedAny(sys.argv[1]).install() from twisted.internet import reactor stdio.StandardIO(WriteChild()) reactor.run()
851
269
# calculate the load on each org import threading # cannot use from wpad_dispatch here, have to import whole module, # because of circular dependency import wpad_dispatch from wpad_utils import * from wlcg_wpad import getiporg orgcleanminutes = 5 orgcleantime = 0 # Minute records keep track of the number of requests in each minute class MinuteRecord: def __init__(self, now, older): self.minute = now # minute of the record self.requests = 0 # number of requests this minute self.next = None # next MinuteRecord if older != None: older.next = self # point older record to this one class OrgData: def __init__(self): self.lock = threading.Lock() # lock for this org self.overloadminute = 0 # minute last overload triggered self.total = 0 # total number of requests tracked self.newest = None # newest MinuteRecord self.oldest = None # oldest MinuteRecord orgdata = {} # lock for adding, deleting, and looking up an org orgdatalock = threading.Lock() # return triple of org name, minutes remaining in an overload, # and percent of limit in the last minutes being tracked def orgload(remoteip, limit, minutes, persist, now): global orgcleantime org = getiporg(remoteip) if org == None: return None, 0, 0 # See if this org is excluded # wlcgwpadconf is occasionally replaced, so use a local variable for it conf = wpad_dispatch.wlcgwpadconf if 'overload' in conf and 'excludes' in conf['overload'] and \ org in conf['overload']['excludes']: return None, 0, 0 now = now / 60 # this function deals only with minutes orgdatalock.acquire() if orgcleantime <= now - orgcleanminutes: # clean out orgs that have had no updates in minutes or overload in # persist minutes, except current org orgcleantime = now numorgs = 0 delorgs = 0 for oldorg in list(orgdata): numorgs += 1 if org == oldorg: continue data = orgdata[oldorg] if persist < now - data.overloadminute and \ data.newest != None and data.newest.minute < now - minutes: # Note that there is a race condition where this could # delete an org from orgdata at the same time as another # request comes in to another thread to prolong it, but # that would only result in the loss of one count, it would # not be fatal. The only way data.newest can equal None # is if the organization is in the process of being created # by another thread, leave that one alone. del orgdata[oldorg] delorgs += 1 if delorgs > 0: orgdatalock.release() logmsg('-', '-', '', 'cleaned load data from ' + str(delorgs) + ' orgs, ' + str(numorgs-delorgs) + ' still active') orgdatalock.acquire() # get the data for this org if org in orgdata: data = orgdata[org] else: data = OrgData() orgdata[org] = data orgdatalock.release() data.lock.acquire() # remove any minute records that are too old record = data.oldest while record != None and record.minute <= now - minutes: data.total -= record.requests record = record.next data.oldest = record record = data.newest if record == None or record.minute != now: # add new minute record record = MinuteRecord( now, record ) data.newest = record if data.oldest == None: data.oldest = record # add one to this minute and the total record.requests += 1 data.total = data.total + 1 percent = int(data.total * 100.0 / limit) if percent > 100: data.overloadminute = now overloadminute = data.overloadminute data.lock.release() return ( org, persist - (now - overloadminute), percent )
4,133
1,157
import requests from datetime import datetime import time headers = { 'Content-Type': 'application/json' } with open('test/logs.txt', 'r') as f: logs = f.read().split('\n') levels = ['INFO', 'WARN', 'DEBUG', 'ERROR', 'TRACE'] messages = [] for i in range(0, len(logs)): if i % 100 == 0: print('{} of {}'.format(i, len(logs))) data = { 'messages': messages } requests.post('http://localhost:9090/insert', json=data, headers=headers) messages = [] now = datetime.now() message = { 'year': now.year, 'month': now.month, 'day': now.day, 'hour': now.hour, 'minute': now.minute, 'second': now.second, 'service': 'foobar', 'message': logs[i], 'level': levels[i % 5] } time.sleep(0.001) messages.append(message)
865
295
import os import sys sys.path.append(os.path.abspath('')) # Raises linting error because not at top of file # Not sure how to resolve this with the pathing from src import uploadJson # noqa: E402 import src.config as config # noqa: E402 # Taking out of commission until new geojson format requested developed # def test_main(): # from src import main # import json # import geojson # #from src import main # main_dict = main.main(['county']) # for v in main_dict.values(): # v_str = json.dumps(v) # v_geojson = geojson.loads(v_str) # assert v_geojson.is_valid == True def test_requirements(): import pkg_resources requirements_path = "requirements.txt" with open(requirements_path) as f: requirements = pkg_resources.parse_requirements(f) for r in requirements: r = str(r) pkg_resources.require(r) # breakpoint() def test_auth(): db = uploadJson.auth_firebase() cook = db.reference('/county_data/17031').get() assert cook['NAME'] == 'Cook County, Illinois' def test_secrets(): assert type(config.CENSUS_KEY) == str assert type(config.FIREBASE_SERVICE_KEY) == str assert config.CENSUS_KEY != '' assert config.FIREBASE_SERVICE_KEY != ''
1,282
419
#!/usr/bin/env python """ Code is personal property of the owner.Rights reserved by the user only. Codes on https://github.com/kt97679/tetris Attempt to spread the open source knowledge :) """ import sys import select import tty import termios import random import os import time import fcntl if (sys.hexversion >> 16) >= 0x202: FCNTL = fcntl else: import FCNTL import contextlib PLAYFIELD_W = 10 PLAYFIELD_H = 20 PLAYFIELD_X = 30 PLAYFIELD_Y = 1 BORDER_COLOR = 'yellow' HELP_X = 58 HELP_Y = 1 HELP_COLOR = 'cyan' SCORE_X = 1 SCORE_Y = 2 SCORE_COLOR = 'green' NEXT_X = 14 NEXT_Y = 11 GAMEOVER_X = 1 GAMEOVER_Y = PLAYFIELD_H + 3 INITIAL_MOVE_DOWN_DELAY = 1.0 DELAY_FACTOR = 0.8 LEVEL_UP = 20 NEXT_EMPTY_CELL = " " PLAYFIELD_EMPTY_CELL = " ." FILLED_CELL = "[]" class TetrisScreen: def __init__(self): self.s = '' self.no_color = False self.color = { 'red': 1, 'green': 2, 'yellow': 3, 'blue': 4, 'fuchsia': 5, 'cyan': 6, 'white': 7 } def xyprint(self, x, y, s): self.s += "\x1b[{0};{1}H{2}".format(y, x, s) def flush(self): sys.stdout.write(self.s) sys.stdout.flush() self.s = '' def puts(self, s): self.s += s def clear_screen(self): self.s += "\x1b[2J" def show_cursor(self): self.s += "\x1b[?25h" def hide_cursor(self): self.s += "\x1b[?25l" def set_fg(self, c): if self.no_color: return self.s += "\x1b[3{0}m".format(self.color.get(c, 7)) def set_bg(self, c): if self.no_color: return self.s += "\x1b[4{0}m".format(self.color.get(c, 7)) def reset_colors(self): self.s += "\x1b[0m" def set_bold(self): self.s += "\x1b[1m" def get_random_color(self): k = self.color.keys() random.shuffle(k) return k[0] def toggle_color(self): self.no_color ^= True class TetrisScreenItem(object): def __init__(self, screen): self.visible = True self.screen = screen def show(self): if self.visible: self.draw(True) def hide(self): if self.visible: self.draw(False) def toggle(self): self.visible ^= True self.draw(self.visible) class TetrisHelp(TetrisScreenItem): def __init__(self, screen): super(TetrisHelp, self).__init__(screen) self.color = HELP_COLOR self.text = [ " Use cursor keys", " or", " s: rotate", "a: left, d: right", " space: drop", " q: quit", " c: toggle color", "n: toggle show next", "h: toggle this help" ] def draw(self, visible): self.screen.set_bold() self.screen.set_fg(self.color) i = 0 for s in self.text: if not visible: s = ' ' * len(s) self.screen.xyprint(HELP_X, HELP_Y + i, s) i += 1 self.screen.reset_colors() class TetrisPlayField: def __init__(self, screen): self.screen = screen self.cells = [[None] * PLAYFIELD_W for i in range(0, PLAYFIELD_H)] def show(self): y = 0 for row in self.cells: self.screen.xyprint(PLAYFIELD_X, PLAYFIELD_Y + y, "") y += 1 for cell in row: if cell == None: self.screen.puts(PLAYFIELD_EMPTY_CELL) else: self.screen.set_fg(cell) self.screen.set_bg(cell) self.screen.puts(FILLED_CELL) self.screen.reset_colors() def flatten_piece(self, piece): for cell in piece.get_cells(): self.cells[cell[1]][cell[0]] = piece.color def process_complete_lines(self): cells = [row for row in self.cells if None in row] complete_lines = PLAYFIELD_H - len(cells) if complete_lines > 0: self.cells = [[None] * PLAYFIELD_W for i in range(0, complete_lines)] + cells return complete_lines def draw_border(self): self.screen.set_bold() self.screen.set_fg(BORDER_COLOR) for y in range(0, PLAYFIELD_H): # 2 because border is 2 characters thick self.screen.xyprint(PLAYFIELD_X - 2, PLAYFIELD_Y + y, "<|") # 2 because each cell on play field is 2 characters wide self.screen.xyprint(PLAYFIELD_X + PLAYFIELD_W * 2, PLAYFIELD_Y+ y, "|>") y = 0 for s in ['==', '\/']: self.screen.xyprint(PLAYFIELD_X, PLAYFIELD_Y + PLAYFIELD_H + y, s * PLAYFIELD_W) y += 1 self.screen.reset_colors() def position_ok(self, cells): return all( (0 <= x < PLAYFIELD_W) and (0 <= y < PLAYFIELD_H) and self.cells[y][x] is None for x, y in cells ) class TetrisPiece(TetrisScreenItem): configurations = [ # 0123 # 4567 # 89ab # cdef [0x1256], # square [0x159d, 0x4567], # line [0x4512, 0x0459], # s [0x0156, 0x1548], # z [0x159a, 0x8456, 0x0159, 0x2654], # l [0x1598, 0x0456, 0x2159, 0xa654], # inverted l [0x1456, 0x1596, 0x4569, 0x4159] # t ] def __init__(self, screen, origin, visible): super(TetrisPiece, self).__init__(screen) self.color = screen.get_random_color() self.data = random.choice(self.configurations) self.symmetry = len(self.data) self.position = 0, 0, random.randint(0, self.symmetry - 1) self.origin = origin self.visible = visible self.empty_cell = NEXT_EMPTY_CELL def get_cells(self, new_position=None): x, y, z = new_position or self.position data = self.data[z] return [[x + ((data >> (i * 4)) & 3), y + ((data >> (i * 4 + 2)) & 3)] for i in range(0, 4)] def draw(self, visible): if visible: self.screen.set_fg(self.color) self.screen.set_bg(self.color) s = FILLED_CELL else: s = self.empty_cell for cell in self.get_cells(): self.screen.xyprint(self.origin[0] + cell[0] * 2, self.origin[1] + cell[1], s) self.screen.reset_colors() def set_xy(self, x, y): self.position = x, y, self.position[2] def new_position(self, dx, dy, dz): x, y, z = self.position return x + dx, y + dy, (z + dz) % self.symmetry class TetrisScore: def __init__(self, screen, tetris_input_processor): self.screen = screen self.tetris_input_processor = tetris_input_processor self.score = 0 self.level = 1 self.lines_completed = 0 def update(self, complete_lines): self.lines_completed += complete_lines self.score += (complete_lines * complete_lines) if self.score > LEVEL_UP * self.level: self.level += 1 self.tetris_input_processor.decrease_delay() self.show() def show(self): self.screen.set_bold() self.screen.set_fg(SCORE_COLOR) self.screen.xyprint(SCORE_X, SCORE_Y, "Lines completed: {0}".format(self.lines_completed)) self.screen.xyprint(SCORE_X, SCORE_Y + 1, "Level: {0}".format(self.level)) self.screen.xyprint(SCORE_X, SCORE_Y + 2, "Score: {0}".format(self.score)) self.screen.reset_colors() class TetrisController: def __init__(self, screen, tetris_input_processor): self.screen = screen self.next_piece_visible = True self.running = True self.help = TetrisHelp(screen) self.score = TetrisScore(screen, tetris_input_processor) self.play_field = TetrisPlayField(screen) self.get_next_piece() self.get_current_piece() self.redraw_screen() screen.flush() def get_current_piece(self): self.next_piece.hide() self.current_piece = self.next_piece self.current_piece.set_xy((PLAYFIELD_W - 4) / 2, 0) if not self.play_field.position_ok(self.current_piece.get_cells()): self.cmd_quit() return self.current_piece.visible = True self.current_piece.empty_cell = PLAYFIELD_EMPTY_CELL self.current_piece.origin = (PLAYFIELD_X, PLAYFIELD_Y) self.current_piece.show() self.get_next_piece() def get_next_piece(self): self.next_piece = TetrisPiece( self.screen, (NEXT_X, NEXT_Y), self.next_piece_visible, ) self.next_piece.show() def redraw_screen(self): self.screen.clear_screen() self.screen.hide_cursor() self.play_field.draw_border() for o in [self.help, self.play_field, self.score, self.next_piece, self.current_piece]: o.show() def cmd_quit(self): self.running = False self.screen.xyprint(GAMEOVER_X, GAMEOVER_Y, "Game over!") self.screen.xyprint(GAMEOVER_X, GAMEOVER_Y + 1, "") self.screen.show_cursor() def process_fallen_piece(self): self.play_field.flatten_piece(self.current_piece) complete_lines = self.play_field.process_complete_lines() if complete_lines > 0: self.score.update(complete_lines) self.play_field.show() def move(self, dx, dy, dz): position = self.current_piece.new_position(dx, dy, dz) if self.play_field.position_ok(self.current_piece.get_cells(position)): self.current_piece.hide() self.current_piece.position = position self.current_piece.show() return True return (dy == 0) def cmd_right(self): self.move(1, 0, 0) def cmd_left(self): self.move(-1, 0, 0) def cmd_rotate(self): self.move(0, 0, 1) def cmd_down(self): if self.move(0, 1, 0): return True self.process_fallen_piece() self.get_current_piece() return False def cmd_drop(self): while self.cmd_down(): pass def toggle_help(self): self.help.toggle() def toggle_next(self): self.next_piece_visible ^= True self.next_piece.toggle() def toggle_color(self): self.screen.toggle_color() self.redraw_screen() @contextlib.contextmanager def nonblocking_input(): fd = sys.stdin try: flags = fcntl.fcntl(fd, FCNTL.F_GETFL) flags = flags | os.O_NONBLOCK fcntl.fcntl(fd, FCNTL.F_SETFL, flags) yield finally: flags = fcntl.fcntl(fd, FCNTL.F_GETFL) flags = flags & ~os.O_NONBLOCK fcntl.fcntl(fd, FCNTL.F_SETFL, flags) @contextlib.contextmanager def tcattr(): try: old_settings = termios.tcgetattr(sys.stdin) yield finally: termios.tcsetattr(sys.stdin, termios.TCSADRAIN, old_settings) class TetrisInputProcessor: delay = INITIAL_MOVE_DOWN_DELAY def decrease_delay(self): self.delay *= DELAY_FACTOR def run(): input_processor = TetrisInputProcessor() with nonblocking_input(), tcattr(): # tty.setcbreak(sys.stdin.fileno()) tty.setraw(sys.stdin.fileno()) key = [0, 0, 0] ts = TetrisScreen() ts.clear_screen() tc = TetrisController(ts, input_processor) commands = { "\x03": tc.cmd_quit, "q": tc.cmd_quit, "C": tc.cmd_right, "d": tc.cmd_right, "D": tc.cmd_left, "a": tc.cmd_left, "A": tc.cmd_rotate, "s": tc.cmd_rotate, " ": tc.cmd_drop, "h": tc.toggle_help, "n": tc.toggle_next, "c": tc.toggle_color } last_move_down_time = time.time() while tc.running: cmd = None now = time.time() select_timeout = input_processor.delay - (now - last_move_down_time) if select_timeout < 0: tc.cmd_down() ts.flush() last_move_down_time = now select_timeout = input_processor.delay if select.select([sys.stdin], [], [], input_processor.delay)[0]: s = sys.stdin.read(16) for c in s: key[2] = key[1] key[1] = key[0] key[0] = c if key[2] == '\x1b' and key[1] == '[': # x1b is ESC cmd = commands.get(key[0], None) else: cmd = commands.get(key[0].lower(), None) if cmd: cmd() ts.flush() if __name__ == '__main__': run()
12,939
4,673
import time import json def export_users(ucm_axl): """ retrieve users from ucm """ try: user_list = ucm_axl.get_users( tagfilter={ "userid": "", "firstName": "", "lastName": "", "directoryUri": "", "telephoneNumber": "", "enableCti": "", "mailid": "", "primaryExtension": {"pattern": "", "routePartitionName": ""}, "enableMobility": "", "homeCluster": "", "associatedPc": "", "enableEmcc": "", "imAndPresenceEnable": "", "serviceProfile": {"_value_1": ""}, "status": "", "userLocale": "", "title": "", "subscribeCallingSearchSpaceName": "", } ) all_users = [] for user in user_list: # print(user) user_details = {} user_details['userid'] = user.userid user_details['firstName'] = user.firstName user_details['lastName'] = user.lastName user_details['telephoneNumber'] = user.telephoneNumber user_details['primaryExtension'] = user.primaryExtension.pattern user_details['directoryUri'] = user.directoryUri user_details['mailid'] = user.mailid all_users.append(user_details) print( f"{user_details.get('userid')} -- {user_details.get('firstName')} {user_details.get('lastName')}: {user_details.get('primaryExtension')}" ) print("-" * 35) print(f"number of users: {len(all_users)}") # print(user_list) # print(json.dumps(all_users, indent=2)) return all_users except Exception as e: print(e) return [] def export_phones(ucm_axl): """ Export Phones """ try: phone_list = ucm_axl.get_phones( tagfilter={ "name": "", "description": "", "product": "", "model": "", "class": "", "protocol": "", "protocolSide": "", "callingSearchSpaceName": "", "devicePoolName": "", "commonDeviceConfigName": "", "commonPhoneConfigName": "", "networkLocation": "", "locationName": "", "mediaResourceListName": "", "networkHoldMohAudioSourceId": "", "userHoldMohAudioSourceId": "", "loadInformation": "", "securityProfileName": "", "sipProfileName": "", "cgpnTransformationCssName": "", "useDevicePoolCgpnTransformCss": "", "numberOfButtons": "", "phoneTemplateName": "", "primaryPhoneName": "", "loginUserId": "", "defaultProfileName": "", "enableExtensionMobility": "", "currentProfileName": "", "loginTime": "", "loginDuration": "", # "currentConfig": "", "ownerUserName": "", "subscribeCallingSearchSpaceName": "", "rerouteCallingSearchSpaceName": "", "allowCtiControlFlag": "", "alwaysUsePrimeLine": "", "alwaysUsePrimeLineForVoiceMessage": "", } ) all_phones = [] for phone in phone_list: # print(phone) phone_details = { "name": phone.name, "description": phone.description, "product": phone.product, "model": phone.model, "protocol": phone.protocol, "protocolSide": phone.protocolSide, "callingSearchSpaceName": phone.callingSearchSpaceName._value_1, "devicePoolName": phone.defaultProfileName._value_1, "commonDeviceConfigName": phone.commonDeviceConfigName._value_1, "commonPhoneConfigName": phone.commonPhoneConfigName._value_1, "networkLocation": phone.networkLocation, "locationName": phone.locationName._value_1, "mediaResourceListName": phone.mediaResourceListName._value_1, "networkHoldMohAudioSourceId": phone.networkHoldMohAudioSourceId, "userHoldMohAudioSourceId": phone.userHoldMohAudioSourceId, "loadInformation": phone.loadInformation, "securityProfileName": phone.securityProfileName._value_1, "sipProfileName": phone.sipProfileName._value_1, "cgpnTransformationCssName": phone.cgpnTransformationCssName._value_1, "useDevicePoolCgpnTransformCss": phone.useDevicePoolCgpnTransformCss, "numberOfButtons": phone.numberOfButtons, "phoneTemplateName": phone.phoneTemplateName._value_1, "primaryPhoneName": phone.primaryPhoneName._value_1, "loginUserId": phone.loginUserId, "defaultProfileName": phone.defaultProfileName._value_1, "enableExtensionMobility": phone.enableExtensionMobility, "currentProfileName": phone.currentProfileName._value_1, "loginTime": phone.loginTime, "loginDuration": phone.loginDuration, # "currentConfig": phone.currentConfig, "ownerUserName": phone.ownerUserName._value_1, "subscribeCallingSearchSpaceName": phone.subscribeCallingSearchSpaceName._value_1, "rerouteCallingSearchSpaceName": phone.rerouteCallingSearchSpaceName._value_1, "allowCtiControlFlag": phone.allowCtiControlFlag, "alwaysUsePrimeLine": phone.alwaysUsePrimeLine, "alwaysUsePrimeLineForVoiceMessage": phone.alwaysUsePrimeLineForVoiceMessage, } line_details = ucm_axl.get_phone(name=phone.name) # print(line_details.lines.line) try: for line in line_details.lines.line: # print(line) phone_details[f"line_{line.index}_dirn"] = line.dirn.pattern phone_details[f"line_{line.index}_routePartitionName"] = line.dirn.routePartitionName._value_1 phone_details[f"line_{line.index}_display"] = line.display phone_details[f"line_{line.index}_e164Mask"] = line.e164Mask except Exception as e: print(e) all_phones.append(phone_details) print( f"exporting: {phone.name}: {phone.model} - {phone.description}") print("-" * 35) print(f"number of phones: {len(all_phones)}") return all_phones except Exception as e: print(e) return [] def export_siptrunks(ucm_axl): try: all_sip_trunks = [] sip_trunks = ucm_axl.get_sip_trunks( tagfilter={ "name": "", "description": "", "devicePoolName": "", "callingSearchSpaceName": "", "sipProfileName": "", "mtpRequired": "", "sigDigits": "", "destAddrIsSrv": "", } ) for siptrunk in sip_trunks: trunk = {} trunk["name"] = siptrunk.name trunk["description"] = siptrunk.description trunk["devicePoolName"] = siptrunk.devicePoolName._value_1 trunk["sipProfileName"] = siptrunk.sipProfileName._value_1 trunk["callingSearchSpace"] = siptrunk.callingSearchSpaceName._value_1 trunk["mtpRequired"] = siptrunk.mtpRequired trunk["sigDigits"] = siptrunk.sigDigits._value_1 # TODO: get_siptrunk details for destinations trunk_details = ucm_axl.get_sip_trunk(name=siptrunk.name) destinations = trunk_details['return']['sipTrunk']['destinations']['destination'] # print(destinations) for count, destination in enumerate(destinations): trunk[f'addressIpv4_{count}'] = destination.addressIpv4 trunk[f'port_{count}'] = destination.port trunk[f'sortOrder_{count}'] = destination.sortOrder all_sip_trunks.append(trunk) # print(siptrunk) print(f"exporting: {siptrunk.name}: {siptrunk.description}") print("-" * 35) print(f"number of siptrunks: {len(all_sip_trunks)}") return all_sip_trunks except Exception as e: print(e) return [] def export_phone_registrations(ucm_axl, ucm_ris): """ Export Phone Registrations """ nodes = ucm_axl.list_process_nodes() del nodes[0] # remove EnterpriseWideData node subs = [] for node in nodes: subs.append(node.name) phones = ucm_axl.get_phones(tagfilter={"name": ""}) all_phones = [] phone_reg = [] reg = {} for phone in phones: all_phones.append(phone.name) def limit(all_phones, n=1000): return [ all_phones[i: i + n] for i in range(0, len(all_phones), n) ] groups = limit(all_phones) for group in groups: registered = ucm_ris.checkRegistration(group, subs) if registered["TotalDevicesFound"] < 1: print("no devices found!") else: reg["user"] = registered["LoginUserId"] reg["regtime"] = time.strftime( "%Y-%m-%d %H:%M:%S", time.localtime(registered["TimeStamp"])) for item in registered["IPAddress"]: reg["ip"] = item[1][0]["IP"] for item in registered["LinesStatus"]: reg["primeline"] = item[1][0]["DirectoryNumber"] reg["name"] = registered["Name"] print(f"exporting: {reg['name']}: {reg['ip']} - {reg['regtime']}") phone_reg.append(reg) print("-" * 35) print(f"number of registered phones: {len(phone_reg)}") return phone_reg def export_translations(ucm_axl): try: all_translations = [] translations = ucm_axl.get_translations() for translation in translations: # print(translation) xlate = {} xlate["pattern"] = translation.pattern xlate["routePartition"] = translation.routePartitionName._value_1 xlate["description"] = translation.description xlate["callingSearchSpace"] = translation.callingSearchSpaceName._value_1 xlate["callingPartyTransformationMask"] = translation.callingPartyTransformationMask xlate["digitDiscardInstructionName"] = translation.digitDiscardInstructionName._value_1 xlate["prefixDigitsOut"] = translation.prefixDigitsOut xlate["calledPartyTransformationMask"] = translation.calledPartyTransformationMask all_translations.append(xlate) print( f"exporting: {xlate['pattern']}: {xlate['routePartition']} - {xlate['description']} --> {xlate['calledPartyTransformationMask']}") print("-" * 35) print(f"number of translations: {len(all_translations)}") return all_translations except Exception as e: return []
11,471
3,130
host = "localhost" port = 9999 dboptions = { "host": "194.67.198.163", "user": "postgres", "password": "werdwerd2012", "database": "zno_bot", 'migrate': True } API_PATH = '/api/' API_VERSION = 'v1' API_URL = API_PATH + API_VERSION
256
119
import pandas as pd import numpy as np from sklearn.datasets import load_iris from sklearn.naive_bayes import GaussianNB import pickle def train_iris_nb(): """Train a GaussianNB model on iris dataset.""" X, y_train = load_iris(return_X_y=True, as_frame=True) colnames = X.columns X_train = X.values model = GaussianNB() model.fit(X_train, y_train) return model def dump_model(model_path, model): """Save model as binary pickle file.""" with open(model_path, 'wb') as file: pickle.dump(model, file) def load_model(model_path): """Load model to return for future use.""" with open(model_path, 'rb') as file: model = pickle.load(file) return model def main(): model = train_iris_nb() dump_model('model.pickle', model) if __name__ == '__main__': main()
829
283
# -*- coding: utf-8 -*- from .parserules import parseAll class UfoscrapyPipeline(object): def process_item(self, item, spider): """ Parsitaan regexin ja datetimen avulla kentät oikeaan muotoon. """ try: item['loc'] = item['loc'][0] except: item['loc'] = "" try: item['shape'] = item['shape'][0] except: item['shape'] = "" try: item['state'] = item['state'][0] except: item['state'] = "" try: item['duration'] = item['duration'][0] except: item['duration'] = "" parseAll(item) return item
746
227
# ***************************************************************************** # precompressed/context_processors.py # ***************************************************************************** """ A set of request processors that return dictionaries to be merged into a template context. Each function takes the request object as its only parameter and returns a dictionary to add to the context. These are referenced from the setting TEMPLATE_CONTEXT_PROCESSORS and used by RequestContext. """ from __future__ import absolute_import, division from __future__ import print_function, unicode_literals from precompressed import utils # ***************************************************************************** # accepts_gzip # ***************************************************************************** def accepts_gzip(request): """ defines ACCEPTS_GZIP -- a boolean which reflects whether the request accepts Content-Type: gzip. """ return {'ACCEPTS_GZIP': utils.accepts_gzip(request)}
1,028
243
#Import modules import os import pandas as pd import numpy as np from pandas import DatetimeIndex import dask import scipy import time import glob import torch import torch.nn as nn from live_plotter import live_plotter import matplotlib.pyplot as plt from mpl_toolkits import mplot3d from functools import partial from abc import ABCMeta, abstractmethod import plottingTools import pytorchModel import loadData class pytorchFwdModel(pytorchModel.pytorchModel) : ####################################################################################################### #Construction functions ####################################################################################################### def __init__(self, learningRate, hyperParameters, nbUnitsPerLayer, nbFactors, modelName = "./bestPyTorchFwdModel"): super().__init__(learningRate, hyperParameters, nbUnitsPerLayer, nbFactors, modelName = modelName) def buildModel(self): self.fe = pytorchModel.Functional_encoder(self.nbFactors + 1) #Neural network architecture return ####################################################################################################### #Evaluation functions ####################################################################################################### def evalBatch(self, batch, code): batchLogMoneyness = self.getLogMoneyness(batch) scaledMoneyness = (batchLogMoneyness.values - self.MeanLogMoneyness) / self.StdLogMoneyness logMoneynessTensor = torch.Tensor(np.expand_dims(scaledMoneyness, 1)).float() #Log moneyness # for j in np.random.choice(len(test[k]), 10): # filt = test[k].nBizDays >= 10 batchLogMat = self.getLogMaturities(batch) scaledMat = (batchLogMat.values - self.MeanLogMaturity) / self.StdLogMaturity logMaturity = torch.tensor( np.expand_dims(scaledMat, 1) , requires_grad=True).float() scaledFwd = (batch[2].values - self.MeanFwd) / self.StdFwd fwdTensor = torch.tensor( np.expand_dims(scaledFwd, 1) ).float() codeTensor = code.repeat(batch[0].shape[0], 1).float() refVol = torch.tensor(batch[0].values) inputTensor = torch.cat((logMoneynessTensor, logMaturity, fwdTensor, codeTensor), dim=1) outputTensor = self.fe( inputTensor )[:, 0] loss = torch.mean( (outputTensor - refVol)[~torch.isnan(outputTensor)] ** 2 )#torch.nanmean( (outputTensor - refVol) ** 2 ) return inputTensor, outputTensor, loss, logMaturity, codeTensor, logMoneynessTensor def commonEvalSingleDayWithoutCalibration(self, initialValueForFactors, dataSetList, computeSensi = False): #Rebuild tensor graph self.restoringGraph() #Build tensor for reconstruction nbObs = 1 if initialValueForFactors.ndim == 1 else initialValueForFactors.shape[0] nbPoints = dataSetList[1].shape[0] if dataSetList[1].ndim == 1 else dataSetList[1].shape[1] nbFactors = self.nbFactors reshapedValueForFactors = np.reshape([initialValueForFactors], (nbObs,nbFactors)) self.code = pytorchModel.Code(nbObs, self.nbFactors, initialValue = reshapedValueForFactors) #Latent variables codeTensor = self.code.code[k, :].repeat(nbPoints, 1) batchLogMoneyness = self.getLogMoneyness(dataSetList) scaledMoneyness = (batchLogMoneyness.values - self.MeanLogMoneyness) / self.StdLogMoneyness logMoneynessTensor = torch.Tensor(np.expand_dims(scaledMoneyness.values, 1)).float() #Log moneyness scaledFwd = (dataSetList[2].values - self.MeanFwd) / self.StdFwd fwdTensor = torch.tensor( np.expand_dims(scaledFwd, 1) ).float() # for j in np.random.choice(len(test[k]), 10): # filt = test[k].nBizDays >= 10 batchLogMat = self.getLogMaturities(dataSetList) scaledMat = (batchLogMat.values - self.MeanLogMaturity) / self.StdLogMaturity logMaturity = torch.tensor( np.expand_dims(scaledMat, 1) ).float() inputTensor = torch.cat((logMoneynessTensor, logMaturity, fwdTensor, codeTensor), dim=1) outputTensor = self.fe( inputTensor )[:, 0] self.restoreWeights() #Build tensor for reconstruction # print("nbPoints : ", nbPoints) # print("initialValueForFactors : ", initialValueForFactors) # print("inputFeatures : ", inputFeatures) # print("outputFeatures : ", outputFeatures) # print("outputTensor : ", self.outputTensor) reconstructedSurface = outputTensor.detach().numpy().reshape(batch[0].shape) inputTensor = torch.cat((strikes, logMaturity, codeTensor), dim=1) #if computeSensi : # inputTensor.requires_grad = True outputTensor = self.fe( inputTensor )[:, 0] reshapedJacobian = None if computeSensi : reshapedJacobian = np.ones((nbObs, nbPoints, nbFactors)) if initialValueForFactors.ndim != 1 else np.ones((nbPoints, nbFactors)) #for p in range(nbPoints) : # output.backward() # jacobian = input.grad.data # reshapedJacobian = tf.reshape(jacobian, shape = [nbObs, nbPoints, nbFactors]) # if self.verbose : # print(reshapedJacobian) calibratedSurfaces = outputTensor factorSensi = None if initialValueForFactors.ndim == 1 : calibratedSurfaces = np.reshape(reconstructedSurface, (nbPoints)) if reshapedJacobian is not None : factorSensi = np.reshape(reshapedJacobian, (nbPoints, nbFactors)) elif initialValueForFactors.ndim == 2 : calibratedSurfaces = np.reshape(reconstructedSurface, (nbObs,nbPoints)) if reshapedJacobian is not None : factorSensi = np.reshape(reshapedJacobian, (nbObs, nbPoints, nbFactors)) return calibratedSurfaces, factorSensi
6,612
2,111
import numpy as np import tensorflow as tf from tensorflow import keras import warnings warnings.filterwarnings('ignore') mnist = tf.keras.datasets.mnist (X_train, y_train), (X_test, y_test) = mnist.load_data() X_train, X_test = X_train / 255.0, X_test / 255.0 X_train.shape = (60000, 28, 28) model = tf.keras.Sequential([ tf.keras.layers.Flatten(input_shape=(28, 28)), tf.keras.layers.Dense(128, activation='relu'), tf.keras.layers.Dropout(0.2), tf.keras.layers.Dense(10) ]) # output shape is (None, 10) ie batch size, 10 # logits is the inverse of sigmoid logits = model(X_train[:1]).numpy() loss_fn = keras.losses.SparseCategoricalCrossentropy(from_logits=True) model.compile(optimizer='adam', loss=loss_fn, metrics=['accuracy']) model.fit(X_train, y_train, epochs=1) model.evaluate(X_test, y_test, verbose=2) probab_model = keras.Sequential([ model, keras.layers.Softmax() ])
911
380
__version__ = "2.1.0.dev0" from pyquil.quil import Program from pyquil.api import list_quantum_computers, get_qc
114
47
#!/bin/env python3.5 import math def test_func(x): return x*x*x+1.1*x*x+0.9*x-1.4 def df_test_func(x): return 3*x*x+2.2*x+0.9 # binary seperation def binary_seperation(func, x, ap=0.001): list(x).sort() x2 = x.pop() x1 = x.pop() y1 = func(x1) y2 = func(x2) if y1 == 0: return x1 if y2 == 0: return x2 if y1*y2 > 0: return "" while True: avg_x = (x1 + x2)/2 avg_y = func(avg_x) if avg_y*y1 > 0: x1 = avg_x elif avg_y*y2 > 0: x2 = avg_x elif avg_y == 0: return avg_x elif abs(x1-x2) < ap: return x1 else: return "" # tan line method def tanline(func, dfunc, x, ap=0.0001): list(x).sort() x2 = x.pop() x1 = x.pop() y1 = func(x1) y2 = func(x2) if y1 == 0: return x1 if y2 == 0: return x2 if y1*y2 > 0: return "" mid_y1 = func((x1+x2)/2) mid_y2 = (y1+y2)/2 if mid_y1 < mid_y2: convex = -1 else: convex = 1 if y1*convex > 0: delta_x = x1 elif y2*convex > 0: delta_x = x2 while True: delta_y = func(delta_x) if abs(delta_y) < ap: return delta_x delta_x = delta_x - func(delta_x)/dfunc(delta_x) if __name__ == '__main__': print(binary_seperation(test_func,[0,1] )) print(tanline(test_func, df_test_func, [0,1]))
1,546
665
import unittest from rdflib import Graph def buildQueryArgs(q): return dict(select="", where="", optional="") class SPARQLParserTest(unittest.TestCase): known_issue = True def setUp(self): self.graph = Graph() pass def tearDown(self): pass tests = [ ("basic", """\ SELECT ?name WHERE { ?a <http://xmlns.com/foaf/0.1/name> ?name }"""), ("simple_prefix", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> SELECT ?name WHERE { ?a foaf:name ?name }"""), ("base_statement", """\ BASE <http://xmlns.com/foaf/0.1/> SELECT ?name WHERE { ?a <name> ?name }"""), ("prefix_and_colon_only_prefix", """\ PREFIX : <http://xmlns.com/foaf/0.1/> PREFIX vcard: <http://www.w3.org/2001/vcard-rdf/3.0#> SELECT ?name ?title WHERE { ?a :name ?name . ?a vcard:TITLE ?title }"""), ("predicate_object_list_notation", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> SELECT ?name ?mbox WHERE { ?x foaf:name ?name ; foaf:mbox ?mbox . }"""), ("object_list_notation", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> SELECT ?x WHERE { ?x foaf:nick "Alice" , "Alice_" . } """), ("escaped_literals", """\ PREFIX tag: <http://xmlns.com/foaf/0.1/> PREFIX vcard: <http://www.w3.org/2001/vcard-rdf/3.0#> SELECT ?name WHERE { ?a tag:name ?name ; vcard:TITLE "escape test vcard:TITLE " ; <tag://test/escaping> "This is a ''' Test \"\"\"" ; <tag://test/escaping> ?d } """), ("key_word_as_variable", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> SELECT ?PREFIX ?WHERE WHERE { ?x foaf:name ?PREFIX ; foaf:mbox ?WHERE . }"""), ("key_word_as_prefix", """\ PREFIX WHERE: <http://xmlns.com/foaf/0.1/> SELECT ?name ?mbox WHERE { ?x WHERE:name ?name ; WHERE:mbox ?mbox . }"""), ("some_test_cases_from_grammar_py_1", """\ SELECT ?title WHERE { <http://example.org/book/book1> <http://purl.org/dc/elements/1.1/title> ?title . }"""), ("some_test_cases_from_grammar_py_2", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> SELECT ?name ?mbox WHERE { ?person foaf:name ?name . OPTIONAL { ?person foaf:mbox ?mbox} }"""), ("some_test_cases_from_grammar_py_3", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> SELECT ?name ?name2 WHERE { ?person foaf:name ?name . OPTIONAL { ?person foaf:knows ?p2 . ?p2 foaf:name ?name2 . } }"""), ("some_test_cases_from_grammar_py_4", """\ PREFIX foaf: <http://xmlns.com/foaf/0.1/> #PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> SELECT ?name ?mbox WHERE { { ?person rdf:type foaf:Person } . OPTIONAL { ?person foaf:name ?name } . OPTIONAL {?person foaf:mbox ?mbox} . }""") ] def _buildQueryArg(q): res = buildQueryArgs(q) if res.get('select', False): assert res["select"] is not None if res.get('where', False): assert res["where"] is not None if res.get('optional', False): assert res["optional"] is not None # result = sparqlGr.query(select, where, optional) # self.assert_(self.graph.query(q) is not None)
3,428
1,335
class Frequency: def __init__(self): self.frequency = 0 def increment(self, i): self.frequency = self.frequency + i def decrement(self, i): self.frequency = self.frequency - i def __str__(self): return str(self.frequency) def __repr__(self): return str(self.frequency) def get_steps(filename): with open(f'data/{filename}', 'r') as f: raw_steps = f.readlines() steps = [] for i in raw_steps: steps.append([i[0], int(i[1:])]) return steps def part_one(): freq = Frequency() steps = get_steps('day1-1.txt') ops = {'+': freq.increment, '-': freq.decrement} for i in steps: ops[i[0]](i[1]) return freq def part_two(): freq = Frequency() steps = get_steps('day1-1.txt') ops = {'+': freq.increment, '-': freq.decrement} current = 0 already_seen = [] while current not in already_seen: for i in steps: if current in already_seen: break already_seen.append(int(str(freq))) ops[i[0]](i[1]) current = int(str(freq)) return freq if __name__ == '__main__': print(f'Part 1: {part_one()}\nPart 2: {part_two()}')
1,228
435
#!/usr/bin/env python3 # -*- coding: future_fstrings -*- from db_sync_tool.utility import mode, system, output, helper from db_sync_tool.remote import client as remote_client from file_sync_tool.transfer import utility def transfer_files(): """ Transfering configured files between clients :return: """ if 'files' in system.config: for config in system.config['files']['config']: output.message( output.Subject.INFO, f'Starting rsync file transfer' ) if 'exclude' not in config: config['exclude'] = [] if mode.get_sync_mode() == mode.SyncMode.PROXY: # Proxy mode: Transferring from origin to local and from local to target utility.generate_temp_dir_name() helper.check_and_create_dump_dir(mode.Client.LOCAL, utility.temp_data_dir) synchronize( origin_path=config[mode.Client.ORIGIN], target_path=utility.temp_data_dir, exclude=config['exclude'], pseudo_client=mode.Client.ORIGIN ) synchronize( origin_path=f'{utility.temp_data_dir}/*', target_path=config[mode.Client.TARGET], exclude=config['exclude'], pseudo_client=mode.Client.TARGET ) utility.remove_temporary_dir() elif mode.get_sync_mode() == mode.SyncMode.SYNC_REMOTE: synchronize( origin_path=config[mode.Client.ORIGIN], target_path=config[mode.Client.TARGET], exclude=config['exclude'], client=mode.Client.ORIGIN, force_remote=True ) else: synchronize( origin_path=config[mode.Client.ORIGIN], target_path=config[mode.Client.TARGET], exclude=config['exclude'] ) else: f'{output.Subject.WARNING} No file sync configuration provided' def synchronize(origin_path, target_path, exclude, client=mode.Client.LOCAL, pseudo_client=None, force_remote=False): """ Using rsync command to synchronize files between systems :param origin_path: String :param target_path: String :param exclude: List :param client: String :param pseudo_client: String Client, which will be forced as remote client. Necessary for proxy transfer. :param force_remote: Boolean :return: """ _remote_client = None if force_remote: remote_client.load_ssh_client_origin() _origin_subject = f'{output.Subject.ORIGIN}{output.CliFormat.BLACK}[REMOTE]{output.CliFormat.ENDC} ' _target_subject = f'{output.Subject.TARGET}{output.CliFormat.BLACK}[REMOTE]{output.CliFormat.ENDC} ' elif mode.is_remote(mode.Client.ORIGIN) and pseudo_client != mode.Client.TARGET: _remote_client = mode.Client.ORIGIN _origin_subject = f'{output.Subject.ORIGIN}{output.CliFormat.BLACK}[REMOTE]{output.CliFormat.ENDC} ' _target_subject = f'{output.Subject.TARGET}{output.CliFormat.BLACK}[LOCAL]{output.CliFormat.ENDC} ' elif mode.is_remote(mode.Client.TARGET) and pseudo_client != mode.Client.ORIGIN: _remote_client = mode.Client.TARGET _origin_subject = f'{output.Subject.ORIGIN}{output.CliFormat.BLACK}[LOCAL]{output.CliFormat.ENDC} ' _target_subject = f'{output.Subject.TARGET}{output.CliFormat.BLACK}[REMOTE]{output.CliFormat.ENDC} ' elif not mode.is_remote(mode.Client.TARGET) and not mode.is_remote(mode.Client.ORIGIN): _origin_subject = f'{output.Subject.ORIGIN}{output.CliFormat.BLACK}[LOCAL]{output.CliFormat.ENDC} ' _target_subject = f'{output.Subject.TARGET}{output.CliFormat.BLACK}[LOCAL]{output.CliFormat.ENDC} ' _origin_name = helper.get_ssh_host_name(mode.Client.ORIGIN, True) if _remote_client == mode.Client.ORIGIN else '' _target_name = helper.get_ssh_host_name(mode.Client.TARGET, True) if _remote_client == mode.Client.TARGET else '' if not system.config['mute']: print( f'{_origin_subject}' f'{_origin_name}' f'{output.CliFormat.BLACK}{origin_path}{output.CliFormat.ENDC}' ) print( f'{_target_subject}' f'{_target_name}' f'{output.CliFormat.BLACK}{target_path}{output.CliFormat.ENDC}' ) _origin_user_host = utility.get_host(mode.Client.ORIGIN) if _remote_client == mode.Client.ORIGIN else '' _target_user_host = utility.get_host(mode.Client.TARGET) if _remote_client == mode.Client.TARGET else '' _output = mode.run_command( f'{utility.get_password_environment(_remote_client)}rsync {utility.get_options()} ' f'{utility.get_authorization(_remote_client)} {utility.get_excludes(exclude)}' f'{_origin_user_host}{origin_path} {_target_user_host}{target_path}', client, True ) utility.read_stats(_output)
5,122
1,587
# coding: utf-8 # # Training Set Analysis # The purpose of this notebook is to compute the kernel density estimate of the PDF between the judgement and each feature in a training set, in order to estimate how each feature is performing. # # # TODO # Modify features to use custom function score queries, and use ML to optimise features given the below plot (plot ideal relationship between judgement and features, and optimise to get as close as possible). # In[51]: import sys import numpy as np import scipy from scipy import stats import matplotlib.pylab as plt models_dir = "/Users/sullid/ONS/dp-search-service/src/main/resources/elastic.ltr/models" model_name = sys.argv[1] # In[52]: class TrainingData(object): """ Class to handle the loading of training sets """ def __init__(self, model_dir, model_name): self.model_dir = model_dir self.model_name = model_name self.data = {} self.load() def load(self): fname = "%s/%s/ons_train.txt" % (self.model_dir, self.model_name) with open(fname, 'r') as f: lines = f.readlines() qid_dict = {} # First collect the qids for line in lines: parts = line.split("\t") qid_part = parts[1] qid = int(qid_part.split(":")[1]) if (qid not in qid_dict): qid_dict[qid] = [] qid_dict[qid].append(line) # Process each line by the qid for qid in qid_dict.keys(): lines = qid_dict[qid] self.data[qid] = {} for line in lines: if (line.startswith("#")): continue parts = line.split("\t") if (len(parts) > 0): for part in parts: if ('#' in part): part = part[0:part.index("#")].strip() key = "J" val = 0.0 if (":" in part): key = part.split(":")[0] val = float(part.split(":")[-1]) else: val = float(part) if (key not in self.data[qid]): self.data[qid][key] = [] self.data[qid][key].append(val) def get(self, qid, item): return np.array(self.data[qid][item]) def qids(self): return self.data.keys() def keys(self, qid): return self.data[qid].keys() def min(self, qid, item): return min(self.get(qid, item)) def max(self, item): return max(self.get(qid, item)) def size(self, qid): return len(self.get(qid, "J")) def numFeatures(self, qid): return len(self.keys(qid)) - 2 # - 2 to account for judgement (J) and qid trainingData = TrainingData(models_dir, model_name) # In[56]: import matplotlib.gridspec as gridspec import matplotlib.pylab as pylab fs=25 params = {'legend.fontsize': 'x-large', 'figure.figsize': (15, 5), 'axes.labelsize': fs, 'axes.titlesize':fs, 'xtick.labelsize':fs, 'ytick.labelsize':fs} pylab.rcParams.update(params) def getValues(trainingData, qid, i): if (i == 0): return "Judgement", trainingData.get(qid, "J") # elif (i == 1 or i == 9): # return str(i), np.log10(trainingData[str(i)]) else: return "Feature %d" % i, trainingData.get(qid, str(i)) def fitKernel(x,y,n=100j): xmin,xmax,ymin,ymax=(x.min(),x.max(),y.min(),y.max()) X, Y = np.mgrid[xmin:xmax:n, ymin:ymax:n] positions = np.vstack([X.ravel(), Y.ravel()]) values = np.vstack([x, y]) kernel = stats.gaussian_kde(values) Z = np.reshape(kernel(positions).T, X.shape) return np.rot90(Z), (xmin, xmax, ymin, ymax) def forceAspect(ax,aspect=1): im = ax.get_images() extent = im[0].get_extent() ax.set_aspect(abs((extent[1]-extent[0])/(extent[3]-extent[2]))/aspect) print "QIDS: ", trainingData.qids() for qid in trainingData.qids(): numFeatures = trainingData.numFeatures(qid) + 1 fig = plt.figure(figsize=(50, 50)) plt.suptitle('qid:%d' % qid, fontsize=fs*1.5) gs = gridspec.GridSpec(numFeatures, numFeatures) for i in range(numFeatures): rowLabel, rowValues = getValues(trainingData, qid, i) labelRow = True for j in range(i+1, numFeatures): colLabel, colValues = getValues(trainingData, qid, j) ax = plt.subplot(gs[i,j-numFeatures]) # ax.text(0.25, 0.5, "ax %d:%d" % (i,j)) if (labelRow): ax.set_ylabel(rowLabel) labelRow = False if (j == (i+1)): ax.set_xlabel(colLabel) try: Z, (xmin,xmax,ymin,ymax) = fitKernel(colValues, rowValues, 200j) extent = [xmin,xmax,ymin,ymax] ax.imshow(Z, cmap=plt.cm.gist_stern_r, extent=extent) ax.set_xlim([xmin, xmax]) ax.set_ylim([ymin, ymax]) forceAspect(ax) except: pass # ax.imshow(np.rot90(Z), cmap=plt.cm.gist_earth_r, # extent=[xmin, xmax, ymin, ymax], aspect=50) # ax.plot(x, y, 'k.', markersize=2) # ax.set_xlim([xmin, xmax]) # ax.set_ylim([ymin, ymax]) plt.show()
5,635
1,911
from operator import attrgetter, itemgetter from typing import ( Sequence, Any, Tuple, Union, List, Iterable, Dict, Callable, NamedTuple, ) SortSpec = NamedTuple("SortSpec", [("sort_key", Union[str, int]), ("reversed", bool)]) def optional_collection(argument, collection_factory): if argument is None: return collection_factory() return argument def sort_in_place( xs: List[Union[Sequence, Any]], specs: Sequence[SortSpec], use_get_item: bool, ): if not hasattr(xs, "sort"): raise AttributeError( "Sortable must be a mutable type with a sort function, e.g. List" ) if use_get_item: for key, reverse in reversed(specs): xs.sort(key=itemgetter(key), reverse=reverse) return xs for key, reverse in reversed(specs): xs.sort(key=attrgetter(key), reverse=reverse) # type: ignore return xs def sort_to_new_list( xs: Iterable[Union[Sequence, Any]], specs: Sequence[Tuple[Union[str, int], bool]], use_get_item: bool, ): if use_get_item: sorted_list = [] for index, spec in enumerate(reversed(specs)): key, reverse = spec if index == 0: sorted_list = sorted(xs, key=itemgetter(key), reverse=reverse) sorted_list = sorted(sorted_list, key=itemgetter(key), reverse=reverse) return sorted_list sorted_list = [] for index, spec in enumerate(reversed(specs)): key, reverse = spec if index == 0: sorted_list = sorted(xs, key=itemgetter(key), reverse=reverse) sorted_list = sorted(sorted_list, key=attrgetter(key), reverse=reverse) # type: ignore return sorted_list def index_list_of_dicts( data: Sequence[Dict[str, Any]], key_field: str ) -> Dict[str, Dict[str, Any]]: # TODO save to utilities result = {} for item in data: key_field_value = item[key_field] # will error if field not found result[str(key_field_value)] = item return result def index_list_of_objects( data: Iterable[Union[Sequence, Any]], key_field, use_get_item: bool, cast_index: Callable = None, ): """ Will index a list of objects based on key_field. Returns a dict with key based on key_field of object Parameters ---------- data : Iterable[Union[Sequence, Any]] [description] key : [type] [description] use_get_item : bool [description] cast_index : Callable, optional [description], by default None Returns ------- [type] [description] """ if use_get_item: indexer = itemgetter(key_field) else: indexer = attrgetter(key_field) result = {} for item in data: key_field_value = indexer(item) if cast_index is not None: key_field_value = cast_index(key_field_value) result[key_field_value] = item return result def index_list_of_objects_multiple( data: Iterable[Union[Sequence, Any]], key_field, use_get_item: bool, cast_index: Callable = None, ) -> Dict[Any, List[Any]]: if use_get_item: indexer = itemgetter(key_field) else: indexer = attrgetter(key_field) result: Dict[Any, List[Any]] = {} for item in data: key_field_value = indexer(item) if cast_index is not None: key_field_value = cast_index(key_field_value) indexed_field = result.get(key_field_value, []) indexed_field.append(item) result[key_field_value] = indexed_field return result
3,623
1,158
from django.contrib import admin from models import Location, ItemTemplate, Log, Inventory, Supplier class ItemTemplateAdmin(admin.ModelAdmin): filter_horizontal = ('supplies', 'suppliers') admin.site.register(Location) admin.site.register(ItemTemplate, ItemTemplateAdmin) admin.site.register(Log) admin.site.register(Inventory) admin.site.register(Supplier)
368
106
from typing import Dict, Type from injector import Module, Binder from core_get.actions.action import Action from core_get.actions.init.init import Init from core_get.actions.init.init_options import InitOptions from core_get.actions.install.install import Install from core_get.actions.install.install_options import InstallOptions from core_get.actions.login.login import Login from core_get.actions.login.login_options import LoginOptions from core_get.actions.package.package import Package from core_get.actions.package.package_options import PackageOptions from core_get.actions.publish.publish import Publish from core_get.actions.publish.publish_options import PublishOptions from core_get.actions.remove.remove import Remove from core_get.actions.remove.remove_options import RemoveOptions from core_get.actions.test.test import Test from core_get.actions.test.test_options import TestOptions from core_get.actions.yank.yank import Yank from core_get.actions.yank.yank_options import YankOptions from core_get.options.options import Options from core_get.utils.injection import MultiMapClassProvider class ActionsModule(Module): def configure(self, binder: Binder) -> None: binder.multibind(Dict[Type[Options], Action], to=MultiMapClassProvider({ InitOptions: Init, InstallOptions: Install, LoginOptions: Login, PackageOptions: Package, PublishOptions: Publish, RemoveOptions: Remove, TestOptions: Test, YankOptions: Yank, }))
1,557
408
from panda3d.core import * # Leave these imports in, they may be used by ptf files. from panda3d.physics import * from . import Particles from . import ForceGroup from direct.directnotify import DirectNotifyGlobal class ParticleEffect(NodePath): notify = DirectNotifyGlobal.directNotify.newCategory('ParticleEffect') pid = 1 def __init__(self, name=None, particles=None): if name is None: name = 'particle-effect-%d' % ParticleEffect.pid ParticleEffect.pid += 1 NodePath.__init__(self, name) # Record particle effect name self.name = name # Enabled flag self.fEnabled = 0 # Dictionary of particles and forceGroups self.particlesDict = {} self.forceGroupDict = {} # The effect's particle system if particles is not None: self.addParticles(particles) self.renderParent = None def cleanup(self): self.removeNode() self.disable() if self.__isValid(): for f in self.forceGroupDict.values(): f.cleanup() for p in self.particlesDict.values(): p.cleanup() del self.forceGroupDict del self.particlesDict del self.renderParent def getName(self): # override NodePath.getName() return self.name def reset(self): self.removeAllForces() self.removeAllParticles() self.forceGroupDict = {} self.particlesDict = {} def start(self, parent=None, renderParent=None): assert self.notify.debug('start() - name: %s' % self.name) self.renderParent = renderParent self.enable() if parent is not None: self.reparentTo(parent) def enable(self): # band-aid added for client crash - grw if self.__isValid(): if self.renderParent: for p in self.particlesDict.values(): p.setRenderParent(self.renderParent.node()) for f in self.forceGroupDict.values(): f.enable() for p in self.particlesDict.values(): p.enable() self.fEnabled = 1 def disable(self): self.detachNode() # band-aid added for client crash - grw if self.__isValid(): for p in self.particlesDict.values(): p.setRenderParent(p.node) for f in self.forceGroupDict.values(): f.disable() for p in self.particlesDict.values(): p.disable() self.fEnabled = 0 def isEnabled(self): """ Note: this may be misleading if enable(), disable() not used """ return self.fEnabled def addForceGroup(self, forceGroup): forceGroup.nodePath.reparentTo(self) forceGroup.particleEffect = self self.forceGroupDict[forceGroup.getName()] = forceGroup # Associate the force group with all particles for i in range(len(forceGroup)): self.addForce(forceGroup[i]) def addForce(self, force): for p in list(self.particlesDict.values()): p.addForce(force) def removeForceGroup(self, forceGroup): # Remove forces from all particles for i in range(len(forceGroup)): self.removeForce(forceGroup[i]) forceGroup.nodePath.removeNode() forceGroup.particleEffect = None self.forceGroupDict.pop(forceGroup.getName(), None) def removeForce(self, force): for p in list(self.particlesDict.values()): p.removeForce(force) def removeAllForces(self): for fg in list(self.forceGroupDict.values()): self.removeForceGroup(fg) def addParticles(self, particles): particles.nodePath.reparentTo(self) self.particlesDict[particles.getName()] = particles # Associate all forces in all force groups with the particles for fg in list(self.forceGroupDict.values()): for i in range(len(fg)): particles.addForce(fg[i]) def removeParticles(self, particles): if particles is None: self.notify.warning('removeParticles() - particles == None!') return particles.nodePath.detachNode() self.particlesDict.pop(particles.getName(), None) # Remove all forces from the particles for fg in list(self.forceGroupDict.values()): for f in fg: particles.removeForce(f) def removeAllParticles(self): for p in list(self.particlesDict.values()): self.removeParticles(p) def getParticlesList(self): return list(self.particlesDict.values()) def getParticlesNamed(self, name): return self.particlesDict.get(name, None) def getParticlesDict(self): return self.particlesDict def getForceGroupList(self): return list(self.forceGroupDict.values()) def getForceGroupNamed(self, name): return self.forceGroupDict.get(name, None) def getForceGroupDict(self): return self.forceGroupDict def saveConfig(self, filename): filename = Filename(filename) with open(filename.toOsSpecific(), 'w') as f: # Add a blank line f.write('\n') # Make sure we start with a clean slate f.write('self.reset()\n') pos = self.getPos() hpr = self.getHpr() scale = self.getScale() f.write('self.setPos(%0.3f, %0.3f, %0.3f)\n' % (pos[0], pos[1], pos[2])) f.write('self.setHpr(%0.3f, %0.3f, %0.3f)\n' % (hpr[0], hpr[1], hpr[2])) f.write('self.setScale(%0.3f, %0.3f, %0.3f)\n' % (scale[0], scale[1], scale[2])) # Save all the particles to file num = 0 for p in list(self.particlesDict.values()): target = 'p%d' % num num = num + 1 f.write(target + ' = Particles.Particles(\'%s\')\n' % p.getName()) p.printParams(f, target) f.write('self.addParticles(%s)\n' % target) # Save all the forces to file num = 0 for fg in list(self.forceGroupDict.values()): target = 'f%d' % num num = num + 1 f.write(target + ' = ForceGroup.ForceGroup(\'%s\')\n' % \ fg.getName()) fg.printParams(f, target) f.write('self.addForceGroup(%s)\n' % target) def loadConfig(self, filename): vfs = VirtualFileSystem.getGlobalPtr() data = vfs.readFile(filename, 1) data = data.replace(b'\r', b'') try: exec(data) except: self.notify.warning('loadConfig: failed to load particle file: '+ repr(filename)) raise def accelerate(self,time,stepCount = 1,stepTime=0.0): for particles in self.getParticlesList(): particles.accelerate(time,stepCount,stepTime) def clearToInitial(self): for particles in self.getParticlesList(): particles.clearToInitial() def softStop(self): for particles in self.getParticlesList(): particles.softStop() def softStart(self, firstBirthDelay=None): if self.__isValid(): for particles in self.getParticlesList(): if firstBirthDelay is not None: particles.softStart(br=-1, first_birth_delay=firstBirthDelay) else: particles.softStart() else: # Not asserting here since we want to crash live clients for more expedient bugfix # (Sorry, live clients) self.notify.error('Trying to start effect(%s) after cleanup.' % (self.getName(),)) def __isValid(self): return hasattr(self, 'forceGroupDict') and \ hasattr(self, 'particlesDict') # Snake-case aliases. is_enabled = isEnabled add_force_group = addForceGroup add_force = addForce remove_force_group = removeForceGroup remove_force = removeForce remove_all_forces = removeAllForces add_particles = addParticles remove_particles = removeParticles remove_all_particles = removeAllParticles get_particles_list = getParticlesList get_particles_named = getParticlesNamed get_particles_dict = getParticlesDict get_force_group_list = getForceGroupList get_force_group_named = getForceGroupNamed get_force_group_dict = getForceGroupDict save_config = saveConfig load_config = loadConfig clear_to_initial = clearToInitial soft_stop = softStop soft_start = softStart
8,773
2,542
from django.core.exceptions import ValidationError from django.db import models from django.utils.functional import cached_property from django.utils.translation import ugettext_lazy as _ from jsonfield import JSONField from orchestra.models.fields import PrivateFileField from orchestra.models.queryset import group_by from . import settings from .methods import PaymentMethod class PaymentSourcesQueryset(models.QuerySet): def get_default(self): return self.filter(is_active=True).first() class PaymentSource(models.Model): account = models.ForeignKey('accounts.Account', verbose_name=_("account"), related_name='paymentsources') method = models.CharField(_("method"), max_length=32, choices=PaymentMethod.get_choices()) data = JSONField(_("data"), default={}) is_active = models.BooleanField(_("active"), default=True) objects = PaymentSourcesQueryset.as_manager() def __str__(self): return "%s (%s)" % (self.label, self.method_class.verbose_name) @cached_property def method_class(self): return PaymentMethod.get(self.method) @cached_property def method_instance(self): """ Per request lived method_instance """ return self.method_class(self) @cached_property def label(self): return self.method_instance.get_label() @cached_property def number(self): return self.method_instance.get_number() def get_bill_context(self): method = self.method_instance return { 'message': method.get_bill_message(), } def get_due_delta(self): return self.method_instance.due_delta def clean(self): self.data = self.method_instance.clean_data() class TransactionQuerySet(models.QuerySet): group_by = group_by def create(self, **kwargs): source = kwargs.get('source') if source is None or not hasattr(source.method_class, 'process'): # Manual payments don't need processing kwargs['state'] = self.model.WAITTING_EXECUTION amount = kwargs.get('amount') if amount == 0: kwargs['state'] = self.model.SECURED return super(TransactionQuerySet, self).create(**kwargs) def secured(self): return self.filter(state=Transaction.SECURED) def exclude_rejected(self): return self.exclude(state=Transaction.REJECTED) def amount(self): return next(iter(self.aggregate(models.Sum('amount')).values())) or 0 def processing(self): return self.filter(state__in=[Transaction.EXECUTED, Transaction.WAITTING_EXECUTION]) class Transaction(models.Model): WAITTING_PROCESSING = 'WAITTING_PROCESSING' # CREATED WAITTING_EXECUTION = 'WAITTING_EXECUTION' # PROCESSED EXECUTED = 'EXECUTED' SECURED = 'SECURED' REJECTED = 'REJECTED' STATES = ( (WAITTING_PROCESSING, _("Waitting processing")), (WAITTING_EXECUTION, _("Waitting execution")), (EXECUTED, _("Executed")), (SECURED, _("Secured")), (REJECTED, _("Rejected")), ) STATE_HELP = { WAITTING_PROCESSING: _("The transaction is created and requires processing by the " "specific payment method."), WAITTING_EXECUTION: _("The transaction is processed and its pending execution on " "the related financial institution."), EXECUTED: _("The transaction is executed on the financial institution."), SECURED: _("The transaction ammount is secured."), REJECTED: _("The transaction has failed and the ammount is lost, a new transaction " "should be created for recharging."), } bill = models.ForeignKey('bills.bill', verbose_name=_("bill"), related_name='transactions') source = models.ForeignKey(PaymentSource, null=True, blank=True, on_delete=models.SET_NULL, verbose_name=_("source"), related_name='transactions') process = models.ForeignKey('payments.TransactionProcess', null=True, blank=True, on_delete=models.SET_NULL, verbose_name=_("process"), related_name='transactions') state = models.CharField(_("state"), max_length=32, choices=STATES, default=WAITTING_PROCESSING) amount = models.DecimalField(_("amount"), max_digits=12, decimal_places=2) currency = models.CharField(max_length=10, default=settings.PAYMENT_CURRENCY) created_at = models.DateTimeField(_("created"), auto_now_add=True) modified_at = models.DateTimeField(_("modified"), auto_now=True) objects = TransactionQuerySet.as_manager() def __str__(self): return "#%i" % self.id @property def account(self): return self.bill.account def clean(self): if not self.pk: amount = self.bill.transactions.exclude(state=self.REJECTED).amount() if amount >= self.bill.total: raise ValidationError( _("Bill %(number)s already has valid transactions that cover bill total amount (%(amount)s).") % { 'number': self.bill.number, 'amount': amount, } ) def get_state_help(self): if self.source: return self.source.method_instance.state_help.get(self.state) or self.STATE_HELP.get(self.state) return self.STATE_HELP.get(self.state) def mark_as_processed(self): self.state = self.WAITTING_EXECUTION self.save(update_fields=('state', 'modified_at')) def mark_as_executed(self): self.state = self.EXECUTED self.save(update_fields=('state', 'modified_at')) def mark_as_secured(self): self.state = self.SECURED self.save(update_fields=('state', 'modified_at')) def mark_as_rejected(self): self.state = self.REJECTED self.save(update_fields=('state', 'modified_at')) class TransactionProcess(models.Model): """ Stores arbitrary data generated by payment methods while processing transactions """ CREATED = 'CREATED' EXECUTED = 'EXECUTED' ABORTED = 'ABORTED' COMMITED = 'COMMITED' STATES = ( (CREATED, _("Created")), (EXECUTED, _("Executed")), (ABORTED, _("Aborted")), (COMMITED, _("Commited")), ) data = JSONField(_("data"), blank=True) file = PrivateFileField(_("file"), blank=True) state = models.CharField(_("state"), max_length=16, choices=STATES, default=CREATED) created_at = models.DateTimeField(_("created"), auto_now_add=True, db_index=True) updated_at = models.DateTimeField(_("updated"), auto_now=True) class Meta: verbose_name_plural = _("Transaction processes") def __str__(self): return '#%i' % self.id def mark_as_executed(self): self.state = self.EXECUTED for transaction in self.transactions.all(): transaction.mark_as_executed() self.save(update_fields=('state', 'updated_at')) def abort(self): self.state = self.ABORTED for transaction in self.transactions.all(): transaction.mark_as_rejected() self.save(update_fields=('state', 'updated_at')) def commit(self): self.state = self.COMMITED for transaction in self.transactions.processing(): transaction.mark_as_secured() self.save(update_fields=('state', 'updated_at'))
7,566
2,325
# -*- coding: utf-8 -*- ########## # IMPORT # ########## import plotly.offline as py import plotly.tools as tls tls.set_credentials_file(username='ducktypers', api_key='fd81wnx3lh') ######## # MAIN # ######## def draw_map(lon, lat, text, titre='NO TITLE'): """ Take 3 list as input, and the title of the map. """ py.plot({ # use `py.iplot` inside the ipython notebook "data": [{ 'type': 'scattergeo', 'locationmode': 'france', 'lon': lon, 'lat': lat, 'text': text, 'mode': 'markers', 'marker': dict( size = 8, opacity = 0.8, line = dict(width=1, color="rgb(102,102,102)") ) }], "layout": { 'title': str(titre), 'geo': dict(scope='europe', projection=dict( type='albers europe' ), showland = True, landcolor = "rgb(200, 200, 200)", subunitcolor = "rgb(217, 217, 217)", countrycolor = "rgb(217, 217, 217)", countrywidth = 1, subunitwidth = 1) } }, filename='interactive_map', # name of the file as saved in your plotly account #sharing='public' )
1,264
439
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: acp.proto """Generated protocol buffer code.""" from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() import custom_components.mbapi2020.proto.gogo_pb2 as gogo__pb2 DESCRIPTOR = _descriptor.FileDescriptor( name='acp.proto', package='proto', syntax='proto3', serialized_options=b'\n\032com.daimler.mbcarkit.proto\220\001\001\320\341\036\001', create_key=_descriptor._internal_create_key, serialized_pb=b'\n\tacp.proto\x12\x05proto\x1a\ngogo.proto\"\xa5\x02\n\x03VVA\"v\n\x0c\x43ommandState\x12\x19\n\x15UNKNOWN_COMMAND_STATE\x10\x00\x12\x0c\n\x07\x43REATED\x10\xf2\x07\x12\r\n\x08\x45NQUEUED\x10\xf8\x07\x12\x0f\n\nPROCESSING\x10\xf4\x07\x12\x0e\n\tSUSPENDED\x10\xf9\x07\x12\r\n\x08\x46INISHED\x10\xfa\x07\"\xa5\x01\n\x10\x43ommandCondition\x12\x1d\n\x19UNKNWON_COMMAND_CONDITION\x10\x00\x12\t\n\x04NONE\x10\xe8\x07\x12\r\n\x08\x41\x43\x43\x45PTED\x10\xe9\x07\x12\r\n\x08REJECTED\x10\xea\x07\x12\x0e\n\tTERMINATE\x10\xeb\x07\x12\x0c\n\x07SUCCESS\x10\xf3\x07\x12\x0b\n\x06\x46\x41ILED\x10\xf5\x07\x12\x10\n\x0bOVERWRITTEN\x10\xf6\x07\x12\x0c\n\x07TIMEOUT\x10\xf7\x07\"\x8f\x0b\n\nVehicleAPI\"~\n\x0c\x43ommandState\x12\x19\n\x15UNKNOWN_COMMAND_STATE\x10\x00\x12\x0e\n\nINITIATION\x10\x01\x12\x0c\n\x08\x45NQUEUED\x10\x02\x12\x0e\n\nPROCESSING\x10\x03\x12\x0b\n\x07WAITING\x10\x04\x12\x0c\n\x08\x46INISHED\x10\x05\x12\n\n\x06\x46\x41ILED\x10\x06\"S\n\x0f\x41ttributeStatus\x12\r\n\tVALUE_SET\x10\x00\x12\x11\n\rVALUE_NOT_SET\x10\x01\x12\x0b\n\x07INVALID\x10\x03\x12\x11\n\rNOT_AVAILABLE\x10\x04\"\xab\t\n\tQueueType\x12\x1b\n\x17UNKNOWNCOMMANDQUEUETYPE\x10\x00\x12\t\n\x05\x44OORS\x10\n\x12\x0b\n\x07\x41UXHEAT\x10\x0b\x12\x0b\n\x07PRECOND\x10\x0c\x12\r\n\tCHARGEOPT\x10\r\x12\x0f\n\x0bMAINTENANCE\x10\x0e\x12\x07\n\x03TCU\x10\x0f\x12\x08\n\x04\x46\x45\x45\x44\x10\x10\x12\x15\n\x11SERVICEACTIVATION\x10\x11\x12\x07\n\x03\x41TP\x10\x12\x12\x0e\n\nASSISTANCE\x10\x13\x12\x08\n\x04RACP\x10\x14\x12\x0f\n\x0bWEEKPROFILE\x10\x15\x12\x13\n\x0fREMOTEDIAGNOSIS\x10\x16\x12\x08\n\x04\x46LSH\x10\x17\x12\x0f\n\x0bTEMPERATURE\x10\x18\x12\x0c\n\x08TRIPCOMP\x10\x19\x12\n\n\x06\x45NGINE\x10\x1a\x12\x0e\n\nTHEFTALARM\x10\x1b\x12\n\n\x06WINDOW\x10\x1c\x12\x0c\n\x08HEADUNIT\x10\x1d\x12\n\n\x06MECALL\x10\x1f\x12\x0f\n\x0bIMMOBILIZER\x10 \x12\x10\n\x0cRENTALSIGNAL\x10!\x12\x07\n\x03\x42\x43\x46\x10\"\x12\x11\n\rPLUGANDCHARGE\x10#\x12\x14\n\x10\x43\x41RSHARINGMODULE\x10$\x12\x0b\n\x07\x42\x41TTERY\x10%\x12\x11\n\rONBOARDFENCES\x10&\x12\x0f\n\x0bSPEEDFENCES\x10\'\x12\x13\n\x0f\x43HARGINGTARIFFS\x10(\x12\r\n\tRTMCONFIG\x10)\x12\x17\n\x13MAINTENANCECOMPUTER\x10*\x12\x0b\n\x07MECALL2\x10+\x12\x19\n\x15\x41UTOMATEDVALETPARKING\x10,\x12\x11\n\rCHARGECONTROL\x10-\x12\x0e\n\nSPEEDALERT\x10.\x12\x1b\n\x17unknowncommandqueuetype\x10\x00\x12\t\n\x05\x64oors\x10\n\x12\x0b\n\x07\x61uxheat\x10\x0b\x12\x0b\n\x07precond\x10\x0c\x12\r\n\tchargeopt\x10\r\x12\x0f\n\x0bmaintenance\x10\x0e\x12\x07\n\x03tcu\x10\x0f\x12\x08\n\x04\x66\x65\x65\x64\x10\x10\x12\x15\n\x11serviceactivation\x10\x11\x12\x07\n\x03\x61tp\x10\x12\x12\x0e\n\nassistance\x10\x13\x12\x08\n\x04racp\x10\x14\x12\x0f\n\x0bweekprofile\x10\x15\x12\x13\n\x0fremotediagnosis\x10\x16\x12\x08\n\x04\x66lsh\x10\x17\x12\x0f\n\x0btemperature\x10\x18\x12\x0c\n\x08tripcomp\x10\x19\x12\n\n\x06\x65ngine\x10\x1a\x12\x0e\n\ntheftalarm\x10\x1b\x12\n\n\x06window\x10\x1c\x12\x0c\n\x08headunit\x10\x1d\x12\n\n\x06mecall\x10\x1f\x12\x0f\n\x0bimmobilizer\x10 \x12\x10\n\x0crentalsignal\x10!\x12\x07\n\x03\x62\x63\x66\x10\"\x12\x11\n\rplugandcharge\x10#\x12\x14\n\x10\x63\x61rsharingmodule\x10$\x12\x0b\n\x07\x62\x61ttery\x10%\x12\x11\n\ronboardfences\x10&\x12\x0f\n\x0bspeedfences\x10\'\x12\x13\n\x0f\x63hargingtariffs\x10(\x12\r\n\trtmconfig\x10)\x12\x17\n\x13maintenancecomputer\x10*\x12\x0b\n\x07mecall2\x10+\x12\x19\n\x15\x61utomatedvaletparking\x10,\x12\x11\n\rchargecontrol\x10-\x12\x0e\n\nspeedalert\x10.\x1a\x02\x10\x01\"\xa9\x31\n\x03\x41\x43P\"\xa1\x31\n\x0b\x43ommandType\x12\x16\n\x12UNKNOWNCOMMANDTYPE\x10\x00\x12\r\n\tDOORSLOCK\x10\x64\x12\x0f\n\x0b\x44OORSUNLOCK\x10n\x12\x0f\n\x0bTRUNKUNLOCK\x10s\x12\x12\n\x0e\x46UELFLAPUNLOCK\x10t\x12\x14\n\x10\x43HARGEFLAPUNLOCK\x10u\x12\x17\n\x13\x43HARGECOUPLERUNLOCK\x10v\x12\x16\n\x12\x44OORSPREPARERENTAL\x10x\x12\x17\n\x12\x44OORSSECUREVEHICLE\x10\x82\x01\x12\x11\n\x0c\x41UXHEATSTART\x10\xac\x02\x12\x10\n\x0b\x41UXHEATSTOP\x10\xb6\x02\x12\x15\n\x10\x41UXHEATCONFIGURE\x10\xc0\x02\x12\x19\n\x14TEMPERATURECONFIGURE\x10\xde\x02\x12\x19\n\x14WEEKPROFILECONFIGURE\x10\xe8\x02\x12\x1b\n\x16WEEKPROFILEV2CONFIGURE\x10\xf2\x02\x12\x11\n\x0cPRECONDSTART\x10\x90\x03\x12\x10\n\x0bPRECONDSTOP\x10\x9a\x03\x12\x15\n\x10PRECONDCONFIGURE\x10\xa4\x03\x12\x1a\n\x15PRECONDCONFIGURESEATS\x10\xa9\x03\x12\x17\n\x12\x43HARGEOPTCONFIGURE\x10\xae\x03\x12\x13\n\x0e\x43HARGEOPTSTART\x10\xb8\x03\x12\x12\n\rCHARGEOPTSTOP\x10\xc2\x03\x12\x0c\n\x07\x46\x45\x45\x44POI\x10\xf4\x03\x12\x11\n\x0c\x46\x45\x45\x44\x46REETEXT\x10\xfe\x03\x12\x10\n\x0b\x45NGINESTART\x10\xa6\x04\x12\x0f\n\nENGINESTOP\x10\xb0\x04\x12\x13\n\x0e\x45NGINEAVPSTART\x10\xba\x04\x12\x0e\n\tTCUWAKEUP\x10\xd8\x04\x12\x10\n\x0bTCUSWUPDATE\x10\xe2\x04\x12\x10\n\x0bTCURCSRESET\x10\xec\x04\x12\x15\n\x10TCUINTERROGATION\x10\xf6\x04\x12\x14\n\x0fSPEEDALERTSTART\x10\xc6\x05\x12\x13\n\x0eSPEEDALERTSTOP\x10\xd0\x05\x12\x0e\n\tFLSHSTART\x10\xee\x05\x12\r\n\x08\x46LSHSTOP\x10\xf8\x05\x12\x10\n\x0bSIGPOSSTART\x10\x82\x06\x12\x16\n\x11\x43ONTRACTCONFIGURE\x10\xa0\x06\x12\x13\n\x0e\x43ONTRACTREMOVE\x10\xaa\x06\x12\x12\n\rROOTCONFIGURE\x10\xb4\x06\x12\x0f\n\nROOTREMOVE\x10\xbe\x06\x12\r\n\x08TRIPCOMP\x10\xd2\x06\x12\x19\n\x14MAINTENANCECONFIGURE\x10\xa2\x07\x12\x1e\n\x19MAINTENANCECOMPUTEROFFSET\x10\xa3\x07\x12\x15\n\x10SHORTTESTEXECUTE\x10\xa7\x07\x12\x1f\n\x1aSERVICEACTIVATIONCONFIGURE\x10\xac\x07\x12\"\n\x1d\x44\x43\x32SERVICEACTIVATIONCONFIGURE\x10\xb1\x07\x12\x13\n\x0e\x44\x43\x32RAWDOWNLOAD\x10\xb6\x07\x12\x1d\n\x18\x41PPLICATIONCONFIGURATION\x10\xbb\x07\x12\x15\n\x10\x44\x43\x32STARTTRACKING\x10\xc0\x07\x12\x10\n\x0b\x41TPSEQUENCE\x10\xde\x07\x12\x1d\n\x18THEFTALARMTOGGLEINTERIOR\x10\xe8\x07\x12\x18\n\x13THEFTALARMTOGGLETOW\x10\xf2\x07\x12 \n\x1bTHEFTALARMSELECTINTERIORTOW\x10\xfc\x07\x12\"\n\x1dTHEFTALARMDESELECTINTERIORTOW\x10\x86\x08\x12\x13\n\x0eTHEFTALARMSTOP\x10\x90\x08\x12\x0f\n\nWINDOWOPEN\x10\xcc\x08\x12\x10\n\x0bWINDOWCLOSE\x10\xd6\x08\x12\x14\n\x0fWINDOWVENTILATE\x10\xe0\x08\x12\x0f\n\nWINDOWMOVE\x10\xe1\x08\x12\r\n\x08ROOFOPEN\x10\xea\x08\x12\x0e\n\tROOFCLOSE\x10\xf4\x08\x12\r\n\x08ROOFLIFT\x10\xfe\x08\x12\r\n\x08ROOFMOVE\x10\xff\x08\x12\x12\n\rBATTERYMAXSOC\x10\xd0\x0f\x12\x19\n\x14\x42\x41TTERYCHARGEPROGRAM\x10\xda\x0f\x12\x1b\n\x16\x43HARGEPROGRAMCONFIGURE\x10\xe4\x0f\x12\x18\n\x13ONBOARDFENCESCREATE\x10\xb4\x10\x12\x18\n\x13ONBOARDFENCESUPDATE\x10\xbe\x10\x12\x18\n\x13ONBOARDFENCESDELETE\x10\xc8\x10\x12\x16\n\x11SPEEDFENCESCREATE\x10\x98\x11\x12\x16\n\x11SPEEDFENCESUPDATE\x10\xa2\x11\x12\x16\n\x11SPEEDFENCESDELETE\x10\xac\x11\x12\x1a\n\x15\x43HARGINGTARIFFSCREATE\x10\xfc\x11\x12\x1a\n\x15\x43HARGINGTARIFFSUPDATE\x10\x86\x12\x12\x1a\n\x15\x43HARGINGTARIFFSDELETE\x10\x90\x12\x12\x14\n\x0fTHEFTALARMSTART\x10\xc4\x13\x12\x1d\n\x18THEFTALARMSELECTINTERIOR\x10\xce\x13\x12\x1f\n\x1aTHEFTALARMDESELECTINTERIOR\x10\xd8\x13\x12\x18\n\x13THEFTALARMSELECTTOW\x10\xe2\x13\x12\x1a\n\x15THEFTALARMDESELECTTOW\x10\xec\x13\x12$\n\x1fTHEFTALARMSELECTDAMAGEDETECTION\x10\xf6\x13\x12&\n!THEFTALARMDESELECTDAMAGEDETECTION\x10\x80\x14\x12%\n THEFTALARMCONFIRMDAMAGEDETECTION\x10\x8a\x14\x12\x11\n\x0cMECALL2START\x10\xa8\x14\x12\x1e\n\x19UDXTRIGGERSYNCHRONIZATION\x10\xb0\t\x12\x19\n\x14UDXACTIVEUSERPROFILE\x10\xba\t\x12\x15\n\x10UDXRESETUSERDATA\x10\xc4\t\x12\x12\n\rUSERPROFSYNCH\x10\xce\t\x12\x12\n\rUSERDATARESET\x10\xd8\t\x12\x17\n\x12PROFACTIVATIONSNAP\x10\xe2\t\x12\x19\n\x14PROFACTIVATIONDIRECT\x10\xe7\t\x12\x13\n\x0eSOFTWAREUPDATE\x10\xec\t\x12\x15\n\x10PUSHNOTIFICATION\x10\xf6\t\x12\x12\n\rMECALLCOMMAND\x10\x9e\n\x12\x14\n\x0fPRECONDSTARTRCS\x10\xf8\n\x12\x13\n\x0ePRECONDSTOPRCS\x10\x82\x0b\x12\x18\n\x13PRECONDCONFIGURERCS\x10\x8c\x0b\x12\x11\n\x0cTCUCONFIGURE\x10\x96\x0b\x12\x1c\n\x17\x45\x44ISONSERVICEACTIVATION\x10\x97\x0b\x12\x11\n\x0cTESTSEQUENCE\x10\x98\x0b\x12\x19\n\x14PRECONDCONFIGURERACP\x10\x99\x0b\x12\x1b\n\x16\x43HARGEOPTCONFIGURERACP\x10\x9a\x0b\x12\x18\n\x13TARIFFTABLEDOWNLOAD\x10\x9b\x0b\x12\x15\n\x10PRECONDSTARTRACP\x10\x9c\x0b\x12\x14\n\x0fPRECONDSTOPRACP\x10\x9d\x0b\x12\x1a\n\x15ROOTCERTIFICATEREMOVE\x10\x9e\x0b\x12\x19\n\x14ONREQUESTPROBEUPLOAD\x10\x9f\x0b\x12\x1c\n\x17ROOTCERTIFICATEDOWNLOAD\x10\xa0\x0b\x12\x1e\n\x19\x43ONTRACTCERTIFICATEREMOVE\x10\xa1\x0b\x12 \n\x1b\x43ONTRACTCERTIFICATEDOWNLOAD\x10\xa2\x0b\x12\x1d\n\x18PROBECONFIGURATIONUPDATE\x10\xa3\x0b\x12\x13\n\x0eRDIAGDELETEECU\x10\xdc\x0b\x12\x16\n\x11RDIAGSTATUSREPORT\x10\xdd\x0b\x12\x13\n\x0eRDIAGEXECUTION\x10\xde\x0b\x12\x19\n\x14IMMOBILIZERCHALLENGE\x10\xc0\x0c\x12\x1d\n\x18IMMOBILIZERSEARCHKEYLINE\x10\xca\x0c\x12\x1e\n\x19IMMOBILIZERRELEASEKEYLINE\x10\xd4\x0c\x12\x1b\n\x16IMMOBILIZERLOCKKEYLINE\x10\xde\x0c\x12\x1b\n\x16IMMOBILIZERLOCKVEHICLE\x10\xdf\x0c\x12\x1e\n\x19IMMOBILIZERRELEASEVEHICLE\x10\xd5\x0c\x12\x14\n\x0fSETRENTALSIGNAL\x10\xa4\r\x12\x19\n\x14\x42LACKCHANNELDOWNLOAD\x10\x88\x0e\x12\x17\n\x12\x42LACKCHANNELUPLOAD\x10\x92\x0e\x12\x11\n\x0c\x43ONFIGURECSM\x10\xec\x0e\x12\x16\n\x11UPDATEVEHICLEINFO\x10\xed\x0e\x12\x16\n\x11RELAYMESSAGETOCSM\x10\xee\x0e\x12\x1c\n\x17RELAYRENTALREQUESTTOCSB\x10\xef\x0e\x12\x16\n\x11RTMDOWNLOADCONFIG\x10\xe0\x12\x12\x12\n\rRTMREADCONFIG\x10\xea\x12\x12\x10\n\x0b\x41VPACTIVATE\x10\x8c\x15\x12\x1b\n\x16\x43HARGECONTROLCONFIGURE\x10\xf0\x15\x12\x16\n\x12unknownCommandType\x10\x00\x12\r\n\tdoorsLock\x10\x64\x12\x0f\n\x0b\x64oorsUnlock\x10n\x12\x0f\n\x0btrunkUnlock\x10s\x12\x12\n\x0e\x66uelflapUnlock\x10t\x12\x14\n\x10\x63hargeflapUnlock\x10u\x12\x17\n\x13\x63hargecouplerUnlock\x10v\x12\x16\n\x12\x64oorsPrepareRental\x10x\x12\x17\n\x12\x64oorsSecureVehicle\x10\x82\x01\x12\x11\n\x0c\x61uxheatStart\x10\xac\x02\x12\x10\n\x0b\x61uxheatStop\x10\xb6\x02\x12\x15\n\x10\x61uxheatConfigure\x10\xc0\x02\x12\x19\n\x14temperatureConfigure\x10\xde\x02\x12\x19\n\x14weekprofileConfigure\x10\xe8\x02\x12\x1b\n\x16weekprofileV2Configure\x10\xf2\x02\x12\x11\n\x0cprecondStart\x10\x90\x03\x12\x10\n\x0bprecondStop\x10\x9a\x03\x12\x15\n\x10precondConfigure\x10\xa4\x03\x12\x1a\n\x15precondConfigureSeats\x10\xa9\x03\x12\x17\n\x12\x63hargeoptConfigure\x10\xae\x03\x12\x13\n\x0e\x63hargeoptStart\x10\xb8\x03\x12\x12\n\rchargeoptStop\x10\xc2\x03\x12\x0c\n\x07\x66\x65\x65\x64Poi\x10\xf4\x03\x12\x11\n\x0c\x66\x65\x65\x64\x46reetext\x10\xfe\x03\x12\x10\n\x0b\x65ngineStart\x10\xa6\x04\x12\x0f\n\nengineStop\x10\xb0\x04\x12\x13\n\x0e\x65ngineAvpstart\x10\xba\x04\x12\x0e\n\ttcuWakeup\x10\xd8\x04\x12\x10\n\x0btcuSwUpdate\x10\xe2\x04\x12\x10\n\x0btcuRcsReset\x10\xec\x04\x12\x15\n\x10tcuInterrogation\x10\xf6\x04\x12\x14\n\x0fspeedalertStart\x10\xc6\x05\x12\x13\n\x0espeedalertStop\x10\xd0\x05\x12\x0e\n\tflshStart\x10\xee\x05\x12\r\n\x08\x66lshStop\x10\xf8\x05\x12\x10\n\x0bsigposStart\x10\x82\x06\x12\x16\n\x11\x63ontractConfigure\x10\xa0\x06\x12\x13\n\x0e\x63ontractRemove\x10\xaa\x06\x12\x12\n\rrootConfigure\x10\xb4\x06\x12\x0f\n\nrootRemove\x10\xbe\x06\x12\r\n\x08tripcomp\x10\xd2\x06\x12\x19\n\x14maintenanceConfigure\x10\xa2\x07\x12\x1e\n\x19maintenanceComputerOffset\x10\xa3\x07\x12\x15\n\x10shorttestExecute\x10\xa7\x07\x12\x1f\n\x1aserviceactivationConfigure\x10\xac\x07\x12\"\n\x1d\x64\x63\x32ServiceactivationConfigure\x10\xb1\x07\x12\x13\n\x0e\x64\x63\x32RawDownload\x10\xb6\x07\x12\x1d\n\x18\x61pplicationConfiguration\x10\xbb\x07\x12\x15\n\x10\x64\x63\x32StartTracking\x10\xc0\x07\x12\x10\n\x0b\x61tpSequence\x10\xde\x07\x12\x1d\n\x18theftalarmToggleInterior\x10\xe8\x07\x12\x18\n\x13theftalarmToggleTow\x10\xf2\x07\x12 \n\x1btheftalarmSelectInteriorTow\x10\xfc\x07\x12\"\n\x1dtheftalarmDeselectInteriorTow\x10\x86\x08\x12\x13\n\x0etheftalarmStop\x10\x90\x08\x12\x0f\n\nwindowOpen\x10\xcc\x08\x12\x10\n\x0bwindowClose\x10\xd6\x08\x12\x14\n\x0fwindowVentilate\x10\xe0\x08\x12\x0f\n\nwindowMove\x10\xe1\x08\x12\r\n\x08roofOpen\x10\xea\x08\x12\x0e\n\troofClose\x10\xf4\x08\x12\r\n\x08roofLift\x10\xfe\x08\x12\r\n\x08roofMove\x10\xff\x08\x12\x12\n\rbatteryMaxsoc\x10\xd0\x0f\x12\x19\n\x14\x62\x61tteryChargeprogram\x10\xda\x0f\x12\x1b\n\x16\x63hargeprogramconfigure\x10\xe4\x0f\x12\x18\n\x13onboardfencesCreate\x10\xb4\x10\x12\x18\n\x13onboardfencesUpdate\x10\xbe\x10\x12\x18\n\x13onboardfencesDelete\x10\xc8\x10\x12\x16\n\x11speedfencesCreate\x10\x98\x11\x12\x16\n\x11speedfencesUpdate\x10\xa2\x11\x12\x16\n\x11speedfencesDelete\x10\xac\x11\x12\x1a\n\x15\x63hargingtariffsCreate\x10\xfc\x11\x12\x1a\n\x15\x63hargingtariffsUpdate\x10\x86\x12\x12\x1a\n\x15\x63hargingtariffsDelete\x10\x90\x12\x12\x14\n\x0ftheftalarmstart\x10\xc4\x13\x12\x1d\n\x18theftalarmselectinterior\x10\xce\x13\x12\x1f\n\x1atheftalarmdeselectinterior\x10\xd8\x13\x12\x18\n\x13theftalarmselecttow\x10\xe2\x13\x12\x1a\n\x15theftalarmdeselecttow\x10\xec\x13\x12$\n\x1ftheftalarmselectdamagedetection\x10\xf6\x13\x12&\n!theftalarmdeselectdamagedetection\x10\x80\x14\x12%\n theftalarmconfirmdamagedetection\x10\x8a\x14\x12\x11\n\x0cmecall2start\x10\xa8\x14\x12\x1e\n\x19udxTriggerSynchronization\x10\xb0\t\x12\x19\n\x14udxActiveUserProfile\x10\xba\t\x12\x15\n\x10udxResetUserData\x10\xc4\t\x12\x12\n\ruserProfSynch\x10\xce\t\x12\x12\n\ruserDataReset\x10\xd8\t\x12\x17\n\x12profActivationSnap\x10\xe2\t\x12\x19\n\x14profActivationDirect\x10\xe7\t\x12\x13\n\x0esoftwareUpdate\x10\xec\t\x12\x15\n\x10pushNotification\x10\xf6\t\x12\x12\n\rmecallcommand\x10\x9e\n\x12\x14\n\x0fprecondStartRcs\x10\xf8\n\x12\x13\n\x0eprecondStopRcs\x10\x82\x0b\x12\x18\n\x13precondConfigureRcs\x10\x8c\x0b\x12\x11\n\x0ctcuConfigure\x10\x96\x0b\x12\x1c\n\x17\x65\x64isonServiceActivation\x10\x97\x0b\x12\x11\n\x0ctestSequence\x10\x98\x0b\x12\x19\n\x14precondConfigureRacp\x10\x99\x0b\x12\x1b\n\x16\x63hargeoptConfigureRacp\x10\x9a\x0b\x12\x18\n\x13tariffTableDownload\x10\x9b\x0b\x12\x15\n\x10precondStartRacp\x10\x9c\x0b\x12\x14\n\x0fprecondStopRacp\x10\x9d\x0b\x12\x1a\n\x15rootCertificateRemove\x10\x9e\x0b\x12\x19\n\x14onRequestProbeUpload\x10\x9f\x0b\x12\x1c\n\x17rootCertificateDownload\x10\xa0\x0b\x12\x1e\n\x19\x63ontractCertificateRemove\x10\xa1\x0b\x12 \n\x1b\x63ontractCertificateDownload\x10\xa2\x0b\x12\x1d\n\x18probeConfigurationUpdate\x10\xa3\x0b\x12\x13\n\x0erdiagDeleteEcu\x10\xdc\x0b\x12\x16\n\x11rdiagStatusReport\x10\xdd\x0b\x12\x13\n\x0erdiagExecution\x10\xde\x0b\x12\x19\n\x14immobilizerChallenge\x10\xc0\x0c\x12\x1d\n\x18immobilizerSearchKeyline\x10\xca\x0c\x12\x1e\n\x19immobilizerReleaseKeyline\x10\xd4\x0c\x12\x1b\n\x16immobilizerLockKeyline\x10\xde\x0c\x12\x1b\n\x16immobilizerLockVehicle\x10\xdf\x0c\x12\x1e\n\x19immobilizerReleaseVehicle\x10\xd5\x0c\x12\x14\n\x0fsetRentalSignal\x10\xa4\r\x12\x19\n\x14\x62lackchannelDownload\x10\x88\x0e\x12\x17\n\x12\x62lackchannelUpload\x10\x92\x0e\x12\x11\n\x0c\x63onfigurecsm\x10\xec\x0e\x12\x16\n\x11updatevehicleinfo\x10\xed\x0e\x12\x16\n\x11relaymessagetocsm\x10\xee\x0e\x12\x1c\n\x17relayrentalrequesttocsb\x10\xef\x0e\x12\x16\n\x11rtmDownloadConfig\x10\xe0\x12\x12\x12\n\rrtmReadConfig\x10\xea\x12\x12\x10\n\x0b\x61vpActivate\x10\x8c\x15\x12\x1b\n\x16\x63hargecontrolconfigure\x10\xf0\x15\x1a\x02\x10\x01\x42#\n\x1a\x63om.daimler.mbcarkit.proto\x90\x01\x01\xd0\xe1\x1e\x01\x62\x06proto3' , dependencies=[gogo__pb2.DESCRIPTOR,]) _VVA_COMMANDSTATE = _descriptor.EnumDescriptor( name='CommandState', full_name='proto.VVA.CommandState', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='UNKNOWN_COMMAND_STATE', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CREATED', index=1, number=1010, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ENQUEUED', index=2, number=1016, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PROCESSING', index=3, number=1012, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SUSPENDED', index=4, number=1017, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FINISHED', index=5, number=1018, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=None, serialized_start=40, serialized_end=158, ) _sym_db.RegisterEnumDescriptor(_VVA_COMMANDSTATE) _VVA_COMMANDCONDITION = _descriptor.EnumDescriptor( name='CommandCondition', full_name='proto.VVA.CommandCondition', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='UNKNWON_COMMAND_CONDITION', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='NONE', index=1, number=1000, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ACCEPTED', index=2, number=1001, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='REJECTED', index=3, number=1002, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TERMINATE', index=4, number=1003, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SUCCESS', index=5, number=1011, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FAILED', index=6, number=1013, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='OVERWRITTEN', index=7, number=1014, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TIMEOUT', index=8, number=1015, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=None, serialized_start=161, serialized_end=326, ) _sym_db.RegisterEnumDescriptor(_VVA_COMMANDCONDITION) _VEHICLEAPI_COMMANDSTATE = _descriptor.EnumDescriptor( name='CommandState', full_name='proto.VehicleAPI.CommandState', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='UNKNOWN_COMMAND_STATE', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='INITIATION', index=1, number=1, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ENQUEUED', index=2, number=2, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PROCESSING', index=3, number=3, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WAITING', index=4, number=4, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FINISHED', index=5, number=5, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FAILED', index=6, number=6, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=None, serialized_start=343, serialized_end=469, ) _sym_db.RegisterEnumDescriptor(_VEHICLEAPI_COMMANDSTATE) _VEHICLEAPI_ATTRIBUTESTATUS = _descriptor.EnumDescriptor( name='AttributeStatus', full_name='proto.VehicleAPI.AttributeStatus', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='VALUE_SET', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='VALUE_NOT_SET', index=1, number=1, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='INVALID', index=2, number=3, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='NOT_AVAILABLE', index=3, number=4, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=None, serialized_start=471, serialized_end=554, ) _sym_db.RegisterEnumDescriptor(_VEHICLEAPI_ATTRIBUTESTATUS) _VEHICLEAPI_QUEUETYPE = _descriptor.EnumDescriptor( name='QueueType', full_name='proto.VehicleAPI.QueueType', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='UNKNOWNCOMMANDQUEUETYPE', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DOORS', index=1, number=10, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='AUXHEAT', index=2, number=11, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECOND', index=3, number=12, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEOPT', index=4, number=13, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MAINTENANCE', index=5, number=14, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TCU', index=6, number=15, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FEED', index=7, number=16, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SERVICEACTIVATION', index=8, number=17, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ATP', index=9, number=18, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ASSISTANCE', index=10, number=19, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RACP', index=11, number=20, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WEEKPROFILE', index=12, number=21, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='REMOTEDIAGNOSIS', index=13, number=22, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FLSH', index=14, number=23, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TEMPERATURE', index=15, number=24, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TRIPCOMP', index=16, number=25, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ENGINE', index=17, number=26, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARM', index=18, number=27, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WINDOW', index=19, number=28, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='HEADUNIT', index=20, number=29, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MECALL', index=21, number=31, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZER', index=22, number=32, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RENTALSIGNAL', index=23, number=33, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='BCF', index=24, number=34, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PLUGANDCHARGE', index=25, number=35, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CARSHARINGMODULE', index=26, number=36, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='BATTERY', index=27, number=37, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ONBOARDFENCES', index=28, number=38, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDFENCES', index=29, number=39, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGINGTARIFFS', index=30, number=40, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RTMCONFIG', index=31, number=41, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MAINTENANCECOMPUTER', index=32, number=42, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MECALL2', index=33, number=43, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='AUTOMATEDVALETPARKING', index=34, number=44, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGECONTROL', index=35, number=45, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDALERT', index=36, number=46, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='unknowncommandqueuetype', index=37, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='doors', index=38, number=10, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='auxheat', index=39, number=11, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precond', index=40, number=12, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeopt', index=41, number=13, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='maintenance', index=42, number=14, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tcu', index=43, number=15, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='feed', index=44, number=16, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='serviceactivation', index=45, number=17, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='atp', index=46, number=18, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='assistance', index=47, number=19, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='racp', index=48, number=20, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='weekprofile', index=49, number=21, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='remotediagnosis', index=50, number=22, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='flsh', index=51, number=23, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='temperature', index=52, number=24, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tripcomp', index=53, number=25, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='engine', index=54, number=26, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarm', index=55, number=27, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='window', index=56, number=28, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='headunit', index=57, number=29, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='mecall', index=58, number=31, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizer', index=59, number=32, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rentalsignal', index=60, number=33, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='bcf', index=61, number=34, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='plugandcharge', index=62, number=35, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='carsharingmodule', index=63, number=36, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='battery', index=64, number=37, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='onboardfences', index=65, number=38, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedfences', index=66, number=39, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargingtariffs', index=67, number=40, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rtmconfig', index=68, number=41, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='maintenancecomputer', index=69, number=42, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='mecall2', index=70, number=43, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='automatedvaletparking', index=71, number=44, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargecontrol', index=72, number=45, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedalert', index=73, number=46, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=b'\020\001', serialized_start=557, serialized_end=1752, ) _sym_db.RegisterEnumDescriptor(_VEHICLEAPI_QUEUETYPE) _ACP_COMMANDTYPE = _descriptor.EnumDescriptor( name='CommandType', full_name='proto.ACP.CommandType', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='UNKNOWNCOMMANDTYPE', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DOORSLOCK', index=1, number=100, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DOORSUNLOCK', index=2, number=110, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TRUNKUNLOCK', index=3, number=115, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FUELFLAPUNLOCK', index=4, number=116, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEFLAPUNLOCK', index=5, number=117, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGECOUPLERUNLOCK', index=6, number=118, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DOORSPREPARERENTAL', index=7, number=120, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DOORSSECUREVEHICLE', index=8, number=130, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='AUXHEATSTART', index=9, number=300, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='AUXHEATSTOP', index=10, number=310, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='AUXHEATCONFIGURE', index=11, number=320, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TEMPERATURECONFIGURE', index=12, number=350, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WEEKPROFILECONFIGURE', index=13, number=360, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WEEKPROFILEV2CONFIGURE', index=14, number=370, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDSTART', index=15, number=400, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDSTOP', index=16, number=410, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDCONFIGURE', index=17, number=420, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDCONFIGURESEATS', index=18, number=425, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEOPTCONFIGURE', index=19, number=430, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEOPTSTART', index=20, number=440, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEOPTSTOP', index=21, number=450, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FEEDPOI', index=22, number=500, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FEEDFREETEXT', index=23, number=510, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ENGINESTART', index=24, number=550, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ENGINESTOP', index=25, number=560, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ENGINEAVPSTART', index=26, number=570, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TCUWAKEUP', index=27, number=600, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TCUSWUPDATE', index=28, number=610, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TCURCSRESET', index=29, number=620, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TCUINTERROGATION', index=30, number=630, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDALERTSTART', index=31, number=710, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDALERTSTOP', index=32, number=720, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FLSHSTART', index=33, number=750, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='FLSHSTOP', index=34, number=760, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SIGPOSSTART', index=35, number=770, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CONTRACTCONFIGURE', index=36, number=800, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CONTRACTREMOVE', index=37, number=810, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOTCONFIGURE', index=38, number=820, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOTREMOVE', index=39, number=830, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TRIPCOMP', index=40, number=850, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MAINTENANCECONFIGURE', index=41, number=930, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MAINTENANCECOMPUTEROFFSET', index=42, number=931, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SHORTTESTEXECUTE', index=43, number=935, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SERVICEACTIVATIONCONFIGURE', index=44, number=940, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DC2SERVICEACTIVATIONCONFIGURE', index=45, number=945, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DC2RAWDOWNLOAD', index=46, number=950, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='APPLICATIONCONFIGURATION', index=47, number=955, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='DC2STARTTRACKING', index=48, number=960, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ATPSEQUENCE', index=49, number=990, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMTOGGLEINTERIOR', index=50, number=1000, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMTOGGLETOW', index=51, number=1010, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMSELECTINTERIORTOW', index=52, number=1020, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMDESELECTINTERIORTOW', index=53, number=1030, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMSTOP', index=54, number=1040, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WINDOWOPEN', index=55, number=1100, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WINDOWCLOSE', index=56, number=1110, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WINDOWVENTILATE', index=57, number=1120, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='WINDOWMOVE', index=58, number=1121, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOFOPEN', index=59, number=1130, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOFCLOSE', index=60, number=1140, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOFLIFT', index=61, number=1150, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOFMOVE', index=62, number=1151, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='BATTERYMAXSOC', index=63, number=2000, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='BATTERYCHARGEPROGRAM', index=64, number=2010, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEPROGRAMCONFIGURE', index=65, number=2020, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ONBOARDFENCESCREATE', index=66, number=2100, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ONBOARDFENCESUPDATE', index=67, number=2110, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ONBOARDFENCESDELETE', index=68, number=2120, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDFENCESCREATE', index=69, number=2200, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDFENCESUPDATE', index=70, number=2210, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SPEEDFENCESDELETE', index=71, number=2220, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGINGTARIFFSCREATE', index=72, number=2300, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGINGTARIFFSUPDATE', index=73, number=2310, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGINGTARIFFSDELETE', index=74, number=2320, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMSTART', index=75, number=2500, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMSELECTINTERIOR', index=76, number=2510, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMDESELECTINTERIOR', index=77, number=2520, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMSELECTTOW', index=78, number=2530, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMDESELECTTOW', index=79, number=2540, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMSELECTDAMAGEDETECTION', index=80, number=2550, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMDESELECTDAMAGEDETECTION', index=81, number=2560, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='THEFTALARMCONFIRMDAMAGEDETECTION', index=82, number=2570, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MECALL2START', index=83, number=2600, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='UDXTRIGGERSYNCHRONIZATION', index=84, number=1200, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='UDXACTIVEUSERPROFILE', index=85, number=1210, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='UDXRESETUSERDATA', index=86, number=1220, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='USERPROFSYNCH', index=87, number=1230, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='USERDATARESET', index=88, number=1240, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PROFACTIVATIONSNAP', index=89, number=1250, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PROFACTIVATIONDIRECT', index=90, number=1255, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SOFTWAREUPDATE', index=91, number=1260, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PUSHNOTIFICATION', index=92, number=1270, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='MECALLCOMMAND', index=93, number=1310, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDSTARTRCS', index=94, number=1400, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDSTOPRCS', index=95, number=1410, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDCONFIGURERCS', index=96, number=1420, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TCUCONFIGURE', index=97, number=1430, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='EDISONSERVICEACTIVATION', index=98, number=1431, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TESTSEQUENCE', index=99, number=1432, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDCONFIGURERACP', index=100, number=1433, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGEOPTCONFIGURERACP', index=101, number=1434, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='TARIFFTABLEDOWNLOAD', index=102, number=1435, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDSTARTRACP', index=103, number=1436, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PRECONDSTOPRACP', index=104, number=1437, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOTCERTIFICATEREMOVE', index=105, number=1438, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ONREQUESTPROBEUPLOAD', index=106, number=1439, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ROOTCERTIFICATEDOWNLOAD', index=107, number=1440, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CONTRACTCERTIFICATEREMOVE', index=108, number=1441, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CONTRACTCERTIFICATEDOWNLOAD', index=109, number=1442, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='PROBECONFIGURATIONUPDATE', index=110, number=1443, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RDIAGDELETEECU', index=111, number=1500, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RDIAGSTATUSREPORT', index=112, number=1501, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RDIAGEXECUTION', index=113, number=1502, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZERCHALLENGE', index=114, number=1600, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZERSEARCHKEYLINE', index=115, number=1610, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZERRELEASEKEYLINE', index=116, number=1620, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZERLOCKKEYLINE', index=117, number=1630, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZERLOCKVEHICLE', index=118, number=1631, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='IMMOBILIZERRELEASEVEHICLE', index=119, number=1621, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='SETRENTALSIGNAL', index=120, number=1700, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='BLACKCHANNELDOWNLOAD', index=121, number=1800, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='BLACKCHANNELUPLOAD', index=122, number=1810, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CONFIGURECSM', index=123, number=1900, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='UPDATEVEHICLEINFO', index=124, number=1901, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RELAYMESSAGETOCSM', index=125, number=1902, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RELAYRENTALREQUESTTOCSB', index=126, number=1903, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RTMDOWNLOADCONFIG', index=127, number=2400, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='RTMREADCONFIG', index=128, number=2410, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='AVPACTIVATE', index=129, number=2700, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='CHARGECONTROLCONFIGURE', index=130, number=2800, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='unknownCommandType', index=131, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='doorsLock', index=132, number=100, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='doorsUnlock', index=133, number=110, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='trunkUnlock', index=134, number=115, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='fuelflapUnlock', index=135, number=116, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeflapUnlock', index=136, number=117, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargecouplerUnlock', index=137, number=118, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='doorsPrepareRental', index=138, number=120, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='doorsSecureVehicle', index=139, number=130, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='auxheatStart', index=140, number=300, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='auxheatStop', index=141, number=310, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='auxheatConfigure', index=142, number=320, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='temperatureConfigure', index=143, number=350, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='weekprofileConfigure', index=144, number=360, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='weekprofileV2Configure', index=145, number=370, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondStart', index=146, number=400, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondStop', index=147, number=410, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondConfigure', index=148, number=420, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondConfigureSeats', index=149, number=425, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeoptConfigure', index=150, number=430, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeoptStart', index=151, number=440, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeoptStop', index=152, number=450, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='feedPoi', index=153, number=500, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='feedFreetext', index=154, number=510, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='engineStart', index=155, number=550, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='engineStop', index=156, number=560, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='engineAvpstart', index=157, number=570, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tcuWakeup', index=158, number=600, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tcuSwUpdate', index=159, number=610, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tcuRcsReset', index=160, number=620, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tcuInterrogation', index=161, number=630, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedalertStart', index=162, number=710, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedalertStop', index=163, number=720, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='flshStart', index=164, number=750, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='flshStop', index=165, number=760, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='sigposStart', index=166, number=770, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='contractConfigure', index=167, number=800, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='contractRemove', index=168, number=810, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rootConfigure', index=169, number=820, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rootRemove', index=170, number=830, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tripcomp', index=171, number=850, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='maintenanceConfigure', index=172, number=930, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='maintenanceComputerOffset', index=173, number=931, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='shorttestExecute', index=174, number=935, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='serviceactivationConfigure', index=175, number=940, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='dc2ServiceactivationConfigure', index=176, number=945, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='dc2RawDownload', index=177, number=950, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='applicationConfiguration', index=178, number=955, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='dc2StartTracking', index=179, number=960, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='atpSequence', index=180, number=990, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmToggleInterior', index=181, number=1000, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmToggleTow', index=182, number=1010, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmSelectInteriorTow', index=183, number=1020, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmDeselectInteriorTow', index=184, number=1030, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmStop', index=185, number=1040, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='windowOpen', index=186, number=1100, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='windowClose', index=187, number=1110, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='windowVentilate', index=188, number=1120, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='windowMove', index=189, number=1121, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='roofOpen', index=190, number=1130, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='roofClose', index=191, number=1140, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='roofLift', index=192, number=1150, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='roofMove', index=193, number=1151, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='batteryMaxsoc', index=194, number=2000, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='batteryChargeprogram', index=195, number=2010, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeprogramconfigure', index=196, number=2020, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='onboardfencesCreate', index=197, number=2100, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='onboardfencesUpdate', index=198, number=2110, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='onboardfencesDelete', index=199, number=2120, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedfencesCreate', index=200, number=2200, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedfencesUpdate', index=201, number=2210, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='speedfencesDelete', index=202, number=2220, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargingtariffsCreate', index=203, number=2300, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargingtariffsUpdate', index=204, number=2310, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargingtariffsDelete', index=205, number=2320, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmstart', index=206, number=2500, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmselectinterior', index=207, number=2510, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmdeselectinterior', index=208, number=2520, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmselecttow', index=209, number=2530, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmdeselecttow', index=210, number=2540, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmselectdamagedetection', index=211, number=2550, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmdeselectdamagedetection', index=212, number=2560, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='theftalarmconfirmdamagedetection', index=213, number=2570, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='mecall2start', index=214, number=2600, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='udxTriggerSynchronization', index=215, number=1200, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='udxActiveUserProfile', index=216, number=1210, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='udxResetUserData', index=217, number=1220, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='userProfSynch', index=218, number=1230, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='userDataReset', index=219, number=1240, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='profActivationSnap', index=220, number=1250, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='profActivationDirect', index=221, number=1255, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='softwareUpdate', index=222, number=1260, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='pushNotification', index=223, number=1270, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='mecallcommand', index=224, number=1310, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondStartRcs', index=225, number=1400, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondStopRcs', index=226, number=1410, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondConfigureRcs', index=227, number=1420, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tcuConfigure', index=228, number=1430, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='edisonServiceActivation', index=229, number=1431, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='testSequence', index=230, number=1432, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondConfigureRacp', index=231, number=1433, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargeoptConfigureRacp', index=232, number=1434, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='tariffTableDownload', index=233, number=1435, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondStartRacp', index=234, number=1436, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='precondStopRacp', index=235, number=1437, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rootCertificateRemove', index=236, number=1438, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='onRequestProbeUpload', index=237, number=1439, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rootCertificateDownload', index=238, number=1440, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='contractCertificateRemove', index=239, number=1441, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='contractCertificateDownload', index=240, number=1442, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='probeConfigurationUpdate', index=241, number=1443, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rdiagDeleteEcu', index=242, number=1500, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rdiagStatusReport', index=243, number=1501, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rdiagExecution', index=244, number=1502, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizerChallenge', index=245, number=1600, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizerSearchKeyline', index=246, number=1610, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizerReleaseKeyline', index=247, number=1620, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizerLockKeyline', index=248, number=1630, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizerLockVehicle', index=249, number=1631, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='immobilizerReleaseVehicle', index=250, number=1621, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='setRentalSignal', index=251, number=1700, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='blackchannelDownload', index=252, number=1800, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='blackchannelUpload', index=253, number=1810, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='configurecsm', index=254, number=1900, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='updatevehicleinfo', index=255, number=1901, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='relaymessagetocsm', index=256, number=1902, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='relayrentalrequesttocsb', index=257, number=1903, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rtmDownloadConfig', index=258, number=2400, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='rtmReadConfig', index=259, number=2410, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='avpActivate', index=260, number=2700, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='chargecontrolconfigure', index=261, number=2800, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=b'\020\001', serialized_start=1763, serialized_end=8068, ) _sym_db.RegisterEnumDescriptor(_ACP_COMMANDTYPE) _VVA = _descriptor.Descriptor( name='VVA', full_name='proto.VVA', filename=None, file=DESCRIPTOR, containing_type=None, create_key=_descriptor._internal_create_key, fields=[ ], extensions=[ ], nested_types=[], enum_types=[ _VVA_COMMANDSTATE, _VVA_COMMANDCONDITION, ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=33, serialized_end=326, ) _VEHICLEAPI = _descriptor.Descriptor( name='VehicleAPI', full_name='proto.VehicleAPI', filename=None, file=DESCRIPTOR, containing_type=None, create_key=_descriptor._internal_create_key, fields=[ ], extensions=[ ], nested_types=[], enum_types=[ _VEHICLEAPI_COMMANDSTATE, _VEHICLEAPI_ATTRIBUTESTATUS, _VEHICLEAPI_QUEUETYPE, ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=329, serialized_end=1752, ) _ACP = _descriptor.Descriptor( name='ACP', full_name='proto.ACP', filename=None, file=DESCRIPTOR, containing_type=None, create_key=_descriptor._internal_create_key, fields=[ ], extensions=[ ], nested_types=[], enum_types=[ _ACP_COMMANDTYPE, ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1755, serialized_end=8068, ) _VVA_COMMANDSTATE.containing_type = _VVA _VVA_COMMANDCONDITION.containing_type = _VVA _VEHICLEAPI_COMMANDSTATE.containing_type = _VEHICLEAPI _VEHICLEAPI_ATTRIBUTESTATUS.containing_type = _VEHICLEAPI _VEHICLEAPI_QUEUETYPE.containing_type = _VEHICLEAPI _ACP_COMMANDTYPE.containing_type = _ACP DESCRIPTOR.message_types_by_name['VVA'] = _VVA DESCRIPTOR.message_types_by_name['VehicleAPI'] = _VEHICLEAPI DESCRIPTOR.message_types_by_name['ACP'] = _ACP _sym_db.RegisterFileDescriptor(DESCRIPTOR) VVA = _reflection.GeneratedProtocolMessageType('VVA', (_message.Message,), { 'DESCRIPTOR' : _VVA, '__module__' : 'acp_pb2' # @@protoc_insertion_point(class_scope:proto.VVA) }) _sym_db.RegisterMessage(VVA) VehicleAPI = _reflection.GeneratedProtocolMessageType('VehicleAPI', (_message.Message,), { 'DESCRIPTOR' : _VEHICLEAPI, '__module__' : 'acp_pb2' # @@protoc_insertion_point(class_scope:proto.VehicleAPI) }) _sym_db.RegisterMessage(VehicleAPI) ACP = _reflection.GeneratedProtocolMessageType('ACP', (_message.Message,), { 'DESCRIPTOR' : _ACP, '__module__' : 'acp_pb2' # @@protoc_insertion_point(class_scope:proto.ACP) }) _sym_db.RegisterMessage(ACP) DESCRIPTOR._options = None _VEHICLEAPI_QUEUETYPE._options = None _ACP_COMMANDTYPE._options = None # @@protoc_insertion_point(module_scope)
89,313
36,201
''' Uses HPI [[https://github.com/karlicoss/HPI/blob/master/doc/MODULES.org#myreddit][reddit]] module ''' from itertools import chain from typing import Set, Optional from ..common import Visit, Loc, extract_urls, Results, logger def index(*, render_markdown: bool = False, renderer: Optional['RedditRenderer'] = None) -> Results: from . import hpi try: from my.reddit.all import submissions, comments, saved, upvoted except ModuleNotFoundError as e: if "No module named 'my.reddit.all'" in str(e): import warnings warnings.warn("DEPRECATED/reddit: Using an old version of HPI, please update") from my.reddit import submissions, comments, saved, upvoted # type: ignore[no-redef] else: raise e if renderer is not None: assert callable(renderer), f"{renderer} is not a callable (should be a subclass of RedditRenderer)" r = renderer(render_markdown=render_markdown) else: r = RedditRenderer(render_markdown=render_markdown) logger.info('processing saves') for s in saved(): try: yield from r._from_save(s) except Exception as e: yield e logger.info('processing comments') for c in comments(): try: yield from r._from_comment(c) except Exception as e: yield e logger.info('processing submissions') for sub in submissions(): try: yield from r._from_submission(sub) except Exception as e: yield e logger.info('processing upvotes') for u in upvoted(): try: yield from r._from_upvote(u) except Exception as e: yield e # mostly here so we can keep track of how the user # wants to render markdown class RedditRenderer: def __init__(self, render_markdown: bool = False): self._link_extractor = None self._parser_cls = None try: from .markdown import TextParser, extract_from_text self._link_extractor = extract_from_text self._parser_cls = TextParser except ImportError as import_err: # TODO: add dummy _link_extractor and _parser_cls classes incase # these are called by a subclass? # only send error if the user is trying to enable this feature if render_markdown: logger.exception(import_err) logger.critical("Could not import markdown module to render reddit markdown. Try 'python3 -m pip install mistletoe'") render_markdown = False # force to be false, couldn't import self.render_markdown = render_markdown def _from_comment(self, i: 'Comment') -> Results: locator = Loc.make( title='Reddit comment', href=i.url, ) yield from self._from_common(i, locator=locator) def _from_submission(self, i: 'Submission') -> Results: locator = Loc.make( title=f'Reddit submission: {i.title}', href=i.url, ) yield from self._from_common(i, locator=locator) def _from_upvote(self, i: 'Upvote') -> Results: locator = Loc.make( title=f'Reddit upvote', href=i.url, ) yield from self._from_common(i, locator=locator) def _from_save(self, i: 'Save') -> Results: locator = Loc.make( title='Reddit save', href=i.url, ) yield from self._from_common(i, locator=locator) # to allow for possible subclassing by the user? def _render_body(self, text: str) -> str: if self.render_markdown and self._parser_cls is not None: return self._parser_cls(text)._doc_ashtml() else: return text def _from_common(self, i: 'RedditBase', locator: Loc) -> Results: urls = [i.url] # TODO this should belong to HPI.. fix permalink handling I guess # ok, it's not present for all of them.. lurl = i.raw.get('link_url') if lurl is not None: urls.append(lurl) lurl = i.raw.get('url') if lurl is not None: urls.append(lurl) context = self._render_body(i.text) emitted: Set[str] = set() for url in chain(urls, extract_urls(i.text)): if url in emitted: continue yield Visit( url=url, dt=i.created, context=context, locator=locator, ) emitted.add(url) # extract from markdown links like [link text](https://...) # incase URLExtract missed any # # this should try to do this, even if the user didn't enable # the render_markdown flag, as it may catch extra links that URLExtract didnt # would still require mistletoe to be installed, but # the user may already have it installed for the auto/markdown modules if self._link_extractor is not None: for res in self._link_extractor(i.text): if isinstance(res, Exception): yield res continue if res.url in emitted: continue yield Visit( url=res.url, dt=i.created, context=context, locator=locator, ) emitted.add(res.url) import typing if typing.TYPE_CHECKING: from my.reddit.common import Submission, Comment, Save, Upvote, RedditBase
5,632
1,598
import optparse import sys from sys import getsizeof import logging from signal import signal, SIGINT import time import requests # MIT License # # Copyright (c) 2022 SaicharanKandukuri # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. from rich.logging import RichHandler FORMAT = "%(message)s" logging.basicConfig( level="NOTSET", format=FORMAT, datefmt="[%X]", handlers=[RichHandler()] ) logging.disable('DEBUG') log = logging.getLogger("rich") class WifiUtils: """class for wifi utils""" def __init__(self, username, password, host, port): self.username = username self.password = password self.host = host self.port = port @classmethod def request(cls, method, username, password, host, port, timeout) -> list: """request method: sends request to wifi host Args: method (str): interaction method "login.xml" or "logout.xml". Defaults to "login.xml". username (str): username assigned by parul university to access wifi password (str): password assigned by parul university to access wifi host (str): hostname of the parul university wifi hotspot/routers Defaults to " port (str): port to send login request. Defaults to "8090". timeout (int): request timeout. Defaults to 10. Returns: list server_request status[true|false] response(xml data returned form server) status_code(web request status code) """ url = ("http://"+host+":"+port+"/"+method) body = ("mode=191&username=" + username + "&password=" + password + "&a=1630404423764&producttype=0" ) headers = { "Host": "http://" + host + ":" + port + "", "Content-Length": str(getsizeof(body)), "User-Agent": "Chrome/92.0.4515.159 Safari/537.36", "Content-Type": "application/x-www-form-urlencoded", "Accept": "*/*", "Origin": "http://" + host + ":" + port, "Referer": "http://" + host + ":" + port + "/", "Accept-Encoding": "gzip defalte", "Accept-Language": "en-US,en;q=0.9", "Connection": "close", } body_array = bytearray(body, 'utf-8') req = requests.post(url, data=body_array, headers=headers, timeout=timeout, verify=False ) return [(req.status_code == 200), req.text, req.status_code] def login(self, username, password, host, port="8090", method="login.xml", timeout=10) -> list: """login: uses request method to send login web request with credentials to wifi host Args: username (str): username assigned by parul university to access wifi password (str): password assigned by parul university to access wifi host (str): hostname of the parul university wifi hotspot/routers Defaults to "10.0.0.11" port (str, optional): port to send login request. Defaults to "8090". method (str, optional): interaction method "login.xml" or "logout.xml". Defaults to "login.xml". timeout (int, optional): request timeout. Defaults to 10. """ return self.request(method, username, password, host, port, timeout) def logout(self, username, password, host, port="8090", method="logout.xml", timeout=10) -> list: """logout: uses request method to send logout web request with credentials to wifi host Args: username (str): username assigned by parul university to access wifi password (str): password assigned by parul university to access wifi host (str): hostname of the parul university wifi hotspot/routers Defaults to "10.0.0.11" port (str, optional): port to send login request. Defaults to "8090". method (str, optional): interaction method "login.xml" or "logout.xml". Defaults to "logout.xml". timeout (int, optional): request timeout. Defaults to 10. """ return self.request(method, username, password, host, port, timeout) # def get_xml_msg(xml): # for later (●'◡'●) # return Et.parse(xml).getroot()[1] def grey_print(_string): """prints outs grey text Args: _string (str) """ print(f"\033[90m{_string}\033[0m") def connection_to(url, timeout=10): """checks if connection to url is available""" try: requests.get(url, timeout=timeout) return True except (requests.ConnectionError, requests.Timeout): return False def keep_alive(username, password, host, port): """keeps connection alive to wifi host""" while True: if connection_to("http://10.0.0.11:8090/"): log.info("connection to router \"available\"") else: log.critical("connection to router \"unavailable\"") if connection_to("https://google.com"): log.info("Connected to the internet") else: log.warning("Not connected to the internet") log.info("Tying to login back") try: log.info(WifiUtils.login(username, password, host, port)) except (requests.ConnectionError, requests.Timeout): log.critical( "Connection error: \"UNSTABLE CONNECTION TO HOST\"") time.sleep(5) def exit_handler(_signal, frame): """captures keyboard interrupts and kill signals & exits with messesage""" log.warning('SIGINT or CTRL-C detected. Exiting gracefully') grey_print("signal:"+str(_signal)) grey_print("frame:"+str(frame)) sys.exit(0) if __name__ == '__main__': signal(SIGINT, exit_handler) parser = optparse.OptionParser() parser.add_option('-u', '--username', dest='username', help='username to login/logout with parul university wifi service') parser.add_option('-p', '--password', dest='password', help='password to login/logout with parul university wifi service') parser.add_option('-H', '--host', dest='host', default='10.0.0.11', type=str) parser.add_option('-P', '--port', dest='port', default='8090', type=str) parser.add_option('-k', '--keep-alive', action='store_true', help='keep connecting to wifi when it gets signed out', default=False) parser.add_option('-o', '--logout', action='store_true', help='logout from wifi', default=False) parser.add_option('-l', '--login', action='store_true', help='login to wifi', default=False) options, args = parser.parse_args() WifiUtils = WifiUtils( options.username, options.password, options.host, options.port) if options.login: log.info("=> login <=") log.info(WifiUtils.login(options.username, options.password, options.host, options.port, )) sys.exit(0) if options.logout: log.info("=> logout <=") log.info(WifiUtils.logout(options.username, options.password, options.host, options.port, )) sys.exit(0) if options.keep_alive: log.info("=> keep alive <=") keep_alive(options.username, options.password, options.host, options.port, )
8,992
2,576
""" sphinx-simulink.application ~~~~~~~~~~~~~~~~~~~~~~~ Embed Simulink diagrams on your documentation. :copyright: Copyright 2016 by Dennis Edward Kalinowski <dekalinowski@gmail.com>. :license: MIT, see LICENSE for details. """ import matlab.engine import os from sphinx.errors import SphinxError from sphinx.util.osutil import ensuredir from sphinxsimulink.diagram import directives,nodes from sphinxsimulink.metadata import __version__ engine = None class SimulinkDiagramError(SphinxError): pass def render_diagram(app, node, docname): global engine uri = node['uri'] # do not regenerate if os.path.exists( uri ): pass ensuredir( os.path.dirname( uri ) ) try: # reuse last engine to save loading time if engine == None: engine = matlab.engine.start_matlab() else: # clean up used engines engine.restoredefaultpath(nargout=0) engine.close('all', nargout=0) engine.bdclose('all', nargout=0) engine.clear('classes', nargout=0) # start engine from document directory engine.cd( os.path.dirname( app.env.doc2path( docname ) ) ) # then, support changing directory (relative to document) dir = node.get('dir') if dir: engine.cd( dir ) # finally, add the MATLAB paths relative to the changed directory pathlist = node.get('addpath') if pathlist: for path in pathlist: engine.addpath( path ) # preload script preload = node.get('preload') if preload: engine.eval( preload + ';', nargout=0) # load system system = node.get('system') if system: engine.load_system( system ); # if subsystem specified, print from this layer subsystem = node.get('subsystem') if subsystem: system = "/".join( [ system, subsystem ] ) # print from Simulink handle to .png engine.eval( "print( get_param( '{}', 'Handle' ), '-dpng', '{}' )". format( system, uri ), nargout=0 ) except matlab.engine.MatlabExecutionError as err: raise SimulinkDiagramError('Unable to render Simulink diagram due ' + 'to MATLAB execution error' ) def process_diagram_nodes(app, doctree, docname): for node in doctree.traverse(nodes.diagram): render_diagram(app, node, docname) node.replace_self(node.children) def terminate_matlab_engine(app, exception): global engine if engine is not None: engine.quit() engine = None def setup(app): app.add_directive('simulink-diagram', directives.SimulinkDiagramDirective) app.connect('doctree-resolved', process_diagram_nodes) app.connect('build-finished', terminate_matlab_engine) return {'version': __version__}
2,974
907
from ._internal.frameworks.pytorch_lightning import load from ._internal.frameworks.pytorch_lightning import save from ._internal.frameworks.pytorch_lightning import load_runner __all__ = ["load", "load_runner", "save"]
221
71
from django import template from django.conf import settings register = template.Library() @register.simple_tag def settings_value(name): """ This Tag allows to access values from the configuration file in the templates. """ try: return settings.__getattr__(name) except AttributeError: return ""
335
83
""" Author: Konstantinos Angelopoulos Date: 04/02/2020 All rights reserved. Feel free to use and modify and if you like it give it a star. Import the Robot's Step Files and Color/Scale/Assemble them using the instructions in /RoboDK/KUKA/KUKA LWR IV+ Description (for Original=kuka_lwr_model_description.json, for custom=custom_kuka_lwr_model_description, for custom2=custom_kuka_lwr_model_description_2) before running the code to complete the robot model. ######################################################################### ######### To quickly color and scale use the next lines of code ######### ######################################################################### from robolink import * # RoboDK API from robodk import * # Robot toolbox RDK = Robolink() for station in RDK.ItemList(): for item in station.Childs(): item.Scale(1000) item.setColor(255/255, 85/255, 0/255, 255/255) ######################################################################## #### For custom2 run these commands before assembling the stl files #### ######################################################################## from robolink import * # RoboDK API from robodk import * # Robot toolbox import numpy as np RDK = Robolink() for station in RDK.ItemList(): for item in station.Childs(): item.setGeometryPose(item.Pose()*rotz(np.pi)) item.Scale(1000) item.setColor(255/255, 85/255, 0/255, 255/255) and after building the mechanism and import it, in order to rotate the robot run: from robolink import * # RoboDK API from robodk import * # Robot toolbox RDK = Robolink() ref = RDK.Item('reference2') ref.setPoseAbs(ref.Pose()*rotz(pi)) ############################################################################################## ##### The original option is just the robot model without any inverted sense and joints ###### ##### home are [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ########################################### ############################################################################################## ##### The custom robot is the real model that has the same limitations, home joints and ###### ##### senses as the REAl KUKA LWR but the X and Y axis system are inverted ################### ############################################################################################## ##### The custom2 robot is the same as the custom option but with the X and Y axis being ##### ##### the same as the REAL KUKA ROBOT ######################################################## ############################################################################################## """ # Start the RoboDK API from robolink.robolink import * from robodk.robodk import * import json import os # ORIGINAL ROBOT DATA with open(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'RoboDK/KUKA/KUKA LWR IV+ Description/kuka_lwr_model_description.json')) as config_file: data = json.load(config_file) original_robot_name = data['Robot name'] original_robot_dof = data['DOF'] original_robot_joint1 = data['Joint 1'] original_robot_joint2 = data['Joint 2'] original_robot_joint3 = data['Joint 3'] original_robot_joint4 = data['Joint 4'] original_robot_joint5 = data['Joint 5'] original_robot_joint6 = data['Joint 6'] original_robot_joint7 = data['Joint 7'] original_robot_joints_build = [original_robot_joint1["Build joints"], original_robot_joint2["Build joints"], original_robot_joint3["Build joints"], original_robot_joint4["Build joints"], original_robot_joint5["Build joints"], original_robot_joint6["Build joints"], original_robot_joint7["Build joints"]] original_robot_joints_home = [original_robot_joint1["Home"], original_robot_joint2["Home"], original_robot_joint3["Home"], original_robot_joint4["Home"], original_robot_joint5["Home"], original_robot_joint6["Home"], original_robot_joint7["Home"]] original_robot_parameters = [data["d1"], data["d3"], data["d5"], data["d7"], data["dtheta1"], data["dtheta2"], data["dtheta3"], data["dtheta4"], data["dtheta5"], data["dtheta6"], data["dtheta7"]] original_robot_joint_senses = [original_robot_joint1["Invert Sense"], original_robot_joint2["Invert Sense"], original_robot_joint3["Invert Sense"], original_robot_joint4["Invert Sense"], original_robot_joint5["Invert Sense"], original_robot_joint6["Invert Sense"], original_robot_joint7["Invert Sense"]] original_robot_joint_lower_limit = [original_robot_joint1["Minimum limit"], original_robot_joint2["Minimum limit"], original_robot_joint3["Minimum limit"], original_robot_joint4["Minimum limit"], original_robot_joint5["Minimum limit"], original_robot_joint6["Minimum limit"], original_robot_joint7["Minimum limit"]] original_robot_joint_upper_limit = [original_robot_joint1["Maximum limit"], original_robot_joint2["Maximum limit"], original_robot_joint3["Maximum limit"], original_robot_joint4["Maximum limit"], original_robot_joint5["Maximum limit"], original_robot_joint6["Maximum limit"], original_robot_joint7["Maximum limit"]] original_robot_base_pose = data["Base shift"] original_robot_tool_pose = data["End-effector shift"] # CUSTOM ROBOT DATA with open(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'RoboDK/KUKA/KUKA LWR IV+ Description/custom_lwr_model_description.json')) as config_file: data = json.load(config_file) custom_robot_name = data['Robot name'] custom_robot_dof = data['DOF'] custom_robot_joint1 = data['Joint 1'] custom_robot_joint2 = data['Joint 2'] custom_robot_joint3 = data['Joint 3'] custom_robot_joint4 = data['Joint 4'] custom_robot_joint5 = data['Joint 5'] custom_robot_joint6 = data['Joint 6'] custom_robot_joint7 = data['Joint 7'] custom_robot_joints_build = [custom_robot_joint1["Build joints"], custom_robot_joint2["Build joints"], custom_robot_joint3["Build joints"], custom_robot_joint4["Build joints"], custom_robot_joint5["Build joints"], custom_robot_joint6["Build joints"], custom_robot_joint7["Build joints"]] custom_robot_joints_home = [custom_robot_joint1["Home"], custom_robot_joint2["Home"], custom_robot_joint3["Home"], custom_robot_joint4["Home"], custom_robot_joint5["Home"], custom_robot_joint6["Home"], custom_robot_joint7["Home"]] custom_robot_parameters = [data["d1"], data["d3"], data["d5"], data["d7"], data["dtheta1"], data["dtheta2"], data["dtheta3"], data["dtheta4"], data["dtheta5"], data["dtheta6"], data["dtheta7"]] custom_robot_joint_senses = [custom_robot_joint1["Invert Sense"], custom_robot_joint2["Invert Sense"], custom_robot_joint3["Invert Sense"], custom_robot_joint4["Invert Sense"], custom_robot_joint5["Invert Sense"], custom_robot_joint6["Invert Sense"], custom_robot_joint7["Invert Sense"]] custom_robot_joint_lower_limit = [custom_robot_joint1["Minimum limit"], custom_robot_joint2["Minimum limit"], custom_robot_joint3["Minimum limit"], custom_robot_joint4["Minimum limit"], custom_robot_joint5["Minimum limit"], custom_robot_joint6["Minimum limit"], custom_robot_joint7["Minimum limit"]] custom_robot_joint_upper_limit = [custom_robot_joint1["Maximum limit"], custom_robot_joint2["Maximum limit"], custom_robot_joint3["Maximum limit"], custom_robot_joint4["Maximum limit"], custom_robot_joint5["Maximum limit"], custom_robot_joint6["Maximum limit"], custom_robot_joint7["Maximum limit"]] custom_robot_base_pose = data["Base shift"] custom_robot_tool_pose = data["End-effector shift"] # CUSTOM 2 ROBOT DATA with open(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'RoboDK/KUKA/KUKA LWR IV+ Description/custom_lwr_model_description_2.json')) as config_file: data = json.load(config_file) custom_2_robot_name = data['Robot name'] custom_2_robot_dof = data['DOF'] custom_2_robot_joint1 = data['Joint 1'] custom_2_robot_joint2 = data['Joint 2'] custom_2_robot_joint3 = data['Joint 3'] custom_2_robot_joint4 = data['Joint 4'] custom_2_robot_joint5 = data['Joint 5'] custom_2_robot_joint6 = data['Joint 6'] custom_2_robot_joint7 = data['Joint 7'] custom_2_robot_joints_build = [custom_2_robot_joint1["Build joints"], custom_2_robot_joint2["Build joints"], custom_2_robot_joint3["Build joints"], custom_2_robot_joint4["Build joints"], custom_2_robot_joint5["Build joints"], custom_2_robot_joint6["Build joints"], custom_2_robot_joint7["Build joints"]] custom_2_robot_joints_home = [custom_2_robot_joint1["Home"], custom_2_robot_joint2["Home"], custom_2_robot_joint3["Home"], custom_2_robot_joint4["Home"], custom_2_robot_joint5["Home"], custom_2_robot_joint6["Home"], custom_2_robot_joint7["Home"]] custom_2_robot_parameters = [data["d1"], data["d3"], data["d5"], data["d7"], data["dtheta1"], data["dtheta2"], data["dtheta3"], data["dtheta4"], data["dtheta5"], data["dtheta6"], data["dtheta7"]] custom_2_robot_joint_senses = [custom_2_robot_joint1["Invert Sense"], custom_2_robot_joint2["Invert Sense"], custom_2_robot_joint3["Invert Sense"], custom_2_robot_joint4["Invert Sense"], custom_2_robot_joint5["Invert Sense"], custom_2_robot_joint6["Invert Sense"], custom_2_robot_joint7["Invert Sense"]] custom_2_robot_joint_lower_limit = [custom_2_robot_joint1["Minimum limit"], custom_2_robot_joint2["Minimum limit"], custom_2_robot_joint3["Minimum limit"], custom_2_robot_joint4["Minimum limit"], custom_2_robot_joint5["Minimum limit"], custom_2_robot_joint6["Minimum limit"], custom_2_robot_joint7["Minimum limit"]] custom_2_robot_joint_upper_limit = [custom_2_robot_joint1["Maximum limit"], custom_2_robot_joint2["Maximum limit"], custom_2_robot_joint3["Maximum limit"], custom_2_robot_joint4["Maximum limit"], custom_2_robot_joint5["Maximum limit"], custom_2_robot_joint6["Maximum limit"], custom_2_robot_joint7["Maximum limit"]] custom_2_robot_base_pose = data["Base shift"] custom_2_robot_tool_pose = data["End-effector shift"] RDK = Robolink() custom = False custom2 = True if custom: robot_name = custom_robot_name DOFs = custom_robot_dof joints_build = custom_robot_joints_build joints_home = custom_robot_joints_home parameters = custom_robot_parameters joints_senses = custom_robot_joint_senses # -1 = Inverted, +1 = Not Inverted lower_limits = custom_robot_joint_lower_limit upper_limits = custom_robot_joint_upper_limit base_pose = xyzrpw_2_pose(custom_robot_base_pose) tool_pose = xyzrpw_2_pose(custom_robot_tool_pose) list_objects = [] elif custom2: robot_name = custom_2_robot_name DOFs = custom_2_robot_dof joints_build = custom_2_robot_joints_build joints_home = custom_2_robot_joints_home parameters = custom_2_robot_parameters joints_senses = custom_2_robot_joint_senses # -1 = Inverted, +1 = Not Inverted lower_limits = custom_2_robot_joint_lower_limit upper_limits = custom_2_robot_joint_upper_limit base_pose = xyzrpw_2_pose(custom_2_robot_base_pose) tool_pose = xyzrpw_2_pose(custom_2_robot_tool_pose) list_objects = [] else: robot_name = original_robot_name DOFs = original_robot_dof joints_build = original_robot_joints_build joints_home = original_robot_joints_home parameters = original_robot_parameters joints_senses = original_robot_joint_senses # -1 = Inverted, +1 = Not Inverted lower_limits = original_robot_joint_lower_limit upper_limits = original_robot_joint_upper_limit base_pose = xyzrpw_2_pose(original_robot_base_pose) tool_pose = xyzrpw_2_pose(original_robot_tool_pose) list_objects = [] for i in range(DOFs + 1): if i == 0: itm = RDK.Item('base', ITEM_TYPE_OBJECT) else: itm = RDK.Item('link_'+str(i), ITEM_TYPE_OBJECT) list_objects.append(itm) new_robot = RDK.BuildMechanism(MAKE_ROBOT_7DOF, list_objects, parameters, joints_build, joints_home, joints_senses, lower_limits, upper_limits, base_pose, tool_pose, robot_name) if not new_robot.Valid(): print("Failed to create the robot. Check input values.") else: print("Robot/mechanism created: " + new_robot.Name())
13,195
4,310
import unittest from context import parser class TVShowFileParserTests(unittest.TestCase): def setUp(self): self.filename = parser.Parser("test.2018.S01E01E02.mkv") def tearDown(self): self.filename = None def testisSubtitleFileSRT(self): self.assertFalse(self.filename.isSubs) if __name__ == '__main__': unittest.main()
368
131
# -*- coding: utf-8 -*- # This file is part of Archdiffer and is released under # the MIT License: http://www.opensource.org/licenses/mit-license.php """ Created on Sat May 12 20:39:46 2018 @author: Pavla Kratochvilova <pavla.kratochvilova@gmail.com> """ from random import choice from datetime import datetime from . import RESTTest from ..constants import STATE_STRINGS from .. import database DATETIMES = [ '1000-01-01 00:00:00', '2018-01-01 00:00:00', '9999-01-01 00:00:00', ] IDS = LIMITS = OFFSETS = ['0', '1', '2', '10', '999999'] STATES = list(STATE_STRINGS.values()) class RESTTestRoutes(RESTTest): """Tests for route 'rest'.""" expected_response = { "/rest": { "methods": [ "GET" ], "routes": { "/rest/comparison_types": { "methods": [ "GET" ], "routes": { "/rest/comparison_types/<int:id>": { "methods": [ "GET" ], "routes": {} } } }, "/rest/comparisons": { "methods": [ "GET" ], "routes": { "/rest/comparisons/<int:id>": { "methods": [ "GET" ], "routes": {} } } } } } } def assert_response(self): """Asert that response is as expected.""" self.assertEqual(self.response, self.expected_response) def test_get_routes(self): """Test GET method on 'rest'.""" self.get('rest') self.assert_code_ok() self.assert_response() class RESTTestLists(RESTTest): """For testing getting lists from the database.""" route = None # Accepted query parameters (values are lists of tested values). param_choices = {} def setUp(self): super().setUp() self.params = {} def set_params(self, parameters): """Randomly set parameters based on the given list. :param list parameters: list of parameters to be set. """ for param in self.param_choices: if param in parameters: self.params[param] = choice(self.param_choices[param]) def form_request(self): """Form GET request based on route and parameters.""" self.get(self.route, params=self.params) def form_request_one(self): """Form GET request for random item based on route and parameters.""" self.get(self.route, choice(IDS), params=self.params) def deep_sort(self, response): """Order all lists in the response. :param list response: response :return list: ordered response """ if type(response) == list: response = sorted(response, key=lambda k: k['id']) new_response = [] for item in response: new_response.append(self.deep_sort(item)) elif type(response) == dict: for key, value in response.items(): response[key] = self.deep_sort(value) return response def assert_response(self, expected): """Assert that response is as expected, aside from lists ordering. :param list expected: expected response """ self.assertEqual( self.deep_sort(self.response), self.deep_sort(expected) ) class RESTTestListsEmpty(RESTTestLists): """Tests for getting lists from empty database. Abstract.""" def run(self, result=None): """Overriden run so that it doesn't run these tests on this class.""" if type(self) == RESTTestListsEmpty: return result return super().run(result) def test_basic(self): """Test getting list - with no params set.""" self.form_request() self.assert_code_ok() self.assert_response([]) def test_basic_one(self): """Test getting instance - with no params set.""" self.form_request_one() self.assert_code_ok() self.assert_response([]) def test_individual_params(self): """Test getting list - for each param set individually.""" for param in self.param_choices: with self.subTest(param=param): self.params = {} self.set_params([param]) self.test_basic() def test_all_params(self): """Test getting list - for all params set.""" self.params = {} self.set_params(self.param_choices) self.test_basic() class RESTTestComparisonsEmpty(RESTTestListsEmpty): """Tests for getting comparisons from empty database.""" route = 'rest/comparisons' param_choices = { 'id': IDS, 'state': STATES, 'before': DATETIMES, 'after': DATETIMES, 'comparison_type_id': IDS, 'comparison_type_name': [''], 'limit': LIMITS, 'offset': OFFSETS, } class RESTTestComparisonTypesEmpty(RESTTestListsEmpty): """Tests for getting comparison types from empty database.""" route = 'rest/comparison_types' param_choices = { 'id': IDS, 'name': [''], 'limit': LIMITS, 'offset': OFFSETS, } class RESTTestListsFilled(RESTTestLists): """Tests for getting lists from filled database. Abstract.""" def run(self, result=None): """Overriden run so that it doesn't run these tests on this class.""" if type(self) == RESTTestListsFilled: return result return super().run(result) def test_params(self): """Run test for each of the tuples_params_results. Check that with given parameters the response is as expected.""" for params, expected in self.tuples_params_results: with self.subTest(**params): self.params = params self.form_request() self.assert_code_ok() self.assert_response(expected) class RESTTestComparisonsFilled(RESTTestListsFilled): """Tests for getting comparisons from filled database.""" route = RESTTestComparisonsEmpty.route param_choices = RESTTestComparisonsEmpty.param_choices def fill_db(self): """Fill database. Called in setUp.""" db_session = database.session() comparison_types = [ database.ComparisonType(id=1, name='1'), database.ComparisonType(id=2, name='2'), ] comparisons = [ database.Comparison( id=1, state=1, time=datetime(9999, 1, 1), comparison_type_id=1 ), database.Comparison( id=2, state=0, time=datetime(1000, 1, 1), comparison_type_id=2 ), database.Comparison( id=3, state=0, time=datetime(2018, 1, 1), comparison_type_id=1 ), database.Comparison( id=4, state=1, time=datetime(2018, 3, 1), comparison_type_id=1 ), ] db_session.add_all(comparison_types) db_session.add_all(comparisons) db_session.commit() db_session.close() # Expected result for request without any parameters. expected = [ { 'id': 1, 'state': STATE_STRINGS[1], 'time': '9999-01-01 00:00:00', 'comparison_type': { 'id': 1, 'name': '1', }, }, { 'id': 2, 'state': STATE_STRINGS[0], 'time': '1000-01-01 00:00:00', 'comparison_type': { 'id': 2, 'name': '2', }, }, { 'id': 3, 'state': STATE_STRINGS[0], 'time': '2018-01-01 00:00:00', 'comparison_type': { 'id': 1, 'name': '1', }, }, { 'id': 4, 'state': STATE_STRINGS[1], 'time': '2018-03-01 00:00:00', 'comparison_type': { 'id': 1, 'name': '1', }, }, ] # Tuples of query parameters and corresponding expected result. tuples_params_results = [ ({}, expected), ({'id': '2'}, [expected[1]]), ({'state': STATE_STRINGS[0]}, [expected[1], expected[2]]), ({'before': '2018-02-01 00:00:00'}, [expected[1], expected[2]]), ({'after': '2018-02-01 00:00:00'}, [expected[0], expected[3]]), ({'comparison_type_id': '1'}, [expected[0], expected[2], expected[3]]), ({'comparison_type_name': '2'}, [expected[1]]), ({'limit': '2'}, [expected[0], expected[1]]), ({'offset': '3'}, [expected[3]]), ( { 'comparison_type_id': '1', 'state': STATE_STRINGS[1], 'after': '2017-01-01 00:00:00', }, [expected[0], expected[3]] ), ] class RESTTestComparisonTypesFilled(RESTTestListsFilled): """Tests for getting comparison types from filled database.""" route = RESTTestComparisonTypesEmpty.route param_choices = RESTTestComparisonTypesEmpty.param_choices def fill_db(self): """Fill database. Called in setUp.""" db_session = database.session() comparison_types = [ database.ComparisonType(id=1, name='1'), database.ComparisonType(id=2, name='2'), database.ComparisonType(id=3, name='3'), database.ComparisonType(id=4, name='4'), database.ComparisonType(id=5, name='5'), ] db_session.add_all(comparison_types) db_session.commit() db_session.close() # Expected result for request without any parameters. expected = [ {'id': 1, 'name': '1'}, {'id': 2, 'name': '2'}, {'id': 3, 'name': '3'}, {'id': 4, 'name': '4'}, {'id': 5, 'name': '5'}, ] # Tuples of query parameters and corresponding expected result. tuples_params_results = [ ({}, expected), ({'id': '2'}, [expected[1]]), ({'name': '4'}, [expected[3]]), ({'limit': '3'}, [expected[0], expected[1], expected[2]]), ({'offset': '3'}, [expected[3], expected[4]]), ({'id': '4', 'name': '4'}, [expected[3]]), ]
10,668
3,228
from urllib.parse import quote from django import template register = template.Library() @register.filter def urlify(value): return quote(value)
150
41
# Generated by Django 2.0.6 on 2018-06-11 23:50 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('producto', '0003_auto_20180611_2248'), ] operations = [ migrations.RenameModel( old_name='Venta', new_name='Ventas', ), ]
330
130
from datetime import datetime from pydantic import BaseModel from typing import Optional, List from fastapi import Query class Service(BaseModel): """ A class for single service """ id: str ptvId: Optional[str] = None type: Optional[str] = None subtype: Optional[str] = None organizations: list name: dict descriptions: dict requirement: dict targetGroups: dict serviceClasses: dict lifeEvents: dict areas: dict lastUpdated: datetime nameAutoTranslated: dict class ServiceClass(BaseModel): """ A class for single service class """ name: str code: str class ServiceChannel(BaseModel): """ A class for single service channel """ id: str ptvId: Optional[str] = None type: Optional[str] = None areaType: Optional[str] = None organizationId: Optional[str] = None serviceIds: list name: dict descriptions: dict webPages: dict emails: dict phoneNumbers: dict areas: dict addresses: dict channelUrls: dict lastUpdated: datetime class ServiceQuery(BaseModel): """ A class for request payoad for service filtering """ include_channels: Optional[bool] = False priorization: Optional[str] = Query("local", regex="^(local|national)$") municipalities: Optional[List[str]] = [] life_events: Optional[List[str]] = [] service_classes: Optional[List[str]] = [] limit_k: Optional[int] = 20 translate_missing_texts: Optional[bool] = False class ServiceRecommendQuery(BaseModel): """ A class for payload for free text service recommendation """ need_text: str municipalities: Optional[List[str]] = [] life_events: Optional[List[str]] = [] service_classes: Optional[List[str]] = [] top_k: Optional[int] = 20 score_threshold: Optional[float] = 0.0 text_recommender: Optional[str] = Query("all", regex="^(nlp|lexical|all)$") language: Optional[str] = Query("fi", regex="^(fi|en|sv)$") translate_missing_texts: Optional[bool] = False class ServiceRecommendConversationQuery(BaseModel): """ A class for payload for service recommendation based on conversation """ mode: Optional[str] = Query("infer", regex="^(search|intent|conversation|infer)$") municipalities: Optional[List[str]] = [] life_events: Optional[List[str]] = [] service_classes: Optional[List[str]] = [] top_k: Optional[int] = 20 score_threshold: Optional[float] = 0.0 text_recommender: Optional[str] = Query("all", regex="^(nlp|lexical|all)$") language: Optional[str] = Query("fi", regex="^(fi|en|sv)$") translate_missing_texts: Optional[bool] = False class ServiceRecommendIntentQuery(BaseModel): """ A class for payload for intent based service recommendation """ intent: str municipalities: Optional[List[str]] = [] life_events: Optional[List[str]] = [] service_classes: Optional[List[str]] = [] translate_missing_texts: Optional[bool] = False class ServiceRecommendIntentAndOptionsQuery(BaseModel): """ A class for payload for intent and options based service recommendation, options can be got from database or can be overwritten in the interface """ intent: str municipalities: Optional[List[str]] = [] life_events: Optional[List[str]] = [] service_classes: Optional[List[str]] = [] score_threshold: Optional[float] = 0.0 need_text: Optional[str] = None text_recommender: Optional[str] = Query("all", regex="^(nlp|lexical|all)$") language: Optional[str] = Query("fi", regex="^(fi|en|sv)$") priorization: Optional[str] = Query("local", regex="^(local|national)$") limit_k: Optional[int] = 20 translate_missing_texts: Optional[bool] = False class ServiceClassRecommendQuery(BaseModel): """ A class for payload for free text service class recommendation """ need_text: str top_k: Optional[int] = 20 class ServiceClassRecommendConversationQuery(BaseModel): """ A class for payload for service class recommendation based on conversation """ top_k: Optional[int] = 20
4,173
1,248
# IBM_PROLOG_BEGIN_TAG # This is an automatically generated prolog. # # $Source: src/test/testcases/testPSUReadSbeMem.py $ # # OpenPOWER sbe Project # # Contributors Listed Below - COPYRIGHT 2017,2019 # [+] International Business Machines Corp. # # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. See the License for the specific language governing # permissions and limitations under the License. # # IBM_PROLOG_END_TAG from __future__ import print_function import sys sys.path.append("targets/p9_nimbus/sbeTest" ) sys.path.append("targets/p9_axone/sbeTest" ) import testPSUUtil import testRegistry as reg import testUtil import testMemUtil #------------------------------- # This is a Test Expected Data #------------------------------- def getdoubleword(dataInInt): hex_string = '0'*(16-len(str(hex(dataInInt))[:18][2:])) + str(hex(dataInInt))[:18][2:] return hex_string def readSeeprom(offset, size, destAddr, primStatus, secStatus): ''' #------------------------------------------------------------------------------------------------------------------------------ # SBE side test data - #------------------------------------------------------------------------------------------------------------------------------ ''' sbe_test_data = ( #----------------------------------------------------------------------------------------------------- # OP Reg ValueToWrite size Test Expected Data Description #----------------------------------------------------------------------------------------------------- # FFDC Size, Pass CMD Size ["write", reg.REG_MBOX0, "0000010000F0D703", 8, "None", "Writing to MBOX0 address"], # seeprom offset, Size ["write", reg.REG_MBOX1, getdoubleword((offset<<32)+size), 8, "None", "Writing to MBOX1 address"], # response Addr ["write", reg.REG_MBOX2, getdoubleword(destAddr), 8, "None", "Writing to MBOX2 address"], ["write", reg.PSU_SBE_DOORBELL_REG_WO_OR, "8000000000000000", 8, "None", "Update SBE Doorbell register to interrupt SBE"], ) ''' #--------------------- # Host side test data - SUCCESS #--------------------- ''' host_test_data_success = ( #---------------------------------------------------------------------------------------------------------------- # OP Reg ValueToWrite size Test Expected Data Description #---------------------------------------------------------------------------------------------------------------- ["read", reg.REG_MBOX4, "0", 8, getdoubleword((primStatus<<48)+(secStatus<<32)+0xF0D703), "Reading Host MBOX4 data to Validate"], ) ''' #----------------------------------------------------------------------- # Do not modify - Used to simulate interrupt on Ringing Doorbell on Host #----------------------------------------------------------------------- ''' host_polling_data = ( #---------------------------------------------------------------------------------------------------------------- # OP Reg ValueToWrite size Test Expected Data Description #---------------------------------------------------------------------------------------------------------------- ["read", reg.PSU_HOST_DOORBELL_REG_WO_OR, "0", 8, "8000000000000000", "Reading Host Doorbell for Interrupt Bit0"], ) # Run Simics initially testUtil.runCycles( 10000000 ); # Intialize the class obj instances regObj = testPSUUtil.registry() # Registry obj def for operation # HOST->SBE data set execution regObj.ExecuteTestOp( testPSUUtil.simSbeObj, sbe_test_data ) print("\n Poll on Host side for INTR ...\n") #Poll on HOST DoorBell Register for interrupt regObj.pollingOn( testPSUUtil.simSbeObj, host_polling_data, 5 ) #SBE->HOST data set execution regObj.ExecuteTestOp( testPSUUtil.simSbeObj, host_test_data_success ) #------------------------- # Main Function #------------------------- def main(): # Run Simics initially testUtil.runCycles( 10000000 ); print("\n Execute SBE Test - Read SBE Mem\n") ''' Test Case 1 ''' readSeeprom(0, 128, 0x08000000, 0, 0) print("SUCCESS: read seeprom valid") # Read data from cache and verify its contents # seeprom header seepprmHdr = 'XIP SEPM' #read from cache readData = testMemUtil.getmem(0x08000000, 0x80, 0x02) for byte in range(len(seepprmHdr)): if( ord(seepprmHdr[byte]) != readData[byte ]): print("Data mismtach at: ", byte) print(" expected: ", ord(seepprmHdr[byte])) print(" Actual: ", readData[byte]) raise Exception('data mistmach'); ''' Test Case 2 ''' readSeeprom(0x38CA0, 0x180, 0x8973780, 0, 0) print("SUCCESS: read seeprom HB testcase") ''' Test Case 3 ''' readSeeprom(0x0, 0x40, 0x08000000, 0x03, 0x19) print("SUCCESS: read seeprom size not aligned") ''' Test Case 4 ''' readSeeprom(0x3fe80, 0x180, 0x08000000, 0x03, 0x19) print("SUCCESS: read seeprom size exceeded") ''' Test Case 5 ''' readSeeprom(0x7, 0x40, 0x08000000, 0x03, 0x19) print("SUCCESS: read seeprom offset not aligned") if __name__ == "__main__": if testUtil.getMachineName() == "axone": try: main() except: print ( "\nTest Suite completed with error(s)" ) testUtil.collectFFDC() raise() print ( "\nTest Suite completed with no errors" ) else: main() if err: print ( "\nTest Suite completed with error(s)" ) #sys.exit(1) else: print ( "\nTest Suite completed with no errors" ) #sys.exit(0);
6,481
2,017
#!/usr/bin/env python3 """Module containing the ClusteringPredict class and the command line interface.""" import argparse import pandas as pd import joblib from biobb_common.generic.biobb_object import BiobbObject from sklearn.preprocessing import StandardScaler from biobb_common.configuration import settings from biobb_common.tools import file_utils as fu from biobb_common.tools.file_utils import launchlogger from biobb_ml.clustering.common import * class ClusteringPredict(BiobbObject): """ | biobb_ml ClusteringPredict | Makes predictions from an input dataset and a given clustering model. | Makes predictions from an input dataset (provided either as a file or as a dictionary property) and a given clustering model fitted with `KMeans <https://scikit-learn.org/stable/modules/generated/sklearn.cluster.KMeans.html>`_ method. Args: input_model_path (str): Path to the input model. File type: input. `Sample file <https://github.com/bioexcel/biobb_ml/raw/master/biobb_ml/test/data/clustering/model_clustering_predict.pkl>`_. Accepted formats: pkl (edam:format_3653). input_dataset_path (str) (Optional): Path to the dataset to predict. File type: input. `Sample file <https://github.com/bioexcel/biobb_ml/raw/master/biobb_ml/test/data/clustering/input_clustering_predict.csv>`_. Accepted formats: csv (edam:format_3752). output_results_path (str): Path to the output results file. File type: output. `Sample file <https://github.com/bioexcel/biobb_ml/raw/master/biobb_ml/test/reference/clustering/ref_output_results_clustering_predict.csv>`_. Accepted formats: csv (edam:format_3752). properties (dic - Python dictionary object containing the tool parameters, not input/output files): * **predictions** (*list*) - (None) List of dictionaries with all values you want to predict targets. It will be taken into account only in case **input_dataset_path** is not provided. Format: [{ 'var1': 1.0, 'var2': 2.0 }, { 'var1': 4.0, 'var2': 2.7 }] for datasets with headers and [[ 1.0, 2.0 ], [ 4.0, 2.7 ]] for datasets without headers. * **remove_tmp** (*bool*) - (True) [WF property] Remove temporal files. * **restart** (*bool*) - (False) [WF property] Do not execute if output files exist. Examples: This is a use example of how to use the building block from Python:: from biobb_ml.clustering.clustering_predict import clustering_predict prop = { 'predictions': [ { 'var1': 1.0, 'var2': 2.0 }, { 'var1': 4.0, 'var2': 2.7 } ] } clustering_predict(input_model_path='/path/to/myModel.pkl', output_results_path='/path/to/newPredictedResults.csv', input_dataset_path='/path/to/myDataset.csv', properties=prop) Info: * wrapped_software: * name: scikit-learn * version: >=0.24.2 * license: BSD 3-Clause * ontology: * name: EDAM * schema: http://edamontology.org/EDAM.owl """ def __init__(self, input_model_path, output_results_path, input_dataset_path=None, properties=None, **kwargs) -> None: properties = properties or {} # Call parent class constructor super().__init__(properties) # Input/Output files self.io_dict = { "in": { "input_model_path": input_model_path, "input_dataset_path": input_dataset_path }, "out": { "output_results_path": output_results_path } } # Properties specific for BB self.predictions = properties.get('predictions', []) self.properties = properties # Check the properties self.check_properties(properties) def check_data_params(self, out_log, err_log): """ Checks all the input/output paths and parameters """ self.io_dict["in"]["input_model_path"] = check_input_path(self.io_dict["in"]["input_model_path"], "input_model_path", out_log, self.__class__.__name__) self.io_dict["out"]["output_results_path"] = check_output_path(self.io_dict["out"]["output_results_path"],"output_results_path", False, out_log, self.__class__.__name__) if self.io_dict["in"]["input_dataset_path"]: self.io_dict["in"]["input_dataset_path"] = check_input_path(self.io_dict["in"]["input_dataset_path"], "input_dataset_path", out_log, self.__class__.__name__) @launchlogger def launch(self) -> int: """Execute the :class:`ClusteringPredict <clustering.clustering_predict.ClusteringPredict>` clustering.clustering_predict.ClusteringPredict object.""" # check input/output paths and parameters self.check_data_params(self.out_log, self.err_log) # Setup Biobb if self.check_restart(): return 0 self.stage_files() fu.log('Getting model from %s' % self.io_dict["in"]["input_model_path"], self.out_log, self.global_log) with open(self.io_dict["in"]["input_model_path"], "rb") as f: while True: try: m = joblib.load(f) if (isinstance(m, KMeans)): new_model = m if isinstance(m, StandardScaler): scaler = m if isinstance(m, dict): variables = m except EOFError: break if self.io_dict["in"]["input_dataset_path"]: # load dataset from input_dataset_path file fu.log('Getting dataset from %s' % self.io_dict["in"]["input_dataset_path"], self.out_log, self.global_log) if 'columns' in variables['predictors']: labels = getHeader(self.io_dict["in"]["input_dataset_path"]) skiprows = 1 else: labels = None skiprows = None new_data_table = pd.read_csv(self.io_dict["in"]["input_dataset_path"], header = None, sep="\s+|;|:|,|\t", engine="python", skiprows=skiprows, names=labels) else: # load dataset from properties if 'columns' in variables['predictors']: # sorting self.properties in the correct order given by variables['predictors']['columns'] index_map = { v: i for i, v in enumerate(variables['predictors']['columns']) } predictions = [] for i, pred in enumerate(self.predictions): sorted_pred = sorted(pred.items(), key=lambda pair: index_map[pair[0]]) predictions.append(dict(sorted_pred)) new_data_table = pd.DataFrame(data=get_list_of_predictors(predictions),columns=get_keys_of_predictors(predictions)) else: predictions = self.predictions new_data_table = pd.DataFrame(data=predictions) if variables['scale']: fu.log('Scaling dataset', self.out_log, self.global_log) new_data = scaler.transform(new_data_table) else: new_data = new_data_table p = new_model.predict(new_data) new_data_table['cluster'] = p fu.log('Predicting results\n\nPREDICTION RESULTS\n\n%s\n' % new_data_table, self.out_log, self.global_log) fu.log('Saving results to %s' % self.io_dict["out"]["output_results_path"], self.out_log, self.global_log) new_data_table.to_csv(self.io_dict["out"]["output_results_path"], index = False, header=True, float_format='%.3f') return 0 def clustering_predict(input_model_path: str, output_results_path: str, input_dataset_path: str = None, properties: dict = None, **kwargs) -> int: """Execute the :class:`ClusteringPredict <clustering.clustering_predict.ClusteringPredict>` class and execute the :meth:`launch() <clustering.clustering_predict.ClusteringPredict.launch>` method.""" return ClusteringPredict(input_model_path=input_model_path, output_results_path=output_results_path, input_dataset_path=input_dataset_path, properties=properties, **kwargs).launch() def main(): """Command line execution of this building block. Please check the command line documentation.""" parser = argparse.ArgumentParser(description="Makes predictions from an input dataset and a given clustering model.", formatter_class=lambda prog: argparse.RawTextHelpFormatter(prog, width=99999)) parser.add_argument('--config', required=False, help='Configuration file') # Specific args of each building block required_args = parser.add_argument_group('required arguments') required_args.add_argument('--input_model_path', required=True, help='Path to the input model. Accepted formats: pkl.') required_args.add_argument('--output_results_path', required=True, help='Path to the output results file. Accepted formats: csv.') parser.add_argument('--input_dataset_path', required=False, help='Path to the dataset to predict. Accepted formats: csv.') args = parser.parse_args() args.config = args.config or "{}" properties = settings.ConfReader(config=args.config).get_prop_dic() # Specific call of each building block clustering_predict(input_model_path=args.input_model_path, output_results_path=args.output_results_path, input_dataset_path=args.input_dataset_path, properties=properties) if __name__ == '__main__': main()
9,780
2,846
from sprake import treeviz # we don't actually have any meaningful tests that we can do, but at least # we can do this
121
39
import sys tmpl = """ #include "yabs_mx_calc_table.h" #include <kernel/matrixnet/mn_sse.h> #include <library/archive/yarchive.h> #include <util/memory/blob.h> #include <util/generic/hash.h> #include <util/generic/ptr.h> #include <util/generic/singleton.h> using namespace NMatrixnet; extern "C" { extern const unsigned char MxFormulas[]; extern const ui32 MxFormulasSize; } namespace { struct TFml: public TBlob, public TMnSseInfo { inline TFml(const TBlob& b) : TBlob(b) , TMnSseInfo(Data(), Size()) { } }; struct TFormulas: public THashMap<size_t, TAutoPtr<TFml>> { inline TFormulas() { TBlob b = TBlob::NoCopy(MxFormulas, MxFormulasSize); TArchiveReader ar(b); %s } inline const TMnSseInfo& at(size_t n) const throw () { return *find(n)->second; } }; %s static func_descr_t yabs_funcs[] = { %s }; } yabs_mx_calc_table_t yabs_mx_calc_table = {YABS_MX_CALC_VERSION, 10000, 0, yabs_funcs}; """ if __name__ == '__main__': init = [] body = [] defs = {} for i in sys.argv[1:]: name = i.replace('.', '_') num = long(name.split('_')[1]) init.append('(*this)[%s] = new TFml(ar.ObjectBlobByKey("%s"));' % (num, '/' + i)) f1 = 'static void yabs_%s(size_t count, const float** args, double* res) {Singleton<TFormulas>()->at(%s).DoCalcRelevs(args, res, count);}' % (name, num) f2 = 'static size_t yabs_%s_factor_count() {return Singleton<TFormulas>()->at(%s).MaxFactorIndex() + 1;}' % (name, num) body.append(f1) body.append(f2) d1 = 'yabs_%s' % name d2 = 'yabs_%s_factor_count' % name defs[num] = '{%s, %s}' % (d1, d2) print tmpl % ('\n'.join(init), '\n\n'.join(body), ',\n'.join((defs.get(i, '{nullptr, nullptr}') for i in range(0, 10000))))
1,926
771